lib/aws/generated/cloud_watch_logs.ex

# WARNING: DO NOT EDIT, AUTO-GENERATED CODE!
# See https://github.com/aws-beam/aws-codegen for more details.

defmodule AWS.CloudWatchLogs do
  @moduledoc """
  You can use Amazon CloudWatch Logs to monitor, store, and access your log files
  from
  EC2 instances, CloudTrail, and other sources.

  You can then retrieve the associated
  log data from CloudWatch Logs using the CloudWatch console. Alternatively, you
  can use
  CloudWatch Logs commands in the Amazon Web Services CLI, CloudWatch Logs API, or
  CloudWatch
  Logs SDK.

  You can use CloudWatch Logs to:

    *

  **Monitor logs from EC2 instances in real time**: You
  can use CloudWatch Logs to monitor applications and systems using log data. For
  example,
  CloudWatch Logs can track the number of errors that occur in your application
  logs. Then,
  it can send you a notification whenever the rate of errors exceeds a threshold
  that you
  specify. CloudWatch Logs uses your log data for monitoring so no code changes
  are
  required. For example, you can monitor application logs for specific literal
  terms (such
  as "NullReferenceException"). You can also count the number of occurrences of a
  literal
  term at a particular position in log data (such as "404" status codes in an
  Apache access
  log). When the term you are searching for is found, CloudWatch Logs reports the
  data to a
  CloudWatch metric that you specify.

    *

  **Monitor CloudTrail logged events**: You
  can create alarms in CloudWatch and receive notifications of particular API
  activity as
  captured by CloudTrail. You can use the notification to perform troubleshooting.

    *

  **Archive log data**: You can use CloudWatch Logs to
  store your log data in highly durable storage. You can change the log retention
  setting so
  that any log events earlier than this setting are automatically deleted. The
  CloudWatch
  Logs agent helps to quickly send both rotated and non-rotated log data off of a
  host and
  into the log service. You can then access the raw log data when you need it.
  """

  alias AWS.Client
  alias AWS.Request

  @typedoc """

  ## Example:
      
      invalid_sequence_token_exception() :: %{
        "expectedSequenceToken" => String.t() | atom(),
        "message" => String.t() | atom()
      }
      
  """
  @type invalid_sequence_token_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_anomalies_response() :: %{
        "anomalies" => list(anomaly()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type list_anomalies_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      add_key_entry() :: %{
        "key" => String.t() | atom(),
        "overwriteIfExists" => boolean(),
        "value" => String.t() | atom()
      }
      
  """
  @type add_key_entry() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      trim_string() :: %{
        "withKeys" => list(String.t() | atom())
      }
      
  """
  @type trim_string() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_resource_policies_request() :: %{
        optional("limit") => integer(),
        optional("nextToken") => String.t() | atom(),
        optional("policyScope") => list(any()),
        optional("resourceArn") => String.t() | atom()
      }
      
  """
  @type describe_resource_policies_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delivery() :: %{
        "arn" => String.t() | atom(),
        "deliveryDestinationArn" => String.t() | atom(),
        "deliveryDestinationType" => list(any()),
        "deliverySourceName" => String.t() | atom(),
        "fieldDelimiter" => String.t() | atom(),
        "id" => String.t() | atom(),
        "recordFields" => list(String.t() | atom()),
        "s3DeliveryConfiguration" => s3_delivery_configuration(),
        "tags" => map()
      }
      
  """
  @type delivery() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      rejected_entity_info() :: %{
        "errorType" => list(any())
      }
      
  """
  @type rejected_entity_info() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_integration_response() :: %{}
      
  """
  @type delete_integration_response() :: %{}

  @typedoc """

  ## Example:
      
      tag_resource_request() :: %{
        required("resourceArn") => String.t() | atom(),
        required("tags") => map()
      }
      
  """
  @type tag_resource_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_log_groups_request() :: %{
        optional("accountIdentifiers") => list(String.t() | atom()),
        optional("includeLinkedAccounts") => boolean(),
        optional("limit") => integer(),
        optional("logGroupClass") => list(any()),
        optional("logGroupIdentifiers") => list(String.t() | atom()),
        optional("logGroupNamePattern") => String.t() | atom(),
        optional("logGroupNamePrefix") => String.t() | atom(),
        optional("nextToken") => String.t() | atom()
      }
      
  """
  @type describe_log_groups_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      parse_cloudfront() :: %{
        "source" => String.t() | atom()
      }
      
  """
  @type parse_cloudfront() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      copy_value() :: %{
        "entries" => list(copy_value_entry())
      }
      
  """
  @type copy_value() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_query_definition_request() :: %{
        required("queryDefinitionId") => String.t() | atom()
      }
      
  """
  @type delete_query_definition_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_log_anomaly_detector_request() :: %{
        required("anomalyDetectorArn") => String.t() | atom()
      }
      
  """
  @type get_log_anomaly_detector_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_delivery_destination_policy_request() :: %{
        required("deliveryDestinationName") => String.t() | atom(),
        required("deliveryDestinationPolicy") => String.t() | atom()
      }
      
  """
  @type put_delivery_destination_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      account_policy() :: %{
        "accountId" => String.t() | atom(),
        "lastUpdatedTime" => float(),
        "policyDocument" => String.t() | atom(),
        "policyName" => String.t() | atom(),
        "policyType" => list(any()),
        "scope" => list(any()),
        "selectionCriteria" => String.t() | atom()
      }
      
  """
  @type account_policy() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_data_source() :: %{
        "dataSourceName" => String.t() | atom(),
        "status" => open_search_resource_status()
      }
      
  """
  @type open_search_data_source() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_index_policy_response() :: %{}
      
  """
  @type delete_index_policy_response() :: %{}

  @typedoc """

  ## Example:
      
      create_export_task_request() :: %{
        optional("destinationPrefix") => String.t() | atom(),
        optional("logStreamNamePrefix") => String.t() | atom(),
        optional("taskName") => String.t() | atom(),
        required("destination") => String.t() | atom(),
        required("from") => float(),
        required("logGroupName") => String.t() | atom(),
        required("to") => float()
      }
      
  """
  @type create_export_task_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      tag_log_group_request() :: %{
        required("logGroupName") => String.t() | atom(),
        required("tags") => map()
      }
      
  """
  @type tag_log_group_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_log_groups_for_query_response() :: %{
        "logGroupIdentifiers" => list(String.t() | atom()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type list_log_groups_for_query_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      type_converter() :: %{
        "entries" => list(type_converter_entry())
      }
      
  """
  @type type_converter() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_query_definitions_request() :: %{
        optional("maxResults") => integer(),
        optional("nextToken") => String.t() | atom(),
        optional("queryDefinitionNamePrefix") => String.t() | atom(),
        optional("queryLanguage") => list(any())
      }
      
  """
  @type describe_query_definitions_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_tags_log_group_response() :: %{
        "tags" => map()
      }
      
  """
  @type list_tags_log_group_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_destination_response() :: %{
        "destination" => destination()
      }
      
  """
  @type put_destination_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      filter_log_events_response() :: %{
        "events" => list(filtered_log_event()),
        "nextToken" => String.t() | atom(),
        "searchedLogStreams" => list(searched_log_stream())
      }
      
  """
  @type filter_log_events_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_resource_config() :: %{
        "applicationArn" => String.t() | atom(),
        "dashboardViewerPrincipals" => list(String.t() | atom()),
        "dataSourceRoleArn" => String.t() | atom(),
        "kmsKeyArn" => String.t() | atom(),
        "retentionDays" => integer()
      }
      
  """
  @type open_search_resource_config() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_destination_request() :: %{
        optional("tags") => map(),
        required("destinationName") => String.t() | atom(),
        required("roleArn") => String.t() | atom(),
        required("targetArn") => String.t() | atom()
      }
      
  """
  @type put_destination_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_query_results_response() :: %{
        "encryptionKey" => String.t() | atom(),
        "queryLanguage" => list(any()),
        "results" => list(list(result_field())()),
        "statistics" => query_statistics(),
        "status" => list(any())
      }
      
  """
  @type get_query_results_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_queries_request() :: %{
        optional("logGroupName") => String.t() | atom(),
        optional("maxResults") => integer(),
        optional("nextToken") => String.t() | atom(),
        optional("queryLanguage") => list(any()),
        optional("status") => list(any())
      }
      
  """
  @type describe_queries_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_account_policy_response() :: %{
        "accountPolicy" => account_policy()
      }
      
  """
  @type put_account_policy_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_transformer_request() :: %{
        required("logGroupIdentifier") => String.t() | atom()
      }
      
  """
  @type get_transformer_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      entity() :: %{
        "attributes" => map(),
        "keyAttributes" => map()
      }
      
  """
  @type entity() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      substitute_string_entry() :: %{
        "from" => String.t() | atom(),
        "source" => String.t() | atom(),
        "to" => String.t() | atom()
      }
      
  """
  @type substitute_string_entry() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      type_converter_entry() :: %{
        "key" => String.t() | atom(),
        "type" => list(any())
      }
      
  """
  @type type_converter_entry() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      split_string_entry() :: %{
        "delimiter" => String.t() | atom(),
        "source" => String.t() | atom()
      }
      
  """
  @type split_string_entry() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      copy_value_entry() :: %{
        "overwriteIfExists" => boolean(),
        "source" => String.t() | atom(),
        "target" => String.t() | atom()
      }
      
  """
  @type copy_value_entry() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_queries_response() :: %{
        "nextToken" => String.t() | atom(),
        "queries" => list(query_info())
      }
      
  """
  @type describe_queries_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_log_events_request() :: %{
        optional("entity") => entity(),
        optional("sequenceToken") => String.t() | atom(),
        required("logEvents") => list(input_log_event()),
        required("logGroupName") => String.t() | atom(),
        required("logStreamName") => String.t() | atom()
      }
      
  """
  @type put_log_events_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_account_policy_request() :: %{
        optional("scope") => list(any()),
        optional("selectionCriteria") => String.t() | atom(),
        required("policyDocument") => String.t() | atom(),
        required("policyName") => String.t() | atom(),
        required("policyType") => list(any())
      }
      
  """
  @type put_account_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      data_already_accepted_exception() :: %{
        "expectedSequenceToken" => String.t() | atom(),
        "message" => String.t() | atom()
      }
      
  """
  @type data_already_accepted_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      export_task_execution_info() :: %{
        "completionTime" => float(),
        "creationTime" => float()
      }
      
  """
  @type export_task_execution_info() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_data_protection_policy_request() :: %{
        required("logGroupIdentifier") => String.t() | atom(),
        required("policyDocument") => String.t() | atom()
      }
      
  """
  @type put_data_protection_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_query_results_request() :: %{
        required("queryId") => String.t() | atom()
      }
      
  """
  @type get_query_results_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      substitute_string() :: %{
        "entries" => list(substitute_string_entry())
      }
      
  """
  @type substitute_string() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_delivery_request() :: %{
        required("id") => String.t() | atom()
      }
      
  """
  @type get_delivery_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      live_tail_session_metadata() :: %{
        "sampled" => boolean()
      }
      
  """
  @type live_tail_session_metadata() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      create_log_group_request() :: %{
        optional("kmsKeyId") => String.t() | atom(),
        optional("logGroupClass") => list(any()),
        optional("tags") => map(),
        required("logGroupName") => String.t() | atom()
      }
      
  """
  @type create_log_group_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      fields_data() :: %{
        "data" => binary()
      }
      
  """
  @type fields_data() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_encryption_policy() :: %{
        "policyName" => String.t() | atom(),
        "status" => open_search_resource_status()
      }
      
  """
  @type open_search_encryption_policy() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_delivery_destination_request() :: %{
        optional("deliveryDestinationConfiguration") => delivery_destination_configuration(),
        optional("deliveryDestinationType") => list(any()),
        optional("outputFormat") => list(any()),
        optional("tags") => map(),
        required("name") => String.t() | atom()
      }
      
  """
  @type put_delivery_destination_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_delivery_source_response() :: %{
        "deliverySource" => delivery_source()
      }
      
  """
  @type get_delivery_source_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      subscription_filter() :: %{
        "applyOnTransformedLogs" => boolean(),
        "creationTime" => float(),
        "destinationArn" => String.t() | atom(),
        "distribution" => list(any()),
        "emitSystemFields" => list(String.t() | atom()),
        "fieldSelectionCriteria" => String.t() | atom(),
        "filterName" => String.t() | atom(),
        "filterPattern" => String.t() | atom(),
        "logGroupName" => String.t() | atom(),
        "roleArn" => String.t() | atom()
      }
      
  """
  @type subscription_filter() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_export_tasks_response() :: %{
        "exportTasks" => list(export_task()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_export_tasks_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      live_tail_session_start() :: %{
        "logEventFilterPattern" => String.t() | atom(),
        "logGroupIdentifiers" => list(String.t() | atom()),
        "logStreamNamePrefixes" => list(String.t() | atom()),
        "logStreamNames" => list(String.t() | atom()),
        "requestId" => String.t() | atom(),
        "sessionId" => String.t() | atom()
      }
      
  """
  @type live_tail_session_start() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      create_log_stream_request() :: %{
        required("logGroupName") => String.t() | atom(),
        required("logStreamName") => String.t() | atom()
      }
      
  """
  @type create_log_stream_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_log_record_response() :: %{
        "logRecord" => map()
      }
      
  """
  @type get_log_record_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      invalid_operation_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type invalid_operation_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_log_groups_response() :: %{
        "logGroups" => list(log_group_summary()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type list_log_groups_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_integration_request() :: %{
        required("integrationName") => String.t() | atom(),
        required("integrationType") => list(any()),
        required("resourceConfig") => list()
      }
      
  """
  @type put_integration_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      create_delivery_response() :: %{
        "delivery" => delivery()
      }
      
  """
  @type create_delivery_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_account_policies_response() :: %{
        "accountPolicies" => list(account_policy()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_account_policies_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      update_delivery_configuration_response() :: %{}
      
  """
  @type update_delivery_configuration_response() :: %{}

  @typedoc """

  ## Example:
      
      update_anomaly_request() :: %{
        optional("anomalyId") => String.t() | atom(),
        optional("baseline") => boolean(),
        optional("patternId") => String.t() | atom(),
        optional("suppressionPeriod") => suppression_period(),
        optional("suppressionType") => list(any()),
        required("anomalyDetectorArn") => String.t() | atom()
      }
      
  """
  @type update_anomaly_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      parse_w_a_f() :: %{
        "source" => String.t() | atom()
      }
      
  """
  @type parse_w_a_f() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_retention_policy_request() :: %{
        required("logGroupName") => String.t() | atom(),
        required("retentionInDays") => integer()
      }
      
  """
  @type put_retention_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      start_query_request() :: %{
        optional("limit") => integer(),
        optional("logGroupIdentifiers") => list(String.t() | atom()),
        optional("logGroupName") => String.t() | atom(),
        optional("logGroupNames") => list(String.t() | atom()),
        optional("queryLanguage") => list(any()),
        required("endTime") => float(),
        required("queryString") => String.t() | atom(),
        required("startTime") => float()
      }
      
  """
  @type start_query_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_deliveries_request() :: %{
        optional("limit") => integer(),
        optional("nextToken") => String.t() | atom()
      }
      
  """
  @type describe_deliveries_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      session_timeout_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type session_timeout_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_log_group_fields_response() :: %{
        "logGroupFields" => list(log_group_field())
      }
      
  """
  @type get_log_group_fields_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      update_log_anomaly_detector_request() :: %{
        optional("anomalyVisibilityTime") => float(),
        optional("evaluationFrequency") => list(any()),
        optional("filterPattern") => String.t() | atom(),
        required("anomalyDetectorArn") => String.t() | atom(),
        required("enabled") => boolean()
      }
      
  """
  @type update_log_anomaly_detector_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_metric_filters_request() :: %{
        optional("filterNamePrefix") => String.t() | atom(),
        optional("limit") => integer(),
        optional("logGroupName") => String.t() | atom(),
        optional("metricName") => String.t() | atom(),
        optional("metricNamespace") => String.t() | atom(),
        optional("nextToken") => String.t() | atom()
      }
      
  """
  @type describe_metric_filters_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      session_streaming_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type session_streaming_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_data_protection_policy_response() :: %{
        "lastUpdatedTime" => float(),
        "logGroupIdentifier" => String.t() | atom(),
        "policyDocument" => String.t() | atom()
      }
      
  """
  @type put_data_protection_policy_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      integration_summary() :: %{
        "integrationName" => String.t() | atom(),
        "integrationStatus" => list(any()),
        "integrationType" => list(any())
      }
      
  """
  @type integration_summary() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      field_index() :: %{
        "fieldIndexName" => String.t() | atom(),
        "firstEventTime" => float(),
        "lastEventTime" => float(),
        "lastScanTime" => float(),
        "logGroupIdentifier" => String.t() | atom()
      }
      
  """
  @type field_index() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      processor() :: %{
        "addKeys" => add_keys(),
        "copyValue" => copy_value(),
        "csv" => csv(),
        "dateTimeConverter" => date_time_converter(),
        "deleteKeys" => delete_keys(),
        "grok" => grok(),
        "listToMap" => list_to_map(),
        "lowerCaseString" => lower_case_string(),
        "moveKeys" => move_keys(),
        "parseCloudfront" => parse_cloudfront(),
        "parseJSON" => parse_json(),
        "parseKeyValue" => parse_key_value(),
        "parsePostgres" => parse_postgres(),
        "parseRoute53" => parse_route53(),
        "parseToOCSF" => parse_to_o_c_s_f(),
        "parseVPC" => parse_vpc(),
        "parseWAF" => parse_w_a_f(),
        "renameKeys" => rename_keys(),
        "splitString" => split_string(),
        "substituteString" => substitute_string(),
        "trimString" => trim_string(),
        "typeConverter" => type_converter(),
        "upperCaseString" => upper_case_string()
      }
      
  """
  @type processor() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      untag_resource_request() :: %{
        required("resourceArn") => String.t() | atom(),
        required("tagKeys") => list(String.t() | atom())
      }
      
  """
  @type untag_resource_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_index_policy_response() :: %{
        "indexPolicy" => index_policy()
      }
      
  """
  @type put_index_policy_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      configuration_template() :: %{
        "allowedActionForAllowVendedLogsDeliveryForResource" => String.t() | atom(),
        "allowedFieldDelimiters" => list(String.t() | atom()),
        "allowedFields" => list(record_field()),
        "allowedOutputFormats" => list(list(any())()),
        "allowedSuffixPathFields" => list(String.t() | atom()),
        "defaultDeliveryConfigValues" => configuration_template_delivery_config_values(),
        "deliveryDestinationType" => list(any()),
        "logType" => String.t() | atom(),
        "resourceType" => String.t() | atom(),
        "service" => String.t() | atom()
      }
      
  """
  @type configuration_template() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_integration_response() :: %{
        "integrationName" => String.t() | atom(),
        "integrationStatus" => list(any())
      }
      
  """
  @type put_integration_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_destinations_response() :: %{
        "destinations" => list(destination()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_destinations_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      live_tail_session_update() :: %{
        "sessionMetadata" => live_tail_session_metadata(),
        "sessionResults" => list(live_tail_session_log_event())
      }
      
  """
  @type live_tail_session_update() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_transformer_response() :: %{
        "creationTime" => float(),
        "lastModifiedTime" => float(),
        "logGroupIdentifier" => String.t() | atom(),
        "transformerConfig" => list(processor())
      }
      
  """
  @type get_transformer_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      transformed_log_record() :: %{
        "eventMessage" => String.t() | atom(),
        "eventNumber" => float(),
        "transformedEventMessage" => String.t() | atom()
      }
      
  """
  @type transformed_log_record() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      query_compile_error_location() :: %{
        "endCharOffset" => integer(),
        "startCharOffset" => integer()
      }
      
  """
  @type query_compile_error_location() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_log_groups_request() :: %{
        optional("accountIdentifiers") => list(String.t() | atom()),
        optional("includeLinkedAccounts") => boolean(),
        optional("limit") => integer(),
        optional("logGroupClass") => list(any()),
        optional("logGroupNamePattern") => String.t() | atom(),
        optional("nextToken") => String.t() | atom()
      }
      
  """
  @type list_log_groups_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      operation_aborted_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type operation_aborted_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_integrations_request() :: %{
        optional("integrationNamePrefix") => String.t() | atom(),
        optional("integrationStatus") => list(any()),
        optional("integrationType") => list(any())
      }
      
  """
  @type list_integrations_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_query_definitions_response() :: %{
        "nextToken" => String.t() | atom(),
        "queryDefinitions" => list(query_definition())
      }
      
  """
  @type describe_query_definitions_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_delivery_response() :: %{
        "delivery" => delivery()
      }
      
  """
  @type get_delivery_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_subscription_filters_request() :: %{
        optional("filterNamePrefix") => String.t() | atom(),
        optional("limit") => integer(),
        optional("nextToken") => String.t() | atom(),
        required("logGroupName") => String.t() | atom()
      }
      
  """
  @type describe_subscription_filters_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_destination_policy_request() :: %{
        optional("forceUpdate") => boolean(),
        required("accessPolicy") => String.t() | atom(),
        required("destinationName") => String.t() | atom()
      }
      
  """
  @type put_destination_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_integration_response() :: %{
        "integrationDetails" => list(),
        "integrationName" => String.t() | atom(),
        "integrationStatus" => list(any()),
        "integrationType" => list(any())
      }
      
  """
  @type get_integration_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      lower_case_string() :: %{
        "withKeys" => list(String.t() | atom())
      }
      
  """
  @type lower_case_string() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      input_log_event() :: %{
        "message" => String.t() | atom(),
        "timestamp" => float()
      }
      
  """
  @type input_log_event() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_retention_policy_request() :: %{
        required("logGroupName") => String.t() | atom()
      }
      
  """
  @type delete_retention_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      malformed_query_exception() :: %{
        "message" => String.t() | atom(),
        "queryCompileError" => query_compile_error()
      }
      
  """
  @type malformed_query_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      live_tail_session_log_event() :: %{
        "ingestionTime" => float(),
        "logGroupIdentifier" => String.t() | atom(),
        "logStreamName" => String.t() | atom(),
        "message" => String.t() | atom(),
        "timestamp" => float()
      }
      
  """
  @type live_tail_session_log_event() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      s3_delivery_configuration() :: %{
        "enableHiveCompatiblePath" => boolean(),
        "suffixPath" => String.t() | atom()
      }
      
  """
  @type s3_delivery_configuration() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_delivery_destination_request() :: %{
        required("name") => String.t() | atom()
      }
      
  """
  @type get_delivery_destination_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      conflict_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type conflict_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      resource_not_found_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type resource_not_found_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_integration_request() :: %{
        optional("force") => boolean(),
        required("integrationName") => String.t() | atom()
      }
      
  """
  @type delete_integration_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_index_policies_request() :: %{
        optional("nextToken") => String.t() | atom(),
        required("logGroupIdentifiers") => list(String.t() | atom())
      }
      
  """
  @type describe_index_policies_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_field_indexes_request() :: %{
        optional("nextToken") => String.t() | atom(),
        required("logGroupIdentifiers") => list(String.t() | atom())
      }
      
  """
  @type describe_field_indexes_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_log_anomaly_detector_request() :: %{
        required("anomalyDetectorArn") => String.t() | atom()
      }
      
  """
  @type delete_log_anomaly_detector_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      disassociate_kms_key_request() :: %{
        optional("logGroupName") => String.t() | atom(),
        optional("resourceIdentifier") => String.t() | atom()
      }
      
  """
  @type disassociate_kms_key_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_transformer_request() :: %{
        required("logGroupIdentifier") => String.t() | atom()
      }
      
  """
  @type delete_transformer_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_export_tasks_request() :: %{
        optional("limit") => integer(),
        optional("nextToken") => String.t() | atom(),
        optional("statusCode") => list(any()),
        optional("taskId") => String.t() | atom()
      }
      
  """
  @type describe_export_tasks_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_collection() :: %{
        "collectionArn" => String.t() | atom(),
        "collectionEndpoint" => String.t() | atom(),
        "status" => open_search_resource_status()
      }
      
  """
  @type open_search_collection() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_log_anomaly_detector_response() :: %{
        "anomalyDetectorStatus" => list(any()),
        "anomalyVisibilityTime" => float(),
        "creationTimeStamp" => float(),
        "detectorName" => String.t() | atom(),
        "evaluationFrequency" => list(any()),
        "filterPattern" => String.t() | atom(),
        "kmsKeyId" => String.t() | atom(),
        "lastModifiedTimeStamp" => float(),
        "logGroupArnList" => list(String.t() | atom())
      }
      
  """
  @type get_log_anomaly_detector_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_query_definition_response() :: %{
        "success" => boolean()
      }
      
  """
  @type delete_query_definition_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_log_streams_response() :: %{
        "logStreams" => list(log_stream()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_log_streams_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_delivery_destination_policy_response() :: %{
        "policy" => policy()
      }
      
  """
  @type put_delivery_destination_policy_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      export_task_status() :: %{
        "code" => list(any()),
        "message" => String.t() | atom()
      }
      
  """
  @type export_task_status() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delivery_destination_configuration() :: %{
        "destinationResourceArn" => String.t() | atom()
      }
      
  """
  @type delivery_destination_configuration() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      suppression_period() :: %{
        "suppressionUnit" => list(any()),
        "value" => integer()
      }
      
  """
  @type suppression_period() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      service_quota_exceeded_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type service_quota_exceeded_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      pattern_token() :: %{
        "dynamicTokenPosition" => integer(),
        "enumerations" => map(),
        "inferredTokenName" => String.t() | atom(),
        "isDynamic" => boolean(),
        "tokenString" => String.t() | atom()
      }
      
  """
  @type pattern_token() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_log_stream_request() :: %{
        required("logGroupName") => String.t() | atom(),
        required("logStreamName") => String.t() | atom()
      }
      
  """
  @type delete_log_stream_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_index_policy_request() :: %{
        required("logGroupIdentifier") => String.t() | atom()
      }
      
  """
  @type delete_index_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      test_transformer_response() :: %{
        "transformedLogs" => list(transformed_log_record())
      }
      
  """
  @type test_transformer_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_data_access_policy() :: %{
        "policyName" => String.t() | atom(),
        "status" => open_search_resource_status()
      }
      
  """
  @type open_search_data_access_policy() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      create_export_task_response() :: %{
        "taskId" => String.t() | atom()
      }
      
  """
  @type create_export_task_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_delivery_destinations_response() :: %{
        "deliveryDestinations" => list(delivery_destination()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_delivery_destinations_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      csv() :: %{
        "columns" => list(String.t() | atom()),
        "delimiter" => String.t() | atom(),
        "quoteCharacter" => String.t() | atom(),
        "source" => String.t() | atom()
      }
      
  """
  @type csv() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      rename_key_entry() :: %{
        "key" => String.t() | atom(),
        "overwriteIfExists" => boolean(),
        "renameTo" => String.t() | atom()
      }
      
  """
  @type rename_key_entry() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      associate_kms_key_request() :: %{
        optional("logGroupName") => String.t() | atom(),
        optional("resourceIdentifier") => String.t() | atom(),
        required("kmsKeyId") => String.t() | atom()
      }
      
  """
  @type associate_kms_key_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_log_record_request() :: %{
        optional("unmask") => boolean(),
        required("logRecordPointer") => String.t() | atom()
      }
      
  """
  @type get_log_record_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_delivery_destination_policy_request() :: %{
        required("deliveryDestinationName") => String.t() | atom()
      }
      
  """
  @type delete_delivery_destination_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      metric_filter_match_record() :: %{
        "eventMessage" => String.t() | atom(),
        "eventNumber" => float(),
        "extractedValues" => map()
      }
      
  """
  @type metric_filter_match_record() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      stop_query_request() :: %{
        required("queryId") => String.t() | atom()
      }
      
  """
  @type stop_query_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_workspace() :: %{
        "status" => open_search_resource_status(),
        "workspaceId" => String.t() | atom()
      }
      
  """
  @type open_search_workspace() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      stop_query_response() :: %{
        "success" => boolean()
      }
      
  """
  @type stop_query_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      create_log_anomaly_detector_response() :: %{
        "anomalyDetectorArn" => String.t() | atom()
      }
      
  """
  @type create_log_anomaly_detector_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_delivery_source_request() :: %{
        optional("tags") => map(),
        required("logType") => String.t() | atom(),
        required("name") => String.t() | atom(),
        required("resourceArn") => String.t() | atom()
      }
      
  """
  @type put_delivery_source_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_field_indexes_response() :: %{
        "fieldIndexes" => list(field_index()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_field_indexes_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      log_group_field() :: %{
        "name" => String.t() | atom(),
        "percent" => integer()
      }
      
  """
  @type log_group_field() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_network_policy() :: %{
        "policyName" => String.t() | atom(),
        "status" => open_search_resource_status()
      }
      
  """
  @type open_search_network_policy() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_transformer_request() :: %{
        required("logGroupIdentifier") => String.t() | atom(),
        required("transformerConfig") => list(processor())
      }
      
  """
  @type put_transformer_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      log_stream() :: %{
        "arn" => String.t() | atom(),
        "creationTime" => float(),
        "firstEventTimestamp" => float(),
        "lastEventTimestamp" => float(),
        "lastIngestionTime" => float(),
        "logStreamName" => String.t() | atom(),
        "storedBytes" => float(),
        "uploadSequenceToken" => String.t() | atom()
      }
      
  """
  @type log_stream() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_tags_for_resource_response() :: %{
        "tags" => map()
      }
      
  """
  @type list_tags_for_resource_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_metric_filter_request() :: %{
        optional("applyOnTransformedLogs") => boolean(),
        optional("emitSystemFieldDimensions") => list(String.t() | atom()),
        optional("fieldSelectionCriteria") => String.t() | atom(),
        required("filterName") => String.t() | atom(),
        required("filterPattern") => String.t() | atom(),
        required("logGroupName") => String.t() | atom(),
        required("metricTransformations") => list(metric_transformation())
      }
      
  """
  @type put_metric_filter_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_integration_request() :: %{
        required("integrationName") => String.t() | atom()
      }
      
  """
  @type get_integration_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_index_policies_response() :: %{
        "indexPolicies" => list(index_policy()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_index_policies_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      filter_log_events_request() :: %{
        optional("endTime") => float(),
        optional("filterPattern") => String.t() | atom(),
        optional("interleaved") => boolean(),
        optional("limit") => integer(),
        optional("logGroupIdentifier") => String.t() | atom(),
        optional("logGroupName") => String.t() | atom(),
        optional("logStreamNamePrefix") => String.t() | atom(),
        optional("logStreamNames") => list(String.t() | atom()),
        optional("nextToken") => String.t() | atom(),
        optional("startTime") => float(),
        optional("unmask") => boolean()
      }
      
  """
  @type filter_log_events_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_deliveries_response() :: %{
        "deliveries" => list(delivery()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_deliveries_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      grok() :: %{
        "match" => String.t() | atom(),
        "source" => String.t() | atom()
      }
      
  """
  @type grok() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      log_group_summary() :: %{
        "logGroupArn" => String.t() | atom(),
        "logGroupClass" => list(any()),
        "logGroupName" => String.t() | atom()
      }
      
  """
  @type log_group_summary() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      service_unavailable_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type service_unavailable_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_destination_request() :: %{
        required("destinationName") => String.t() | atom()
      }
      
  """
  @type delete_destination_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      parse_key_value() :: %{
        "destination" => String.t() | atom(),
        "fieldDelimiter" => String.t() | atom(),
        "keyPrefix" => String.t() | atom(),
        "keyValueDelimiter" => String.t() | atom(),
        "nonMatchValue" => String.t() | atom(),
        "overwriteIfExists" => boolean(),
        "source" => String.t() | atom()
      }
      
  """
  @type parse_key_value() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_account_policy_request() :: %{
        required("policyName") => String.t() | atom(),
        required("policyType") => list(any())
      }
      
  """
  @type delete_account_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_log_groups_response() :: %{
        "logGroups" => list(log_group()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_log_groups_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      upper_case_string() :: %{
        "withKeys" => list(String.t() | atom())
      }
      
  """
  @type upper_case_string() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      untag_log_group_request() :: %{
        required("logGroupName") => String.t() | atom(),
        required("tags") => list(String.t() | atom())
      }
      
  """
  @type untag_log_group_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_subscription_filter_request() :: %{
        optional("applyOnTransformedLogs") => boolean(),
        optional("distribution") => list(any()),
        optional("emitSystemFields") => list(String.t() | atom()),
        optional("fieldSelectionCriteria") => String.t() | atom(),
        optional("roleArn") => String.t() | atom(),
        required("destinationArn") => String.t() | atom(),
        required("filterName") => String.t() | atom(),
        required("filterPattern") => String.t() | atom(),
        required("logGroupName") => String.t() | atom()
      }
      
  """
  @type put_subscription_filter_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_delivery_sources_response() :: %{
        "deliverySources" => list(delivery_source()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_delivery_sources_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      start_live_tail_response() :: %{
        "responseStream" => list()
      }
      
  """
  @type start_live_tail_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_log_groups_for_query_request() :: %{
        optional("maxResults") => integer(),
        optional("nextToken") => String.t() | atom(),
        required("queryId") => String.t() | atom()
      }
      
  """
  @type list_log_groups_for_query_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_lifecycle_policy() :: %{
        "policyName" => String.t() | atom(),
        "status" => open_search_resource_status()
      }
      
  """
  @type open_search_lifecycle_policy() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delivery_destination() :: %{
        "arn" => String.t() | atom(),
        "deliveryDestinationConfiguration" => delivery_destination_configuration(),
        "deliveryDestinationType" => list(any()),
        "name" => String.t() | atom(),
        "outputFormat" => list(any()),
        "tags" => map()
      }
      
  """
  @type delivery_destination() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_to_map() :: %{
        "flatten" => boolean(),
        "flattenedElement" => list(any()),
        "key" => String.t() | atom(),
        "source" => String.t() | atom(),
        "target" => String.t() | atom(),
        "valueKey" => String.t() | atom()
      }
      
  """
  @type list_to_map() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      query_compile_error() :: %{
        "location" => query_compile_error_location(),
        "message" => String.t() | atom()
      }
      
  """
  @type query_compile_error() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_delivery_destination_response() :: %{
        "deliveryDestination" => delivery_destination()
      }
      
  """
  @type get_delivery_destination_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      create_log_anomaly_detector_request() :: %{
        optional("anomalyVisibilityTime") => float(),
        optional("detectorName") => String.t() | atom(),
        optional("evaluationFrequency") => list(any()),
        optional("filterPattern") => String.t() | atom(),
        optional("kmsKeyId") => String.t() | atom(),
        optional("tags") => map(),
        required("logGroupArnList") => list(String.t() | atom())
      }
      
  """
  @type create_log_anomaly_detector_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_delivery_destination_policy_response() :: %{
        "policy" => policy()
      }
      
  """
  @type get_delivery_destination_policy_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      update_delivery_configuration_request() :: %{
        optional("fieldDelimiter") => String.t() | atom(),
        optional("recordFields") => list(String.t() | atom()),
        optional("s3DeliveryConfiguration") => s3_delivery_configuration(),
        required("id") => String.t() | atom()
      }
      
  """
  @type update_delivery_configuration_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_log_group_fields_request() :: %{
        optional("logGroupIdentifier") => String.t() | atom(),
        optional("logGroupName") => String.t() | atom(),
        optional("time") => float()
      }
      
  """
  @type get_log_group_fields_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_query_definition_request() :: %{
        optional("clientToken") => String.t() | atom(),
        optional("logGroupNames") => list(String.t() | atom()),
        optional("queryDefinitionId") => String.t() | atom(),
        optional("queryLanguage") => list(any()),
        required("name") => String.t() | atom(),
        required("queryString") => String.t() | atom()
      }
      
  """
  @type put_query_definition_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      export_task() :: %{
        "destination" => String.t() | atom(),
        "destinationPrefix" => String.t() | atom(),
        "executionInfo" => export_task_execution_info(),
        "from" => float(),
        "logGroupName" => String.t() | atom(),
        "status" => export_task_status(),
        "taskId" => String.t() | atom(),
        "taskName" => String.t() | atom(),
        "to" => float()
      }
      
  """
  @type export_task() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      metric_transformation() :: %{
        "defaultValue" => float(),
        "dimensions" => map(),
        "metricName" => String.t() | atom(),
        "metricNamespace" => String.t() | atom(),
        "metricValue" => String.t() | atom(),
        "unit" => list(any())
      }
      
  """
  @type metric_transformation() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_tags_log_group_request() :: %{
        required("logGroupName") => String.t() | atom()
      }
      
  """
  @type list_tags_log_group_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      resource_policy() :: %{
        "lastUpdatedTime" => float(),
        "policyDocument" => String.t() | atom(),
        "policyName" => String.t() | atom(),
        "policyScope" => list(any()),
        "resourceArn" => String.t() | atom(),
        "revisionId" => String.t() | atom()
      }
      
  """
  @type resource_policy() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      anomaly() :: %{
        "active" => boolean(),
        "anomalyDetectorArn" => String.t() | atom(),
        "anomalyId" => String.t() | atom(),
        "description" => String.t() | atom(),
        "firstSeen" => float(),
        "histogram" => map(),
        "isPatternLevelSuppression" => boolean(),
        "lastSeen" => float(),
        "logGroupArnList" => list(String.t() | atom()),
        "logSamples" => list(log_event()),
        "patternId" => String.t() | atom(),
        "patternRegex" => String.t() | atom(),
        "patternString" => String.t() | atom(),
        "patternTokens" => list(pattern_token()),
        "priority" => String.t() | atom(),
        "state" => list(any()),
        "suppressed" => boolean(),
        "suppressedDate" => float(),
        "suppressedUntil" => float()
      }
      
  """
  @type anomaly() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_anomalies_request() :: %{
        optional("anomalyDetectorArn") => String.t() | atom(),
        optional("limit") => integer(),
        optional("nextToken") => String.t() | atom(),
        optional("suppressionState") => list(any())
      }
      
  """
  @type list_anomalies_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_delivery_destination_response() :: %{
        "deliveryDestination" => delivery_destination()
      }
      
  """
  @type put_delivery_destination_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_delivery_sources_request() :: %{
        optional("limit") => integer(),
        optional("nextToken") => String.t() | atom()
      }
      
  """
  @type describe_delivery_sources_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      rename_keys() :: %{
        "entries" => list(rename_key_entry())
      }
      
  """
  @type rename_keys() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      start_query_response() :: %{
        "queryId" => String.t() | atom()
      }
      
  """
  @type start_query_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_log_object_response() :: %{
        "fieldStream" => list()
      }
      
  """
  @type get_log_object_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      start_live_tail_request() :: %{
        optional("logEventFilterPattern") => String.t() | atom(),
        optional("logStreamNamePrefixes") => list(String.t() | atom()),
        optional("logStreamNames") => list(String.t() | atom()),
        required("logGroupIdentifiers") => list(String.t() | atom())
      }
      
  """
  @type start_live_tail_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      filtered_log_event() :: %{
        "eventId" => String.t() | atom(),
        "ingestionTime" => float(),
        "logStreamName" => String.t() | atom(),
        "message" => String.t() | atom(),
        "timestamp" => float()
      }
      
  """
  @type filtered_log_event() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_data_protection_policy_request() :: %{
        required("logGroupIdentifier") => String.t() | atom()
      }
      
  """
  @type delete_data_protection_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      policy() :: %{
        "deliveryDestinationPolicy" => String.t() | atom()
      }
      
  """
  @type policy() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_resource_policy_request() :: %{
        optional("expectedRevisionId") => String.t() | atom(),
        optional("policyDocument") => String.t() | atom(),
        optional("policyName") => String.t() | atom(),
        optional("resourceArn") => String.t() | atom()
      }
      
  """
  @type put_resource_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_delivery_source_request() :: %{
        required("name") => String.t() | atom()
      }
      
  """
  @type delete_delivery_source_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      access_denied_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type access_denied_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      invalid_parameter_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type invalid_parameter_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_integration_details() :: %{
        "accessPolicy" => open_search_data_access_policy(),
        "application" => open_search_application(),
        "collection" => open_search_collection(),
        "dataSource" => open_search_data_source(),
        "encryptionPolicy" => open_search_encryption_policy(),
        "lifecyclePolicy" => open_search_lifecycle_policy(),
        "networkPolicy" => open_search_network_policy(),
        "workspace" => open_search_workspace()
      }
      
  """
  @type open_search_integration_details() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      move_key_entry() :: %{
        "overwriteIfExists" => boolean(),
        "source" => String.t() | atom(),
        "target" => String.t() | atom()
      }
      
  """
  @type move_key_entry() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_log_events_request() :: %{
        optional("endTime") => float(),
        optional("limit") => integer(),
        optional("logGroupIdentifier") => String.t() | atom(),
        optional("logGroupName") => String.t() | atom(),
        optional("nextToken") => String.t() | atom(),
        optional("startFromHead") => boolean(),
        optional("startTime") => float(),
        optional("unmask") => boolean(),
        required("logStreamName") => String.t() | atom()
      }
      
  """
  @type get_log_events_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      parse_to_o_c_s_f() :: %{
        "eventSource" => list(any()),
        "ocsfVersion" => list(any()),
        "source" => String.t() | atom()
      }
      
  """
  @type parse_to_o_c_s_f() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      add_keys() :: %{
        "entries" => list(add_key_entry())
      }
      
  """
  @type add_keys() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      rejected_log_events_info() :: %{
        "expiredLogEventEndIndex" => integer(),
        "tooNewLogEventStartIndex" => integer(),
        "tooOldLogEventEndIndex" => integer()
      }
      
  """
  @type rejected_log_events_info() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_data_protection_policy_request() :: %{
        required("logGroupIdentifier") => String.t() | atom()
      }
      
  """
  @type get_data_protection_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_log_events_response() :: %{
        "events" => list(output_log_event()),
        "nextBackwardToken" => String.t() | atom(),
        "nextForwardToken" => String.t() | atom()
      }
      
  """
  @type get_log_events_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      output_log_event() :: %{
        "ingestionTime" => float(),
        "message" => String.t() | atom(),
        "timestamp" => float()
      }
      
  """
  @type output_log_event() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      searched_log_stream() :: %{
        "logStreamName" => String.t() | atom(),
        "searchedCompletely" => boolean()
      }
      
  """
  @type searched_log_stream() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_account_policies_request() :: %{
        optional("accountIdentifiers") => list(String.t() | atom()),
        optional("nextToken") => String.t() | atom(),
        optional("policyName") => String.t() | atom(),
        required("policyType") => list(any())
      }
      
  """
  @type describe_account_policies_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      metric_filter() :: %{
        "applyOnTransformedLogs" => boolean(),
        "creationTime" => float(),
        "emitSystemFieldDimensions" => list(String.t() | atom()),
        "fieldSelectionCriteria" => String.t() | atom(),
        "filterName" => String.t() | atom(),
        "filterPattern" => String.t() | atom(),
        "logGroupName" => String.t() | atom(),
        "metricTransformations" => list(metric_transformation())
      }
      
  """
  @type metric_filter() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      validation_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type validation_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_tags_for_resource_request() :: %{
        required("resourceArn") => String.t() | atom()
      }
      
  """
  @type list_tags_for_resource_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      destination() :: %{
        "accessPolicy" => String.t() | atom(),
        "arn" => String.t() | atom(),
        "creationTime" => float(),
        "destinationName" => String.t() | atom(),
        "roleArn" => String.t() | atom(),
        "targetArn" => String.t() | atom()
      }
      
  """
  @type destination() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_log_events_response() :: %{
        "nextSequenceToken" => String.t() | atom(),
        "rejectedEntityInfo" => rejected_entity_info(),
        "rejectedLogEventsInfo" => rejected_log_events_info()
      }
      
  """
  @type put_log_events_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_delivery_source_response() :: %{
        "deliverySource" => delivery_source()
      }
      
  """
  @type put_delivery_source_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_delivery_request() :: %{
        required("id") => String.t() | atom()
      }
      
  """
  @type delete_delivery_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      throttling_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type throttling_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      log_group() :: %{
        "arn" => String.t() | atom(),
        "creationTime" => float(),
        "dataProtectionStatus" => list(any()),
        "inheritedProperties" => list(list(any())()),
        "kmsKeyId" => String.t() | atom(),
        "logGroupArn" => String.t() | atom(),
        "logGroupClass" => list(any()),
        "logGroupName" => String.t() | atom(),
        "metricFilterCount" => integer(),
        "retentionInDays" => integer(),
        "storedBytes" => float()
      }
      
  """
  @type log_group() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_log_anomaly_detectors_request() :: %{
        optional("filterLogGroupArn") => String.t() | atom(),
        optional("limit") => integer(),
        optional("nextToken") => String.t() | atom()
      }
      
  """
  @type list_log_anomaly_detectors_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_log_streams_request() :: %{
        optional("descending") => boolean(),
        optional("limit") => integer(),
        optional("logGroupIdentifier") => String.t() | atom(),
        optional("logGroupName") => String.t() | atom(),
        optional("logStreamNamePrefix") => String.t() | atom(),
        optional("nextToken") => String.t() | atom(),
        optional("orderBy") => list(any())
      }
      
  """
  @type describe_log_streams_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      result_field() :: %{
        "field" => String.t() | atom(),
        "value" => String.t() | atom()
      }
      
  """
  @type result_field() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_resource_policy_response() :: %{
        "resourcePolicy" => resource_policy(),
        "revisionId" => String.t() | atom()
      }
      
  """
  @type put_resource_policy_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_delivery_source_request() :: %{
        required("name") => String.t() | atom()
      }
      
  """
  @type get_delivery_source_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      date_time_converter() :: %{
        "locale" => String.t() | atom(),
        "matchPatterns" => list(String.t() | atom()),
        "source" => String.t() | atom(),
        "sourceTimezone" => String.t() | atom(),
        "target" => String.t() | atom(),
        "targetFormat" => String.t() | atom(),
        "targetTimezone" => String.t() | atom()
      }
      
  """
  @type date_time_converter() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      limit_exceeded_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type limit_exceeded_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_metric_filter_request() :: %{
        required("filterName") => String.t() | atom(),
        required("logGroupName") => String.t() | atom()
      }
      
  """
  @type delete_metric_filter_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_resource_policies_response() :: %{
        "nextToken" => String.t() | atom(),
        "resourcePolicies" => list(resource_policy())
      }
      
  """
  @type describe_resource_policies_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_subscription_filters_response() :: %{
        "nextToken" => String.t() | atom(),
        "subscriptionFilters" => list(subscription_filter())
      }
      
  """
  @type describe_subscription_filters_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      query_statistics() :: %{
        "bytesScanned" => float(),
        "estimatedBytesSkipped" => float(),
        "estimatedRecordsSkipped" => float(),
        "logGroupsScanned" => float(),
        "recordsMatched" => float(),
        "recordsScanned" => float()
      }
      
  """
  @type query_statistics() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_data_protection_policy_response() :: %{
        "lastUpdatedTime" => float(),
        "logGroupIdentifier" => String.t() | atom(),
        "policyDocument" => String.t() | atom()
      }
      
  """
  @type get_data_protection_policy_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_resource_status() :: %{
        "status" => list(any()),
        "statusMessage" => String.t() | atom()
      }
      
  """
  @type open_search_resource_status() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delivery_source() :: %{
        "arn" => String.t() | atom(),
        "logType" => String.t() | atom(),
        "name" => String.t() | atom(),
        "resourceArns" => list(String.t() | atom()),
        "service" => String.t() | atom(),
        "tags" => map()
      }
      
  """
  @type delivery_source() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_resource_policy_request() :: %{
        optional("expectedRevisionId") => String.t() | atom(),
        optional("policyName") => String.t() | atom(),
        optional("resourceArn") => String.t() | atom()
      }
      
  """
  @type delete_resource_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_log_object_request() :: %{
        optional("unmask") => boolean(),
        required("logObjectPointer") => String.t() | atom()
      }
      
  """
  @type get_log_object_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      anomaly_detector() :: %{
        "anomalyDetectorArn" => String.t() | atom(),
        "anomalyDetectorStatus" => list(any()),
        "anomalyVisibilityTime" => float(),
        "creationTimeStamp" => float(),
        "detectorName" => String.t() | atom(),
        "evaluationFrequency" => list(any()),
        "filterPattern" => String.t() | atom(),
        "kmsKeyId" => String.t() | atom(),
        "lastModifiedTimeStamp" => float(),
        "logGroupArnList" => list(String.t() | atom())
      }
      
  """
  @type anomaly_detector() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      test_transformer_request() :: %{
        required("logEventMessages") => list(String.t() | atom()),
        required("transformerConfig") => list(processor())
      }
      
  """
  @type test_transformer_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_log_anomaly_detectors_response() :: %{
        "anomalyDetectors" => list(anomaly_detector()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type list_log_anomaly_detectors_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      parse_route53() :: %{
        "source" => String.t() | atom()
      }
      
  """
  @type parse_route53() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_delivery_destination_request() :: %{
        required("name") => String.t() | atom()
      }
      
  """
  @type delete_delivery_destination_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      internal_streaming_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type internal_streaming_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      cancel_export_task_request() :: %{
        required("taskId") => String.t() | atom()
      }
      
  """
  @type cancel_export_task_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      resource_already_exists_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type resource_already_exists_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_configuration_templates_response() :: %{
        "configurationTemplates" => list(configuration_template()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_configuration_templates_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_subscription_filter_request() :: %{
        required("filterName") => String.t() | atom(),
        required("logGroupName") => String.t() | atom()
      }
      
  """
  @type delete_subscription_filter_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_destinations_request() :: %{
        optional("DestinationNamePrefix") => String.t() | atom(),
        optional("limit") => integer(),
        optional("nextToken") => String.t() | atom()
      }
      
  """
  @type describe_destinations_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      test_metric_filter_response() :: %{
        "matches" => list(metric_filter_match_record())
      }
      
  """
  @type test_metric_filter_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      move_keys() :: %{
        "entries" => list(move_key_entry())
      }
      
  """
  @type move_keys() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      parse_json() :: %{
        "destination" => String.t() | atom(),
        "source" => String.t() | atom()
      }
      
  """
  @type parse_json() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      open_search_application() :: %{
        "applicationArn" => String.t() | atom(),
        "applicationEndpoint" => String.t() | atom(),
        "applicationId" => String.t() | atom(),
        "status" => open_search_resource_status()
      }
      
  """
  @type open_search_application() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      query_definition() :: %{
        "lastModified" => float(),
        "logGroupNames" => list(String.t() | atom()),
        "name" => String.t() | atom(),
        "queryDefinitionId" => String.t() | atom(),
        "queryLanguage" => list(any()),
        "queryString" => String.t() | atom()
      }
      
  """
  @type query_definition() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_keys() :: %{
        "withKeys" => list(String.t() | atom())
      }
      
  """
  @type delete_keys() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      get_delivery_destination_policy_request() :: %{
        required("deliveryDestinationName") => String.t() | atom()
      }
      
  """
  @type get_delivery_destination_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_delivery_destinations_request() :: %{
        optional("limit") => integer(),
        optional("nextToken") => String.t() | atom()
      }
      
  """
  @type describe_delivery_destinations_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      create_delivery_request() :: %{
        optional("fieldDelimiter") => String.t() | atom(),
        optional("recordFields") => list(String.t() | atom()),
        optional("s3DeliveryConfiguration") => s3_delivery_configuration(),
        optional("tags") => map(),
        required("deliveryDestinationArn") => String.t() | atom(),
        required("deliverySourceName") => String.t() | atom()
      }
      
  """
  @type create_delivery_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      record_field() :: %{
        "mandatory" => boolean(),
        "name" => String.t() | atom()
      }
      
  """
  @type record_field() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      delete_log_group_request() :: %{
        required("logGroupName") => String.t() | atom()
      }
      
  """
  @type delete_log_group_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      test_metric_filter_request() :: %{
        required("filterPattern") => String.t() | atom(),
        required("logEventMessages") => list(String.t() | atom())
      }
      
  """
  @type test_metric_filter_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      parse_postgres() :: %{
        "source" => String.t() | atom()
      }
      
  """
  @type parse_postgres() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      query_info() :: %{
        "createTime" => float(),
        "logGroupName" => String.t() | atom(),
        "queryId" => String.t() | atom(),
        "queryLanguage" => list(any()),
        "queryString" => String.t() | atom(),
        "status" => list(any())
      }
      
  """
  @type query_info() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      configuration_template_delivery_config_values() :: %{
        "fieldDelimiter" => String.t() | atom(),
        "recordFields" => list(String.t() | atom()),
        "s3DeliveryConfiguration" => s3_delivery_configuration()
      }
      
  """
  @type configuration_template_delivery_config_values() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      parse_vpc() :: %{
        "source" => String.t() | atom()
      }
      
  """
  @type parse_vpc() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_index_policy_request() :: %{
        required("logGroupIdentifier") => String.t() | atom(),
        required("policyDocument") => String.t() | atom()
      }
      
  """
  @type put_index_policy_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      split_string() :: %{
        "entries" => list(split_string_entry())
      }
      
  """
  @type split_string() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      unrecognized_client_exception() :: %{
        "message" => String.t() | atom()
      }
      
  """
  @type unrecognized_client_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      too_many_tags_exception() :: %{
        "message" => String.t() | atom(),
        "resourceName" => String.t() | atom()
      }
      
  """
  @type too_many_tags_exception() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_configuration_templates_request() :: %{
        optional("deliveryDestinationTypes") => list(list(any())()),
        optional("limit") => integer(),
        optional("logTypes") => list(String.t() | atom()),
        optional("nextToken") => String.t() | atom(),
        optional("resourceTypes") => list(String.t() | atom()),
        optional("service") => String.t() | atom()
      }
      
  """
  @type describe_configuration_templates_request() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      describe_metric_filters_response() :: %{
        "metricFilters" => list(metric_filter()),
        "nextToken" => String.t() | atom()
      }
      
  """
  @type describe_metric_filters_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      index_policy() :: %{
        "lastUpdateTime" => float(),
        "logGroupIdentifier" => String.t() | atom(),
        "policyDocument" => String.t() | atom(),
        "policyName" => String.t() | atom(),
        "source" => list(any())
      }
      
  """
  @type index_policy() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      log_event() :: %{
        "message" => String.t() | atom(),
        "timestamp" => float()
      }
      
  """
  @type log_event() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      list_integrations_response() :: %{
        "integrationSummaries" => list(integration_summary())
      }
      
  """
  @type list_integrations_response() :: %{(String.t() | atom()) => any()}

  @typedoc """

  ## Example:
      
      put_query_definition_response() :: %{
        "queryDefinitionId" => String.t() | atom()
      }
      
  """
  @type put_query_definition_response() :: %{(String.t() | atom()) => any()}

  @type associate_kms_key_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type cancel_export_task_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | invalid_operation_exception()

  @type create_delivery_errors() ::
          throttling_exception()
          | validation_exception()
          | access_denied_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()
          | resource_not_found_exception()
          | conflict_exception()

  @type create_export_task_errors() ::
          resource_already_exists_exception()
          | limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type create_log_anomaly_detector_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type create_log_group_errors() ::
          resource_already_exists_exception()
          | limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | operation_aborted_exception()

  @type create_log_stream_errors() ::
          resource_already_exists_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type delete_account_policy_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_data_protection_policy_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_delivery_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()
          | resource_not_found_exception()
          | conflict_exception()

  @type delete_delivery_destination_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()
          | resource_not_found_exception()
          | conflict_exception()

  @type delete_delivery_destination_policy_errors() ::
          validation_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | conflict_exception()

  @type delete_delivery_source_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()
          | resource_not_found_exception()
          | conflict_exception()

  @type delete_destination_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_index_policy_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_integration_errors() ::
          validation_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type delete_log_anomaly_detector_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_log_group_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_log_stream_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_metric_filter_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_query_definition_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type delete_resource_policy_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_retention_policy_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_subscription_filter_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type delete_transformer_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()
          | invalid_operation_exception()

  @type describe_account_policies_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type describe_configuration_templates_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type describe_deliveries_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()

  @type describe_delivery_destinations_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()

  @type describe_delivery_sources_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()

  @type describe_destinations_errors() ::
          invalid_parameter_exception() | service_unavailable_exception()

  @type describe_export_tasks_errors() ::
          invalid_parameter_exception() | service_unavailable_exception()

  @type describe_field_indexes_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type describe_index_policies_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type describe_log_groups_errors() ::
          invalid_parameter_exception() | service_unavailable_exception()

  @type describe_log_streams_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type describe_metric_filters_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type describe_queries_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type describe_query_definitions_errors() ::
          invalid_parameter_exception() | service_unavailable_exception()

  @type describe_resource_policies_errors() ::
          invalid_parameter_exception() | service_unavailable_exception()

  @type describe_subscription_filters_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type disassociate_kms_key_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type filter_log_events_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type get_data_protection_policy_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type get_delivery_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()
          | resource_not_found_exception()

  @type get_delivery_destination_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()
          | resource_not_found_exception()

  @type get_delivery_destination_policy_errors() ::
          validation_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type get_delivery_source_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()
          | resource_not_found_exception()

  @type get_integration_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type get_log_anomaly_detector_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type get_log_events_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type get_log_group_fields_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type get_log_object_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | access_denied_exception()
          | resource_not_found_exception()
          | invalid_operation_exception()

  @type get_log_record_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type get_query_results_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type get_transformer_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | invalid_operation_exception()

  @type list_anomalies_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type list_integrations_errors() ::
          invalid_parameter_exception() | service_unavailable_exception()

  @type list_log_anomaly_detectors_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type list_log_groups_errors() ::
          invalid_parameter_exception() | service_unavailable_exception()

  @type list_log_groups_for_query_errors() ::
          invalid_parameter_exception()
          | access_denied_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type list_tags_for_resource_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type list_tags_log_group_errors() ::
          service_unavailable_exception() | resource_not_found_exception()

  @type put_account_policy_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | operation_aborted_exception()

  @type put_data_protection_policy_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type put_delivery_destination_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()
          | resource_not_found_exception()
          | conflict_exception()

  @type put_delivery_destination_policy_errors() ::
          validation_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | conflict_exception()

  @type put_delivery_source_errors() ::
          throttling_exception()
          | validation_exception()
          | service_unavailable_exception()
          | service_quota_exceeded_exception()
          | resource_not_found_exception()
          | conflict_exception()

  @type put_destination_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | operation_aborted_exception()

  @type put_destination_policy_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | operation_aborted_exception()

  @type put_index_policy_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type put_integration_errors() ::
          limit_exceeded_exception()
          | validation_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()

  @type put_log_events_errors() ::
          unrecognized_client_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | data_already_accepted_exception()
          | invalid_sequence_token_exception()

  @type put_metric_filter_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()
          | invalid_operation_exception()

  @type put_query_definition_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type put_resource_policy_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type put_retention_policy_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type put_subscription_filter_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()
          | invalid_operation_exception()

  @type put_transformer_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()
          | invalid_operation_exception()

  @type start_live_tail_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | access_denied_exception()
          | resource_not_found_exception()
          | invalid_operation_exception()

  @type start_query_errors() ::
          limit_exceeded_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | malformed_query_exception()

  @type stop_query_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type tag_log_group_errors() :: invalid_parameter_exception() | resource_not_found_exception()

  @type tag_resource_errors() ::
          too_many_tags_exception()
          | invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type test_metric_filter_errors() ::
          invalid_parameter_exception() | service_unavailable_exception()

  @type test_transformer_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | invalid_operation_exception()

  @type untag_log_group_errors() :: resource_not_found_exception()

  @type untag_resource_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()

  @type update_anomaly_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  @type update_delivery_configuration_errors() ::
          throttling_exception()
          | validation_exception()
          | access_denied_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | conflict_exception()

  @type update_log_anomaly_detector_errors() ::
          invalid_parameter_exception()
          | service_unavailable_exception()
          | resource_not_found_exception()
          | operation_aborted_exception()

  def metadata do
    %{
      api_version: "2014-03-28",
      content_type: "application/x-amz-json-1.1",
      credential_scope: nil,
      endpoint_prefix: "logs",
      global?: false,
      hostname: nil,
      protocol: "json",
      service_id: "CloudWatch Logs",
      signature_version: "v4",
      signing_name: "logs",
      target_prefix: "Logs_20140328"
    }
  end

  @doc """
  Associates the specified KMS key with either one log group in the
  account, or with all stored CloudWatch Logs query insights results in the
  account.

  When you use `AssociateKmsKey`, you specify either the
  `logGroupName` parameter or the `resourceIdentifier` parameter. You
  can't specify both of those parameters in the same operation.

    *
  Specify the `logGroupName` parameter to cause log events ingested into that
  log group to be encrypted with that key. Only the log events ingested after the
  key is
  associated are encrypted with that key.

  Associating a KMS key with a log group overrides any existing
  associations between the log group and a KMS key. After a KMS key is associated
  with a log group, all newly ingested data for the log group
  is encrypted using the KMS key. This association is stored as long as the
  data encrypted with the KMS key is still within CloudWatch Logs. This
  enables CloudWatch Logs to decrypt this data whenever it is requested.

  Associating a key with a log group does not cause the results of queries of that
  log
  group to be encrypted with that key. To have query results encrypted with a KMS
  key, you must use an `AssociateKmsKey` operation with the
  `resourceIdentifier` parameter that specifies a `query-result`
  resource.

    *
  Specify the `resourceIdentifier` parameter with a `query-result`
  resource, to use that key to encrypt the stored results of all future
  [StartQuery](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_StartQuery.html) operations in the account. The response from a
  [GetQueryResults](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_GetQueryResults.html)
  operation will still return the query results in plain
  text.

  Even if you have not associated a key with your query results, the query results
  are
  encrypted when stored, using the default CloudWatch Logs method.

  If you run a query from a monitoring account that queries logs in a source
  account,
  the query results key from the monitoring account, if any, is used.

  If you delete the key that is used to encrypt log events or log group query
  results,
  then all the associated stored log events or query results that were encrypted
  with that key
  will be unencryptable and unusable.

  CloudWatch Logs supports only symmetric KMS keys. Do not associate an
  asymmetric KMS key with your log group or query results. For more
  information, see [Using Symmetric and Asymmetric Keys](https://docs.aws.amazon.com/kms/latest/developerguide/symmetric-asymmetric.html).

  It can take up to 5 minutes for this operation to take effect.

  If you attempt to associate a KMS key with a log group but the KMS key does not
  exist or the KMS key is disabled, you receive an
  `InvalidParameterException` error.
  """
  @spec associate_kms_key(map(), associate_kms_key_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, associate_kms_key_errors()}
  def associate_kms_key(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "AssociateKmsKey", input, options)
  end

  @doc """
  Cancels the specified export task.

  The task must be in the `PENDING` or `RUNNING` state.
  """
  @spec cancel_export_task(map(), cancel_export_task_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, cancel_export_task_errors()}
  def cancel_export_task(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "CancelExportTask", input, options)
  end

  @doc """
  Creates a *delivery*.

  A delivery is a connection between a logical
  *delivery source* and a logical *delivery destination*
  that you have already created.

  Only some Amazon Web Services services support being configured as a delivery
  source using
  this operation. These services are listed as **Supported [V2 Permissions]** in the table at [Enabling logging from Amazon Web Services
  services.](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/AWS-logs-and-resource-policy.html)

  A delivery destination can represent a log group in CloudWatch Logs, an Amazon
  S3 bucket, a delivery stream in Firehose, or X-Ray.

  To configure logs delivery between a supported Amazon Web Services service and a
  destination, you must do the following:

    *
  Create a delivery source, which is a logical object that represents the resource
  that
  is actually sending the logs. For more information, see
  [PutDeliverySource](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliverySource.html). 

    *
  Create a *delivery destination*, which is a logical object that
  represents the actual delivery destination. For more information, see
  [PutDeliveryDestination](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliveryDestination.html).

    *
  If you are delivering logs cross-account, you must use
  [PutDeliveryDestinationPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliveryDestinationPolicy.html) in the destination account to assign an IAM policy to the destination. This
  policy allows delivery to that destination.

    *
  Use `CreateDelivery` to create a *delivery* by pairing
  exactly one delivery source and one delivery destination.

  You can configure a single delivery source to send logs to multiple destinations
  by
  creating multiple deliveries. You can also create multiple deliveries to
  configure multiple
  delivery sources to send logs to the same delivery destination.

  To update an existing delivery configuration, use
  [UpdateDeliveryConfiguration](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_UpdateDeliveryConfiguration.html).
  """
  @spec create_delivery(map(), create_delivery_request(), list()) ::
          {:ok, create_delivery_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, create_delivery_errors()}
  def create_delivery(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "CreateDelivery", input, options)
  end

  @doc """
  Creates an export task so that you can efficiently export data from a log group
  to an
  Amazon S3 bucket.

  When you perform a `CreateExportTask` operation, you must use
  credentials that have permission to write to the S3 bucket that you specify as
  the
  destination.

  Exporting log data to S3 buckets that are encrypted by KMS is supported.
  Exporting log data to Amazon S3 buckets that have S3 Object Lock enabled with a
  retention period is also supported.

  Exporting to S3 buckets that are encrypted with AES-256 is supported.

  This is an asynchronous call. If all the required information is provided, this
  operation initiates an export task and responds with the ID of the task. After
  the task has
  started, you can use
  [DescribeExportTasks](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_DescribeExportTasks.html) to get the status of the export task. Each account can only
  have one active (`RUNNING` or `PENDING`) export task at a time. To
  cancel an export task, use
  [CancelExportTask](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_CancelExportTask.html).

  You can export logs from multiple log groups or multiple time ranges to the same
  S3
  bucket. To separate log data for each export task, specify a prefix to be used
  as the Amazon
  S3 key prefix for all exported objects.

  We recommend that you don't regularly export to Amazon S3 as a way to
  continuously archive your logs. For that use case, we instead recommend that you
  use
  subscriptions. For more information about subscriptions, see [Real-time processing of log data
  with
  subscriptions](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/Subscriptions.html).

  Time-based sorting on chunks of log data inside an exported file is not
  guaranteed. You
  can sort the exported log field data by using Linux utilities.
  """
  @spec create_export_task(map(), create_export_task_request(), list()) ::
          {:ok, create_export_task_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, create_export_task_errors()}
  def create_export_task(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "CreateExportTask", input, options)
  end

  @doc """
  Creates an *anomaly detector* that regularly scans one or more log
  groups and look for patterns and anomalies in the logs.

  An anomaly detector can help surface issues by automatically discovering
  anomalies in your
  log event traffic. An anomaly detector uses machine learning algorithms to scan
  log events and
  find *patterns*. A pattern is a shared text structure that recurs among
  your log fields. Patterns provide a useful tool for analyzing large sets of logs
  because a
  large number of log events can often be compressed into a few patterns.

  The anomaly detector uses pattern recognition to find `anomalies`, which are
  unusual log events. It uses the `evaluationFrequency` to compare current log
  events
  and patterns with trained baselines.

  Fields within a pattern are called *tokens*. Fields that vary within a
  pattern, such as a request ID or timestamp, are referred to as *dynamic
  tokens* and represented by `<*>`.

  The following is an example of a pattern:

  `[INFO] Request time: <*> ms` 
  This pattern represents log events like `[INFO] Request time: 327 ms` and other
  similar log events that differ only by the number, in this csse 327. When the
  pattern is
  displayed, the different numbers are replaced by `<*>`

  Any parts of log events that are masked as sensitive data are not scanned for
  anomalies.
  For more information about masking sensitive data, see [Help protect sensitive log
  data with
  masking](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/mask-sensitive-log-data.html).
  """
  @spec create_log_anomaly_detector(map(), create_log_anomaly_detector_request(), list()) ::
          {:ok, create_log_anomaly_detector_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, create_log_anomaly_detector_errors()}
  def create_log_anomaly_detector(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "CreateLogAnomalyDetector", input, options)
  end

  @doc """
  Creates a log group with the specified name.

  You can create up to 1,000,000 log groups
  per Region per account.

  You must use the following guidelines when naming a log group:

    *
  Log group names must be unique within a Region for an Amazon Web Services
  account.

    *
  Log group names can be between 1 and 512 characters long.

    *
  Log group names consist of the following characters: a-z, A-Z, 0-9, '_'
  (underscore), '-' (hyphen), '/' (forward slash), '.' (period), and '#' (number
  sign)

    *
  Log group names can't start with the string `aws/`

  When you create a log group, by default the log events in the log group do not
  expire.
  To set a retention policy so that events expire and are deleted after a
  specified time, use
  [PutRetentionPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutRetentionPolicy.html). 
  If you associate an KMS key with the log group, ingested data is
  encrypted using the KMS key. This association is stored as long as the data
  encrypted with the KMS key is still within CloudWatch Logs. This enables
  CloudWatch Logs to decrypt this data whenever it is requested.

  If you attempt to associate a KMS key with the log group but the KMS key does
  not exist or the KMS key is disabled, you receive an
  `InvalidParameterException` error.

  CloudWatch Logs supports only symmetric KMS keys. Do not associate an
  asymmetric KMS key with your log group. For more information, see [Using
  Symmetric and Asymmetric
  Keys](https://docs.aws.amazon.com/kms/latest/developerguide/symmetric-asymmetric.html).
  """
  @spec create_log_group(map(), create_log_group_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, create_log_group_errors()}
  def create_log_group(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "CreateLogGroup", input, options)
  end

  @doc """
  Creates a log stream for the specified log group.

  A log stream is a sequence of log
  events that originate from a single source, such as an application instance or a
  resource that
  is being monitored.

  There is no limit on the number of log streams that you can create for a log
  group.
  There is a limit of 50 TPS on `CreateLogStream` operations, after which
  transactions are throttled.

  You must use the following guidelines when naming a log stream:

    *
  Log stream names must be unique within the log group.

    *
  Log stream names can be between 1 and 512 characters long.

    *
  Don't use ':' (colon) or '*' (asterisk) characters.
  """
  @spec create_log_stream(map(), create_log_stream_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, create_log_stream_errors()}
  def create_log_stream(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "CreateLogStream", input, options)
  end

  @doc """
  Deletes a CloudWatch Logs account policy.

  This stops the account-wide policy from
  applying to log groups in the account. If you delete a data protection policy or
  subscription
  filter policy, any log-group level policies of those types remain in effect.

  To use this operation, you must be signed on with the correct permissions
  depending on the
  type of policy that you are deleting.

    *
  To delete a data protection policy, you must have the
  `logs:DeleteDataProtectionPolicy` and `logs:DeleteAccountPolicy`
  permissions.

    *
  To delete a subscription filter policy, you must have the
  `logs:DeleteSubscriptionFilter` and `logs:DeleteAccountPolicy`
  permissions.

    *
  To delete a transformer policy, you must have the `logs:DeleteTransformer`
  and `logs:DeleteAccountPolicy` permissions.

    *
  To delete a field index policy, you must have the `logs:DeleteIndexPolicy`
  and `logs:DeleteAccountPolicy` permissions.

  If you delete a field index policy, the indexing of the log events that happened
  before
  you deleted the policy will still be used for up to 30 days to improve
  CloudWatch Logs
  Insights queries.
  """
  @spec delete_account_policy(map(), delete_account_policy_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_account_policy_errors()}
  def delete_account_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteAccountPolicy", input, options)
  end

  @doc """
  Deletes the data protection policy from the specified log group.

  For more information about data protection policies, see
  [PutDataProtectionPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDataProtectionPolicy.html).
  """
  @spec delete_data_protection_policy(map(), delete_data_protection_policy_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_data_protection_policy_errors()}
  def delete_data_protection_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteDataProtectionPolicy", input, options)
  end

  @doc """
  Deletes a *delivery*.

  A delivery is a connection between a logical
  *delivery source* and a logical *delivery
  destination*. Deleting a delivery only deletes the connection between the
  delivery
  source and delivery destination. It does not delete the delivery destination or
  the delivery
  source.
  """
  @spec delete_delivery(map(), delete_delivery_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_delivery_errors()}
  def delete_delivery(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteDelivery", input, options)
  end

  @doc """
  Deletes a *delivery destination*.

  A delivery is a connection between a
  logical *delivery source* and a logical *delivery
  destination*.

  You can't delete a delivery destination if any current deliveries are associated
  with it.
  To find whether any deliveries are associated with this delivery destination,
  use the
  [DescribeDeliveries](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_DescribeDeliveries.html)
  operation and check the `deliveryDestinationArn`
  field in the results.
  """
  @spec delete_delivery_destination(map(), delete_delivery_destination_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_delivery_destination_errors()}
  def delete_delivery_destination(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteDeliveryDestination", input, options)
  end

  @doc """
  Deletes a delivery destination policy.

  For more information about these policies, see
  [PutDeliveryDestinationPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliveryDestinationPolicy.html).
  """
  @spec delete_delivery_destination_policy(
          map(),
          delete_delivery_destination_policy_request(),
          list()
        ) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_delivery_destination_policy_errors()}
  def delete_delivery_destination_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteDeliveryDestinationPolicy", input, options)
  end

  @doc """
  Deletes a *delivery source*.

  A delivery is a connection between a
  logical *delivery source* and a logical *delivery
  destination*.

  You can't delete a delivery source if any current deliveries are associated with
  it. To
  find whether any deliveries are associated with this delivery source, use the
  [DescribeDeliveries](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_DescribeDeliveries.html)
  operation and check the `deliverySourceName` field in
  the results.
  """
  @spec delete_delivery_source(map(), delete_delivery_source_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_delivery_source_errors()}
  def delete_delivery_source(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteDeliverySource", input, options)
  end

  @doc """
  Deletes the specified destination, and eventually disables all the subscription
  filters
  that publish to it.

  This operation does not delete the physical resource encapsulated by the
  destination.
  """
  @spec delete_destination(map(), delete_destination_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_destination_errors()}
  def delete_destination(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteDestination", input, options)
  end

  @doc """
  Deletes a log-group level field index policy that was applied to a single log
  group.

  The
  indexing of the log events that happened before you delete the policy will still
  be used for
  as many as 30 days to improve CloudWatch Logs Insights queries.

  You can't use this operation to delete an account-level index policy. Instead,
  use
  [DeletAccountPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_DeleteAccountPolicy.html).

  If you delete a log-group level field index policy and there is an account-level
  field
  index policy, in a few minutes the log group begins using that account-wide
  policy to index
  new incoming log events.
  """
  @spec delete_index_policy(map(), delete_index_policy_request(), list()) ::
          {:ok, delete_index_policy_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_index_policy_errors()}
  def delete_index_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteIndexPolicy", input, options)
  end

  @doc """
  Deletes the integration between CloudWatch Logs and OpenSearch Service.

  If your
  integration has active vended logs dashboards, you must specify `true` for the
  `force` parameter, otherwise the operation will fail. If you delete the
  integration by setting `force` to `true`, all your vended logs
  dashboards powered by OpenSearch Service will be deleted and the data that was
  on them will no
  longer be accessible.
  """
  @spec delete_integration(map(), delete_integration_request(), list()) ::
          {:ok, delete_integration_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_integration_errors()}
  def delete_integration(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteIntegration", input, options)
  end

  @doc """
  Deletes the specified CloudWatch Logs anomaly detector.
  """
  @spec delete_log_anomaly_detector(map(), delete_log_anomaly_detector_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_log_anomaly_detector_errors()}
  def delete_log_anomaly_detector(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteLogAnomalyDetector", input, options)
  end

  @doc """
  Deletes the specified log group and permanently deletes all the archived log
  events
  associated with the log group.
  """
  @spec delete_log_group(map(), delete_log_group_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_log_group_errors()}
  def delete_log_group(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteLogGroup", input, options)
  end

  @doc """
  Deletes the specified log stream and permanently deletes all the archived log
  events
  associated with the log stream.
  """
  @spec delete_log_stream(map(), delete_log_stream_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_log_stream_errors()}
  def delete_log_stream(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteLogStream", input, options)
  end

  @doc """
  Deletes the specified metric filter.
  """
  @spec delete_metric_filter(map(), delete_metric_filter_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_metric_filter_errors()}
  def delete_metric_filter(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteMetricFilter", input, options)
  end

  @doc """
  Deletes a saved CloudWatch Logs Insights query definition.

  A query definition contains
  details about a saved CloudWatch Logs Insights query.

  Each `DeleteQueryDefinition` operation can delete one query definition.

  You must have the `logs:DeleteQueryDefinition` permission to be able to perform
  this operation.
  """
  @spec delete_query_definition(map(), delete_query_definition_request(), list()) ::
          {:ok, delete_query_definition_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_query_definition_errors()}
  def delete_query_definition(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteQueryDefinition", input, options)
  end

  @doc """
  Deletes a resource policy from this account.

  This revokes the access of the identities
  in that policy to put log events to this account.
  """
  @spec delete_resource_policy(map(), delete_resource_policy_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_resource_policy_errors()}
  def delete_resource_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteResourcePolicy", input, options)
  end

  @doc """
  Deletes the specified retention policy.

  Log events do not expire if they belong to log groups without a retention
  policy.
  """
  @spec delete_retention_policy(map(), delete_retention_policy_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_retention_policy_errors()}
  def delete_retention_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteRetentionPolicy", input, options)
  end

  @doc """
  Deletes the specified subscription filter.
  """
  @spec delete_subscription_filter(map(), delete_subscription_filter_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_subscription_filter_errors()}
  def delete_subscription_filter(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteSubscriptionFilter", input, options)
  end

  @doc """
  Deletes the log transformer for the specified log group.

  As soon as you do this, the
  transformation of incoming log events according to that transformer stops. If
  this account has
  an account-level transformer that applies to this log group, the log group
  begins using that
  account-level transformer when this log-group level transformer is deleted.

  After you delete a transformer, be sure to edit any metric filters or
  subscription filters
  that relied on the transformed versions of the log events.
  """
  @spec delete_transformer(map(), delete_transformer_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, delete_transformer_errors()}
  def delete_transformer(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DeleteTransformer", input, options)
  end

  @doc """
  Returns a list of all CloudWatch Logs account policies in the account.

  To use this operation, you must be signed on with the correct permissions
  depending on the
  type of policy that you are retrieving information for.

    *
  To see data protection policies, you must have the
  `logs:GetDataProtectionPolicy` and `logs:DescribeAccountPolicies`
  permissions.

    *
  To see subscription filter policies, you must have the
  `logs:DescribeSubscriptionFilters` and
  `logs:DescribeAccountPolicies` permissions.

    *
  To see transformer policies, you must have the `logs:GetTransformer` and
  `logs:DescribeAccountPolicies` permissions.

    *
  To see field index policies, you must have the `logs:DescribeIndexPolicies`
  and `logs:DescribeAccountPolicies` permissions.
  """
  @spec describe_account_policies(map(), describe_account_policies_request(), list()) ::
          {:ok, describe_account_policies_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_account_policies_errors()}
  def describe_account_policies(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeAccountPolicies", input, options)
  end

  @doc """
  Use this operation to return the valid and default values that are used when
  creating
  delivery sources, delivery destinations, and deliveries.

  For more information about
  deliveries, see
  [CreateDelivery](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_CreateDelivery.html).
  """
  @spec describe_configuration_templates(
          map(),
          describe_configuration_templates_request(),
          list()
        ) ::
          {:ok, describe_configuration_templates_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_configuration_templates_errors()}
  def describe_configuration_templates(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeConfigurationTemplates", input, options)
  end

  @doc """
  Retrieves a list of the deliveries that have been created in the account.

  A *delivery* is a connection between a [
  *delivery
  source*
  ](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliverySource.html)
  and a [
  *delivery destination*
  ](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliveryDestination.html).

  A delivery source represents an Amazon Web Services resource that sends logs to
  an logs
  delivery destination. The destination can be CloudWatch Logs, Amazon S3,
  Firehose or X-Ray. Only some Amazon Web Services services support being
  configured as a delivery source. These services are listed in [Enable logging from
  Amazon Web Services
  services.](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/AWS-logs-and-resource-policy.html)
  """
  @spec describe_deliveries(map(), describe_deliveries_request(), list()) ::
          {:ok, describe_deliveries_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_deliveries_errors()}
  def describe_deliveries(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeDeliveries", input, options)
  end

  @doc """
  Retrieves a list of the delivery destinations that have been created in the
  account.
  """
  @spec describe_delivery_destinations(map(), describe_delivery_destinations_request(), list()) ::
          {:ok, describe_delivery_destinations_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_delivery_destinations_errors()}
  def describe_delivery_destinations(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeDeliveryDestinations", input, options)
  end

  @doc """
  Retrieves a list of the delivery sources that have been created in the account.
  """
  @spec describe_delivery_sources(map(), describe_delivery_sources_request(), list()) ::
          {:ok, describe_delivery_sources_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_delivery_sources_errors()}
  def describe_delivery_sources(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeDeliverySources", input, options)
  end

  @doc """
  Lists all your destinations.

  The results are ASCII-sorted by destination
  name.
  """
  @spec describe_destinations(map(), describe_destinations_request(), list()) ::
          {:ok, describe_destinations_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_destinations_errors()}
  def describe_destinations(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeDestinations", input, options)
  end

  @doc """
  Lists the specified export tasks.

  You can list all your export tasks or filter the
  results based on task ID or task status.
  """
  @spec describe_export_tasks(map(), describe_export_tasks_request(), list()) ::
          {:ok, describe_export_tasks_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_export_tasks_errors()}
  def describe_export_tasks(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeExportTasks", input, options)
  end

  @doc """
  Returns a list of custom and default field indexes which are discovered in log
  data.

  For
  more information about field index policies, see
  [PutIndexPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutIndexPolicy.html).
  """
  @spec describe_field_indexes(map(), describe_field_indexes_request(), list()) ::
          {:ok, describe_field_indexes_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_field_indexes_errors()}
  def describe_field_indexes(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeFieldIndexes", input, options)
  end

  @doc """
  Returns the field index policies of the specified log group.

  For more information about
  field index policies, see
  [PutIndexPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutIndexPolicy.html).   If a specified log group has a log-group level index policy, that policy is
  returned by
  this operation.

  If a specified log group doesn't have a log-group level index policy, but an
  account-wide
  index policy applies to it, that account-wide policy is returned by this
  operation.

  To find information about only account-level policies, use
  [DescribeAccountPolicies](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_DescribeAccountPolicies.html)
  instead.
  """
  @spec describe_index_policies(map(), describe_index_policies_request(), list()) ::
          {:ok, describe_index_policies_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_index_policies_errors()}
  def describe_index_policies(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeIndexPolicies", input, options)
  end

  @doc """
  Returns information about log groups.

  You can return all your log groups or filter the
  results by prefix. The results are ASCII-sorted by log group name.

  CloudWatch Logs doesn't support IAM policies that control access to the
  `DescribeLogGroups` action by using the

  ```
  aws:ResourceTag/*key-name*

  ```

  condition key. Other CloudWatch
  Logs actions do support the use of the

  ```
  aws:ResourceTag/*key-name*

  ```

  condition key to control access.
  For more information about using tags to control access, see [Controlling access to Amazon Web Services
  resources using
  tags](https://docs.aws.amazon.com/IAM/latest/UserGuide/access_tags.html).

  If you are using CloudWatch cross-account observability, you can use this
  operation
  in a monitoring account and view data from the linked source accounts. For more
  information,
  see [CloudWatch cross-account observability](https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/CloudWatch-Unified-Cross-Account.html).
  """
  @spec describe_log_groups(map(), describe_log_groups_request(), list()) ::
          {:ok, describe_log_groups_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_log_groups_errors()}
  def describe_log_groups(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeLogGroups", input, options)
  end

  @doc """
  Lists the log streams for the specified log group.

  You can list all the log streams or
  filter the results by prefix. You can also control how the results are ordered.

  You can specify the log group to search by using either `logGroupIdentifier` or
  `logGroupName`. You must include one of these two parameters, but you can't
  include both.

  This operation has a limit of 25 transactions per second, after which
  transactions are
  throttled.

  If you are using CloudWatch cross-account observability, you can use this
  operation
  in a monitoring account and view data from the linked source accounts. For more
  information,
  see [CloudWatch cross-account observability](https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/CloudWatch-Unified-Cross-Account.html).
  """
  @spec describe_log_streams(map(), describe_log_streams_request(), list()) ::
          {:ok, describe_log_streams_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_log_streams_errors()}
  def describe_log_streams(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeLogStreams", input, options)
  end

  @doc """
  Lists the specified metric filters.

  You can list all of the metric filters or filter
  the results by log name, prefix, metric name, or metric namespace. The results
  are
  ASCII-sorted by filter name.
  """
  @spec describe_metric_filters(map(), describe_metric_filters_request(), list()) ::
          {:ok, describe_metric_filters_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_metric_filters_errors()}
  def describe_metric_filters(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeMetricFilters", input, options)
  end

  @doc """
  Returns a list of CloudWatch Logs Insights queries that are scheduled, running,
  or have
  been run recently in this account.

  You can request all queries or limit it to queries of a
  specific log group or queries with a certain status.
  """
  @spec describe_queries(map(), describe_queries_request(), list()) ::
          {:ok, describe_queries_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_queries_errors()}
  def describe_queries(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeQueries", input, options)
  end

  @doc """
  This operation returns a paginated list of your saved CloudWatch Logs Insights
  query
  definitions.

  You can retrieve query definitions from the current account or from a source
  account that is linked to the current account.

  You can use the `queryDefinitionNamePrefix` parameter to limit the results to
  only the query definitions that have names that start with a certain string.
  """
  @spec describe_query_definitions(map(), describe_query_definitions_request(), list()) ::
          {:ok, describe_query_definitions_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_query_definitions_errors()}
  def describe_query_definitions(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeQueryDefinitions", input, options)
  end

  @doc """
  Lists the resource policies in this account.
  """
  @spec describe_resource_policies(map(), describe_resource_policies_request(), list()) ::
          {:ok, describe_resource_policies_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_resource_policies_errors()}
  def describe_resource_policies(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeResourcePolicies", input, options)
  end

  @doc """
  Lists the subscription filters for the specified log group.

  You can list all the
  subscription filters or filter the results by prefix. The results are
  ASCII-sorted by filter
  name.
  """
  @spec describe_subscription_filters(map(), describe_subscription_filters_request(), list()) ::
          {:ok, describe_subscription_filters_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, describe_subscription_filters_errors()}
  def describe_subscription_filters(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DescribeSubscriptionFilters", input, options)
  end

  @doc """
  Disassociates the specified KMS key from the specified log group or
  from all CloudWatch Logs Insights query results in the account.

  When you use `DisassociateKmsKey`, you specify either the
  `logGroupName` parameter or the `resourceIdentifier` parameter. You
  can't specify both of those parameters in the same operation.

    *
  Specify the `logGroupName` parameter to stop using the KMS key to encrypt future
  log events ingested and stored in the log group.
  Instead, they will be encrypted with the default CloudWatch Logs method. The log
  events
  that were ingested while the key was associated with the log group are still
  encrypted
  with that key. Therefore, CloudWatch Logs will need permissions for the key
  whenever
  that data is accessed.

    *
  Specify the `resourceIdentifier` parameter with the
  `query-result` resource to stop using the KMS key to
  encrypt the results of all future
  [StartQuery](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_StartQuery.html)
  operations in the account. They will instead be encrypted with the default
  CloudWatch Logs method. The results from queries that ran while the key was
  associated with
  the account are still encrypted with that key. Therefore, CloudWatch Logs will
  need
  permissions for the key whenever that data is accessed.

  It can take up to 5 minutes for this operation to take effect.
  """
  @spec disassociate_kms_key(map(), disassociate_kms_key_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, disassociate_kms_key_errors()}
  def disassociate_kms_key(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "DisassociateKmsKey", input, options)
  end

  @doc """
  Lists log events from the specified log group.

  You can list all the log events or
  filter the results using one or more of the following:

    *
  A filter pattern

    *
  A time range

    *
  The log stream name, or a log stream name prefix that matches multiple log
  streams

  You must have the `logs:FilterLogEvents` permission to perform this
  operation.

  You can specify the log group to search by using either `logGroupIdentifier` or
  `logGroupName`. You must include one of these two parameters, but you can't
  include both.

  `FilterLogEvents` is a paginated operation. Each page returned can contain up
  to 1 MB of log events or up to 10,000 log events. A returned page might only be
  partially
  full, or even empty. For example, if the result of a query would return 15,000
  log events, the
  first page isn't guaranteed to have 10,000 log events even if they all fit into
  1 MB.

  Partially full or empty pages don't necessarily mean that pagination is
  finished. If the
  results include a `nextToken`, there might be more log events available. You can
  return these additional log events by providing the nextToken in a subsequent
  `FilterLogEvents` operation. If the results don't include a
  `nextToken`, then pagination is finished.

  Specifying the `limit` parameter only guarantees that a single page doesn't
  return more log events than the specified limit, but it might return fewer
  events than the
  limit. This is the expected API behavior.

  The returned log events are sorted by event timestamp, the timestamp when the
  event was
  ingested by CloudWatch Logs, and the ID of the `PutLogEvents` request.

  If you are using CloudWatch cross-account observability, you can use this
  operation
  in a monitoring account and view data from the linked source accounts. For more
  information,
  see [CloudWatch cross-account observability](https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/CloudWatch-Unified-Cross-Account.html).

  If you are using [log transformation](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatch-Logs-Transformation.html),
  the `FilterLogEvents` operation returns only the
  original versions of log events, before they were transformed. To view the
  transformed
  versions, you must use a [CloudWatch Logs query.](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/AnalyzingLogData.html)
  """
  @spec filter_log_events(map(), filter_log_events_request(), list()) ::
          {:ok, filter_log_events_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, filter_log_events_errors()}
  def filter_log_events(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "FilterLogEvents", input, options)
  end

  @doc """
  Returns information about a log group data protection policy.
  """
  @spec get_data_protection_policy(map(), get_data_protection_policy_request(), list()) ::
          {:ok, get_data_protection_policy_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_data_protection_policy_errors()}
  def get_data_protection_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetDataProtectionPolicy", input, options)
  end

  @doc """
  Returns complete information about one logical *delivery*.

  A delivery
  is a connection between a [
  *delivery
  source*
  ](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliverySource.html)
  and a [
  *delivery destination*
  ](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliveryDestination.html).

  A delivery source represents an Amazon Web Services resource that sends logs to
  an logs
  delivery destination. The destination can be CloudWatch Logs, Amazon S3, or
  Firehose. Only some Amazon Web Services services support being configured as a
  delivery
  source. These services are listed in [Enable logging from Amazon Web Services
  services.](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/AWS-logs-and-resource-policy.html)

  You need to specify the delivery `id` in this operation. You can find the IDs
  of the deliveries in your account with the
  [DescribeDeliveries](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_DescribeDeliveries.html)
  operation.
  """
  @spec get_delivery(map(), get_delivery_request(), list()) ::
          {:ok, get_delivery_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_delivery_errors()}
  def get_delivery(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetDelivery", input, options)
  end

  @doc """
  Retrieves complete information about one delivery destination.
  """
  @spec get_delivery_destination(map(), get_delivery_destination_request(), list()) ::
          {:ok, get_delivery_destination_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_delivery_destination_errors()}
  def get_delivery_destination(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetDeliveryDestination", input, options)
  end

  @doc """
  Retrieves the delivery destination policy assigned to the delivery destination
  that you
  specify.

  For more information about delivery destinations and their policies, see
  [PutDeliveryDestinationPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliveryDestinationPolicy.html).
  """
  @spec get_delivery_destination_policy(map(), get_delivery_destination_policy_request(), list()) ::
          {:ok, get_delivery_destination_policy_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_delivery_destination_policy_errors()}
  def get_delivery_destination_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetDeliveryDestinationPolicy", input, options)
  end

  @doc """
  Retrieves complete information about one delivery source.
  """
  @spec get_delivery_source(map(), get_delivery_source_request(), list()) ::
          {:ok, get_delivery_source_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_delivery_source_errors()}
  def get_delivery_source(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetDeliverySource", input, options)
  end

  @doc """
  Returns information about one integration between CloudWatch Logs and OpenSearch
  Service.
  """
  @spec get_integration(map(), get_integration_request(), list()) ::
          {:ok, get_integration_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_integration_errors()}
  def get_integration(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetIntegration", input, options)
  end

  @doc """
  Retrieves information about the log anomaly detector that you specify.

  The KMS key ARN detected is valid.
  """
  @spec get_log_anomaly_detector(map(), get_log_anomaly_detector_request(), list()) ::
          {:ok, get_log_anomaly_detector_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_log_anomaly_detector_errors()}
  def get_log_anomaly_detector(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetLogAnomalyDetector", input, options)
  end

  @doc """
  Lists log events from the specified log stream.

  You can list all of the log events or
  filter using a time range.

  `GetLogEvents` is a paginated operation. Each page returned can contain up to 1
  MB of log events or up to 10,000 log events. A returned page might only be
  partially full, or
  even empty. For example, if the result of a query would return 15,000 log
  events, the first
  page isn't guaranteed to have 10,000 log events even if they all fit into 1 MB.

  Partially full or empty pages don't necessarily mean that pagination is
  finished. As long
  as the `nextBackwardToken` or `nextForwardToken` returned is NOT equal
  to the `nextToken` that you passed into the API call, there might be more log
  events available. The token that you use depends on the direction you want to
  move in along
  the log stream. The returned tokens are never null.

  If you set `startFromHead` to `true` and you don’t include
  `endTime` in your request, you can end up in a situation where the pagination
  doesn't terminate. This can happen when the new log events are being added to
  the target log
  streams faster than they are being read. This situation is a good use case for
  the CloudWatch Logs
  [Live Tail](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatchLogs_LiveTail.html)
  feature.

  If you are using CloudWatch cross-account observability, you can use this
  operation
  in a monitoring account and view data from the linked source accounts. For more
  information,
  see [CloudWatch cross-account observability](https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/CloudWatch-Unified-Cross-Account.html).

  You can specify the log group to search by using either `logGroupIdentifier` or
  `logGroupName`. You must include one of these two parameters, but you can't
  include both.

  If you are using [log transformation](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatch-Logs-Transformation.html),
  the `GetLogEvents` operation returns only the original
  versions of log events, before they were transformed. To view the transformed
  versions, you
  must use a [CloudWatch Logs query.](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/AnalyzingLogData.html)
  """
  @spec get_log_events(map(), get_log_events_request(), list()) ::
          {:ok, get_log_events_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_log_events_errors()}
  def get_log_events(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetLogEvents", input, options)
  end

  @doc """
  Returns a list of the fields that are included in log events in the specified
  log group.

  Includes the percentage of log events that contain each field. The search is
  limited to a time
  period that you specify.

  You can specify the log group to search by using either `logGroupIdentifier` or
  `logGroupName`. You must specify one of these parameters, but you can't specify
  both.

  In the results, fields that start with `@` are fields generated by CloudWatch
  Logs. For example, `@timestamp` is the timestamp of each log event. For more
  information about the fields that are generated by CloudWatch logs, see
  [Supported Logs and Discovered
  Fields](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CWL_AnalyzeLogData-discoverable-fields.html).

  The response results are sorted by the frequency percentage, starting with the
  highest
  percentage.

  If you are using CloudWatch cross-account observability, you can use this
  operation
  in a monitoring account and view data from the linked source accounts. For more
  information,
  see [CloudWatch cross-account observability](https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/CloudWatch-Unified-Cross-Account.html).
  """
  @spec get_log_group_fields(map(), get_log_group_fields_request(), list()) ::
          {:ok, get_log_group_fields_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_log_group_fields_errors()}
  def get_log_group_fields(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetLogGroupFields", input, options)
  end

  @doc """
  Retrieves a large logging object (LLO) and streams it back.

  This API is used to fetch the
  content of large portions of log events that have been ingested through the
  PutOpenTelemetryLogs API. When log events contain fields that would cause the
  total event size
  to exceed 1MB, CloudWatch Logs automatically processes up to 10 fields, starting
  with the
  largest fields. Each field is truncated as needed to keep the total event size
  as close to 1MB
  as possible. The excess portions are stored as Large Log Objects (LLOs) and
  these fields are
  processed separately and LLO reference system fields (in the format
  `@ptr.$[path.to.field]`) are added. The path in the reference field reflects the original JSON structure where the large field was located. For example, this
  could be
  `@ptr.$['input']['message']`, `@ptr.$['AAA']['BBB']['CCC']['DDD']`, `@ptr.$['AAA']`, or any other path matching your log structure.
  """
  @spec get_log_object(map(), get_log_object_request(), list()) ::
          {:ok, get_log_object_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_log_object_errors()}
  def get_log_object(%Client{} = client, input, options \\ []) do
    meta = metadata() |> Map.put_new(:host_prefix, "streaming-")

    Request.request_post(client, meta, "GetLogObject", input, options)
  end

  @doc """
  Retrieves all of the fields and values of a single log event.

  All fields are retrieved,
  even if the original query that produced the `logRecordPointer` retrieved only a
  subset of fields. Fields are returned as field name/field value pairs.

  The full unparsed log event is returned within `@message`.
  """
  @spec get_log_record(map(), get_log_record_request(), list()) ::
          {:ok, get_log_record_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_log_record_errors()}
  def get_log_record(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetLogRecord", input, options)
  end

  @doc """
  Returns the results from the specified query.

  Only the fields requested in the query are returned, along with a `@ptr` field,
  which is the identifier for the log record. You can use the value of `@ptr` in a
  [GetLogRecord](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_GetLogRecord.html)  operation to get the full log record.

  `GetQueryResults` does not start running a query. To run a query, use
  [StartQuery](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_StartQuery.html).
  For more information about how long results of previous queries are
  available, see [CloudWatch Logs quotas](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/cloudwatch_limits_cwl.html).

  If the value of the `Status` field in the output is `Running`, this
  operation returns only partial results. If you see a value of `Scheduled` or
  `Running` for the status, you can retry the operation later to see the final
  results.

  If you are using CloudWatch cross-account observability, you can use this
  operation
  in a monitoring account to start queries in linked source accounts. For more
  information, see
  [CloudWatch cross-account observability](https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/CloudWatch-Unified-Cross-Account.html).
  """
  @spec get_query_results(map(), get_query_results_request(), list()) ::
          {:ok, get_query_results_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_query_results_errors()}
  def get_query_results(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetQueryResults", input, options)
  end

  @doc """
  Returns the information about the log transformer associated with this log
  group.

  This operation returns data only for transformers created at the log group
  level. To get
  information for an account-level transformer, use
  [DescribeAccountPolicies](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_DescribeAccountPolicies.html).
  """
  @spec get_transformer(map(), get_transformer_request(), list()) ::
          {:ok, get_transformer_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, get_transformer_errors()}
  def get_transformer(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "GetTransformer", input, options)
  end

  @doc """
  Returns a list of anomalies that log anomaly detectors have found.

  For details about the
  structure format of each anomaly object that is returned, see the example in
  this
  section.
  """
  @spec list_anomalies(map(), list_anomalies_request(), list()) ::
          {:ok, list_anomalies_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, list_anomalies_errors()}
  def list_anomalies(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "ListAnomalies", input, options)
  end

  @doc """
  Returns a list of integrations between CloudWatch Logs and other services in
  this
  account.

  Currently, only one integration can be created in an account, and this
  integration
  must be with OpenSearch Service.
  """
  @spec list_integrations(map(), list_integrations_request(), list()) ::
          {:ok, list_integrations_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, list_integrations_errors()}
  def list_integrations(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "ListIntegrations", input, options)
  end

  @doc """
  Retrieves a list of the log anomaly detectors in the account.
  """
  @spec list_log_anomaly_detectors(map(), list_log_anomaly_detectors_request(), list()) ::
          {:ok, list_log_anomaly_detectors_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, list_log_anomaly_detectors_errors()}
  def list_log_anomaly_detectors(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "ListLogAnomalyDetectors", input, options)
  end

  @doc """
  Returns a list of log groups in the Region in your account.

  If you are performing this
  action in a monitoring account, you can choose to also return log groups from
  source accounts
  that are linked to the monitoring account. For more information about using
  cross-account
  observability to set up monitoring accounts and source accounts, see [
  CloudWatch cross-account
  observability](https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/CloudWatch-Unified-Cross-Account.html).

  You can optionally filter the list by log group class and by using regular
  expressions in
  your request to match strings in the log group names.

  This operation is paginated. By default, your first use of this operation
  returns 50
  results, and includes a token to use in a subsequent operation to return more
  results.
  """
  @spec list_log_groups(map(), list_log_groups_request(), list()) ::
          {:ok, list_log_groups_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, list_log_groups_errors()}
  def list_log_groups(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "ListLogGroups", input, options)
  end

  @doc """
  Returns a list of the log groups that were analyzed during a single CloudWatch
  Logs
  Insights query.

  This can be useful for queries that use log group name prefixes or the
  `filterIndex` command, because the log groups are dynamically selected in these
  cases.

  For more information about field indexes, see [Create field indexes to improve query performance and reduce
  costs](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatchLogs-Field-Indexing.html).
  """
  @spec list_log_groups_for_query(map(), list_log_groups_for_query_request(), list()) ::
          {:ok, list_log_groups_for_query_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, list_log_groups_for_query_errors()}
  def list_log_groups_for_query(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "ListLogGroupsForQuery", input, options)
  end

  @doc """
  Displays the tags associated with a CloudWatch Logs resource.

  Currently, log groups and
  destinations support tagging.
  """
  @spec list_tags_for_resource(map(), list_tags_for_resource_request(), list()) ::
          {:ok, list_tags_for_resource_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, list_tags_for_resource_errors()}
  def list_tags_for_resource(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "ListTagsForResource", input, options)
  end

  @doc """

  The ListTagsLogGroup operation is on the path to deprecation.

  We recommend that you use
  [ListTagsForResource](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_ListTagsForResource.html)
  instead.

  Lists the tags for the specified log group.
  """
  @spec list_tags_log_group(map(), list_tags_log_group_request(), list()) ::
          {:ok, list_tags_log_group_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, list_tags_log_group_errors()}
  def list_tags_log_group(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "ListTagsLogGroup", input, options)
  end

  @doc """
  Creates an account-level data protection policy, subscription filter policy,
  field index
  policy, transformer policy, or metric extraction policy that applies to all log
  groups or a
  subset of log groups in the account.

  To use this operation, you must be signed on with the correct permissions
  depending on the
  type of policy that you are creating.

    *
  To create a data protection policy, you must have the
  `logs:PutDataProtectionPolicy` and `logs:PutAccountPolicy`
  permissions.

    *
  To create a subscription filter policy, you must have the
  `logs:PutSubscriptionFilter` and `logs:PutAccountPolicy`
  permissions.

    *
  To create a transformer policy, you must have the `logs:PutTransformer` and
  `logs:PutAccountPolicy` permissions.

    *
  To create a field index policy, you must have the `logs:PutIndexPolicy` and
  `logs:PutAccountPolicy` permissions.

    *
  To create a metric extraction policy, you must have the
  `logs:PutMetricExtractionPolicy` and `logs:PutAccountPolicy`
  permissions.

  ## Data protection policy

  A data protection policy can help safeguard sensitive data that's ingested by
  your log
  groups by auditing and masking the sensitive log data. Each account can have
  only one
  account-level data protection policy.

  Sensitive data is detected and masked when it is ingested into a log group. When
  you set
  a data protection policy, log events ingested into the log groups before that
  time are not
  masked.

  If you use `PutAccountPolicy` to create a data protection policy for your whole
  account, it applies to both existing log groups and all log groups that are
  created later in
  this account. The account-level policy is applied to existing log groups with
  eventual
  consistency. It might take up to 5 minutes before sensitive data in existing log
  groups begins
  to be masked.

  By default, when a user views a log event that includes masked data, the
  sensitive data is
  replaced by asterisks. A user who has the `logs:Unmask` permission can use a
  [GetLogEvents](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_GetLogEvents.html) or
  [FilterLogEvents](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_FilterLogEvents.html)
  operation with the `unmask` parameter set to
  `true` to view the unmasked log events. Users with the `logs:Unmask`
  can also view unmasked data in the CloudWatch Logs console by running a
  CloudWatch Logs
  Insights query with the `unmask` query command.

  For more information, including a list of types of data that can be audited and
  masked,
  see [Protect sensitive log data with
  masking](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/mask-sensitive-log-data.html).

  To use the `PutAccountPolicy` operation for a data protection policy, you must
  be signed on with the `logs:PutDataProtectionPolicy` and
  `logs:PutAccountPolicy` permissions.

  The `PutAccountPolicy` operation applies to all log groups in the account. You
  can use
  [PutDataProtectionPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDataProtectionPolicy.html) to create a data protection policy that applies to just one
  log group. If a log group has its own data protection policy and the account
  also has an
  account-level data protection policy, then the two policies are cumulative. Any
  sensitive term
  specified in either policy is masked.

  ## Subscription filter policy

  A subscription filter policy sets up a real-time feed of log events from
  CloudWatch Logs to other Amazon Web Services services. Account-level
  subscription filter policies apply to
  both existing log groups and log groups that are created later in this account.
  Supported
  destinations are Kinesis Data Streams, Firehose, and Lambda. When log
  events are sent to the receiving service, they are Base64 encoded and compressed
  with the GZIP
  format.

  The following destinations are supported for subscription filters:

    *
  An Kinesis Data Streams data stream in the same account as the subscription
  policy, for
  same-account delivery.

    *
  An Firehose data stream in the same account as the subscription policy, for
  same-account delivery.

    *
  A Lambda function in the same account as the subscription policy, for
  same-account delivery.

    *
  A logical destination in a different account created with
  [PutDestination](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDestination.html),
  for cross-account delivery. Kinesis Data Streams and Firehose are supported as
  logical destinations.

  Each account can have one account-level subscription filter policy per Region.
  If you are
  updating an existing filter, you must specify the correct name in `PolicyName`.
  To
  perform a `PutAccountPolicy` subscription filter operation for any destination
  except a Lambda function, you must also have the `iam:PassRole`
  permission.

  ## Transformer policy

  Creates or updates a *log transformer policy* for your account. You use
  log transformers to transform log events into a different format, making them
  easier for you
  to process and analyze. You can also transform logs from different sources into
  standardized
  formats that contain relevant, source-specific information. After you have
  created a
  transformer, CloudWatch Logs performs this transformation at the time of log
  ingestion. You
  can then refer to the transformed versions of the logs during operations such as
  querying with
  CloudWatch Logs Insights or creating metric filters or subscription filters.

  You can also use a transformer to copy metadata from metadata keys into the log
  events
  themselves. This metadata can include log group name, log stream name, account
  ID and
  Region.

  A transformer for a log group is a series of processors, where each processor
  applies one
  type of transformation to the log events ingested into this log group. For more
  information
  about the available processors to use in a transformer, see [ Processors that you can
  use](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatch-Logs-Transformation.html#CloudWatch-Logs-Transformation-Processors).

  Having log events in standardized format enables visibility across your
  applications for
  your log analysis, reporting, and alarming needs. CloudWatch Logs provides
  transformation
  for common log types with out-of-the-box transformation templates for major
  Amazon Web Services
  log sources such as VPC flow logs, Lambda, and Amazon RDS. You can use
  pre-built transformation templates or create custom transformation policies.

  You can create transformers only for the log groups in the Standard log class.

  You can have one account-level transformer policy that applies to all log groups
  in the
  account. Or you can create as many as 20 account-level transformer policies that
  are each
  scoped to a subset of log groups with the `selectionCriteria` parameter. If you
  have multiple account-level transformer policies with selection criteria, no two
  of them can
  use the same or overlapping log group name prefixes. For example, if you have
  one policy
  filtered to log groups that start with `my-log`, you can't have another field
  index
  policy filtered to `my-logpprod` or `my-logging`.

  CloudWatch Logs provides default field indexes for all log groups in the
  Standard log
  class. Default field indexes are automatically available for the following
  fields:

    *

  `@aws.region`

    *

  `@aws.account`

    *

  `@source.log`

    *

  `traceId`

  Default field indexes are in addition to any custom field indexes you define
  within your
  policy. Default field indexes are not counted towards your field index quota.

  You can also set up a transformer at the log-group level. For more information,
  see
  [PutTransformer](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutTransformer.html). If there is both a log-group level transformer created with
  `PutTransformer` and an account-level transformer that could apply to the same
  log group, the log group uses only the log-group level transformer. It ignores
  the
  account-level transformer.

  ## Field index policy

  You can use field index policies to create indexes on fields found in log events
  in the
  log group. Creating field indexes can help lower the scan volume for CloudWatch
  Logs
  Insights queries that reference those fields, because these queries attempt to
  skip the
  processing of log events that are known to not match the indexed field. Good
  fields to index
  are fields that you often need to query for and fields or values that match only
  a small
  fraction of the total log events. Common examples of indexes include request ID,
  session ID,
  user IDs, or instance IDs. For more information, see [Create field indexes
  to improve query performance and reduce
  costs](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatchLogs-Field-Indexing.html)

  To find the fields that are in your log group events, use the
  [GetLogGroupFields](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_GetLogGroupFields.html) operation.

  For example, suppose you have created a field index for `requestId`. Then, any
  CloudWatch Logs Insights query on that log group that includes

  ```
  requestId =
  *value*

  ```

  or

  ```
  requestId in [*value*,
  *value*, ...]
  ```

  will attempt to process only the log events where
  the indexed field matches the specified value.

  Matches of log events to the names of indexed fields are case-sensitive. For
  example, an
  indexed field of `RequestId` won't match a log event containing
  `requestId`.

  You can have one account-level field index policy that applies to all log groups
  in the
  account. Or you can create as many as 20 account-level field index policies that
  are each
  scoped to a subset of log groups with the `selectionCriteria` parameter. If you
  have multiple account-level index policies with selection criteria, no two of
  them can use the
  same or overlapping log group name prefixes. For example, if you have one policy
  filtered to
  log groups that start with `my-log`, you can't have another field index policy
  filtered to `my-logpprod` or `my-logging`.

  If you create an account-level field index policy in a monitoring account in
  cross-account
  observability, the policy is applied only to the monitoring account and not to
  any source
  accounts.

  If you want to create a field index policy for a single log group, you can use
  [PutIndexPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutIndexPolicy.html) instead of `PutAccountPolicy`. If you do so, that log
  group will use only that log-group level policy, and will ignore the
  account-level policy that
  you create with
  [PutAccountPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutAccountPolicy.html).

  ## Metric extraction policy

  A metric extraction policy controls whether CloudWatch Metrics can be created
  through the
  Embedded Metrics Format (EMF) for log groups in your account. By default, EMF
  metric creation
  is enabled for all log groups. You can use metric extraction policies to disable
  EMF metric
  creation for your entire account or specific log groups.

  When a policy disables EMF metric creation for a log group, log events in the
  EMF format
  are still ingested, but no CloudWatch Metrics are created from them.

  Creating a policy disables metrics for AWS features that use EMF to create
  metrics, such
  as CloudWatch Container Insights and CloudWatch Application Signals. To prevent
  turning off
  those features by accident, we recommend that you exclude the underlying
  log-groups through
  a selection-criteria such as

  ```
  LogGroupNamePrefix NOT IN ["/aws/containerinsights", "/aws/ecs/containerinsights", "/aws/application-signals/data"]
  ```

  .

  Each account can have either one account-level metric extraction policy that
  applies to
  all log groups, or up to 5 policies that are each scoped to a subset of log
  groups with the
  `selectionCriteria` parameter. The selection criteria supports filtering by
  `LogGroupName` and `LogGroupNamePrefix` using the operators
  `IN` and `NOT IN`. You can specify up to 50 values in each
  `IN` or `NOT IN` list.

  The selection criteria can be specified in these formats:

  `LogGroupName IN ["log-group-1", "log-group-2"]` 

  `LogGroupNamePrefix NOT IN ["/aws/prefix1", "/aws/prefix2"]`

  If you have multiple account-level metric extraction policies with selection
  criteria, no
  two of them can have overlapping criteria. For example, if you have one policy
  with selection
  criteria `LogGroupNamePrefix IN ["my-log"]`, you can't have another metric extraction policy with selection criteria `LogGroupNamePrefix IN
  ["/my-log-prod"]`
  or `LogGroupNamePrefix IN ["/my-logging"]`, as the set of log groups matching
  these
  prefixes would be a subset of the log groups matching the first policy's prefix,
  creating an
  overlap.

  When using `NOT IN`, only one policy with this operator is allowed per
  account.

  When combining policies with `IN` and `NOT IN` operators, the
  overlap check ensures that policies don't have conflicting effects. Two policies
  with
  `IN` and `NOT IN` operators do not overlap if and only if every value
  in the `IN `policy is completely contained within some value in the

  ```
  NOT
  IN
  ```

  policy. For example:

    *
  If you have a `NOT IN` policy for prefix `"/aws/lambda"`, you
  can create an `IN` policy for the exact log group name
  `"/aws/lambda/function1"` because the set of log groups matching
  `"/aws/lambda/function1"` is a subset of the log groups matching
  `"/aws/lambda"`.

    *
  If you have a `NOT IN` policy for prefix `"/aws/lambda"`, you
  cannot create an `IN` policy for prefix `"/aws"` because the set of
  log groups matching `"/aws"` is not a subset of the log groups matching
  `"/aws/lambda"`.
  """
  @spec put_account_policy(map(), put_account_policy_request(), list()) ::
          {:ok, put_account_policy_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_account_policy_errors()}
  def put_account_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutAccountPolicy", input, options)
  end

  @doc """
  Creates a data protection policy for the specified log group.

  A data protection policy can
  help safeguard sensitive data that's ingested by the log group by auditing and
  masking the
  sensitive log data.

  Sensitive data is detected and masked when it is ingested into the log group.
  When you
  set a data protection policy, log events ingested into the log group before that
  time are
  not masked.

  By default, when a user views a log event that includes masked data, the
  sensitive data is
  replaced by asterisks. A user who has the `logs:Unmask` permission can use a
  [GetLogEvents](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_GetLogEvents.html) or
  [FilterLogEvents](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_FilterLogEvents.html)
  operation with the `unmask` parameter set to
  `true` to view the unmasked log events. Users with the `logs:Unmask`
  can also view unmasked data in the CloudWatch Logs console by running a
  CloudWatch Logs
  Insights query with the `unmask` query command.

  For more information, including a list of types of data that can be audited and
  masked,
  see [Protect sensitive log data with
  masking](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/mask-sensitive-log-data.html).

  The `PutDataProtectionPolicy` operation applies to only the specified log
  group. You can also use
  [PutAccountPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutAccountPolicy.html)
  to create an account-level data protection policy that applies to
  all log groups in the account, including both existing log groups and log groups
  that are
  created level. If a log group has its own data protection policy and the account
  also has an
  account-level data protection policy, then the two policies are cumulative. Any
  sensitive term
  specified in either policy is masked.
  """
  @spec put_data_protection_policy(map(), put_data_protection_policy_request(), list()) ::
          {:ok, put_data_protection_policy_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_data_protection_policy_errors()}
  def put_data_protection_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutDataProtectionPolicy", input, options)
  end

  @doc """
  Creates or updates a logical *delivery destination*.

  A delivery
  destination is an Amazon Web Services resource that represents an Amazon Web
  Services service
  that logs can be sent to. CloudWatch Logs, Amazon S3, and Firehose are
  supported as logs delivery destinations and X-Ray as the trace delivery
  destination.

  To configure logs delivery between a supported Amazon Web Services service and a
  destination, you must do the following:

    *
  Create a delivery source, which is a logical object that represents the resource
  that
  is actually sending the logs. For more information, see
  [PutDeliverySource](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliverySource.html). 

    *
  Use `PutDeliveryDestination` to create a *delivery
  destination* in the same account of the actual delivery destination. The
  delivery destination that you create is a logical object that represents the
  actual
  delivery destination.

    *
  If you are delivering logs cross-account, you must use
  [PutDeliveryDestinationPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliveryDestinationPolicy.html)
  in the destination account to assign an IAM policy to the destination. This
  policy allows delivery to that destination.

    *
  Use `CreateDelivery` to create a *delivery* by pairing
  exactly one delivery source and one delivery destination. For more information,
  see
  [CreateDelivery](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_CreateDelivery.html). 

  You can configure a single delivery source to send logs to multiple destinations
  by
  creating multiple deliveries. You can also create multiple deliveries to
  configure multiple
  delivery sources to send logs to the same delivery destination.

  Only some Amazon Web Services services support being configured as a delivery
  source. These
  services are listed as **Supported [V2 Permissions]** in the
  table at [Enabling logging from Amazon Web Services
  services.](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/AWS-logs-and-resource-policy.html)

  If you use this operation to update an existing delivery destination, all the
  current
  delivery destination parameters are overwritten with the new parameter values
  that you
  specify.
  """
  @spec put_delivery_destination(map(), put_delivery_destination_request(), list()) ::
          {:ok, put_delivery_destination_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_delivery_destination_errors()}
  def put_delivery_destination(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutDeliveryDestination", input, options)
  end

  @doc """
  Creates and assigns an IAM policy that grants permissions to CloudWatch Logs to
  deliver logs cross-account to a specified destination in this account.

  To
  configure the delivery of logs from an Amazon Web Services service in another
  account to a logs
  delivery destination in the current account, you must do the following:

    *
  Create a delivery source, which is a logical object that represents the resource
  that
  is actually sending the logs. For more information, see
  [PutDeliverySource](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliverySource.html). 

    *
  Create a *delivery destination*, which is a logical object that
  represents the actual delivery destination. For more information, see
  [PutDeliveryDestination](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliveryDestination.html).

    *
  Use this operation in the destination account to assign an IAM policy
  to the destination. This policy allows delivery to that destination.

    *
  Create a *delivery* by pairing exactly one delivery source and one
  delivery destination. For more information, see
  [CreateDelivery](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_CreateDelivery.html). 

  Only some Amazon Web Services services support being configured as a delivery
  source. These
  services are listed as **Supported [V2 Permissions]** in the
  table at [Enabling logging from Amazon Web Services
  services.](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/AWS-logs-and-resource-policy.html)

  The contents of the policy must include two statements. One statement enables
  general logs
  delivery, and the other allows delivery to the chosen destination. See the
  examples for the
  needed policies.
  """
  @spec put_delivery_destination_policy(map(), put_delivery_destination_policy_request(), list()) ::
          {:ok, put_delivery_destination_policy_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_delivery_destination_policy_errors()}
  def put_delivery_destination_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutDeliveryDestinationPolicy", input, options)
  end

  @doc """
  Creates or updates a logical *delivery source*.

  A delivery source
  represents an Amazon Web Services resource that sends logs to an logs delivery
  destination. The
  destination can be CloudWatch Logs, Amazon S3, Firehose or X-Ray for sending
  traces.

  To configure logs delivery between a delivery destination and an Amazon Web
  Services
  service that is supported as a delivery source, you must do the following:

    *
  Use `PutDeliverySource` to create a delivery source, which is a logical
  object that represents the resource that is actually sending the logs.

    *
  Use `PutDeliveryDestination` to create a *delivery
  destination*, which is a logical object that represents the actual delivery
  destination. For more information, see
  [PutDeliveryDestination](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliveryDestination.html). 

    *
  If you are delivering logs cross-account, you must use
  [PutDeliveryDestinationPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDeliveryDestinationPolicy.html)
  in the destination account to assign an IAM policy to the destination. This
  policy allows delivery to that destination.

    *
  Use `CreateDelivery` to create a *delivery* by pairing
  exactly one delivery source and one delivery destination. For more information,
  see
  [CreateDelivery](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_CreateDelivery.html). 

  You can configure a single delivery source to send logs to multiple destinations
  by
  creating multiple deliveries. You can also create multiple deliveries to
  configure multiple
  delivery sources to send logs to the same delivery destination.

  Only some Amazon Web Services services support being configured as a delivery
  source. These
  services are listed as **Supported [V2 Permissions]** in the
  table at [Enabling logging from Amazon Web Services
  services.](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/AWS-logs-and-resource-policy.html)

  If you use this operation to update an existing delivery source, all the current
  delivery
  source parameters are overwritten with the new parameter values that you
  specify.
  """
  @spec put_delivery_source(map(), put_delivery_source_request(), list()) ::
          {:ok, put_delivery_source_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_delivery_source_errors()}
  def put_delivery_source(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutDeliverySource", input, options)
  end

  @doc """
  Creates or updates a destination.

  This operation is used only to create destinations
  for cross-account subscriptions.

  A destination encapsulates a physical resource (such as an Amazon Kinesis
  stream). With
  a destination, you can subscribe to a real-time stream of log events for a
  different account,
  ingested using
  [PutLogEvents](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutLogEvents.html).   Through an access policy, a destination controls what is written to it. By
  default,
  `PutDestination` does not set any access policy with the destination, which
  means
  a cross-account user cannot call
  [PutSubscriptionFilter](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutSubscriptionFilter.html)
  against this destination. To enable this, the destination
  owner must call
  [PutDestinationPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDestinationPolicy.html)
  after `PutDestination`.

  To perform a `PutDestination` operation, you must also have the
  `iam:PassRole` permission.
  """
  @spec put_destination(map(), put_destination_request(), list()) ::
          {:ok, put_destination_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_destination_errors()}
  def put_destination(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutDestination", input, options)
  end

  @doc """
  Creates or updates an access policy associated with an existing destination.

  An access
  policy is an [IAM policy
  document](https://docs.aws.amazon.com/IAM/latest/UserGuide/policies_overview.html)
  that is used to authorize claims to register a subscription filter
  against a given destination.
  """
  @spec put_destination_policy(map(), put_destination_policy_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_destination_policy_errors()}
  def put_destination_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutDestinationPolicy", input, options)
  end

  @doc """
  Creates or updates a *field index policy* for the specified log group.

  Only log groups in the Standard log class support field index policies. For more
  information
  about log classes, see [Log classes](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatch_Logs_Log_Classes.html).

  You can use field index policies to create *field indexes* on fields
  found in log events in the log group. Creating field indexes speeds up and
  lowers the costs
  for CloudWatch Logs Insights queries that reference those field indexes, because
  these
  queries attempt to skip the processing of log events that are known to not match
  the indexed
  field. Good fields to index are fields that you often need to query for and
  fields or values
  that match only a small fraction of the total log events. Common examples of
  indexes include
  request ID, session ID, userID, and instance IDs. For more information, see
  [Create field indexes to improve query performance and reduce costs](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatchLogs-Field-Indexing.html).

  To find the fields that are in your log group events, use the
  [GetLogGroupFields](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_GetLogGroupFields.html) operation.

  For example, suppose you have created a field index for `requestId`. Then, any
  CloudWatch Logs Insights query on that log group that includes

  ```
  requestId =
  *value*

  ```

  or

  ```
  requestId IN [*value*,
  *value*, ...]
  ```

  will process fewer log events to reduce costs, and
  have improved performance.

  CloudWatch Logs provides default field indexes for all log groups in the
  Standard log
  class. Default field indexes are automatically available for the following
  fields:

    *

  `@aws.region`

    *

  `@aws.account`

    *

  `@source.log`

    *

  `traceId`

  Default field indexes are in addition to any custom field indexes you define
  within your
  policy. Default field indexes are not counted towards your field index quota.

  Each index policy has the following quotas and restrictions:

    *
  As many as 20 fields can be included in the policy.

    *
  Each field name can include as many as 100 characters.

  Matches of log events to the names of indexed fields are case-sensitive. For
  example, a
  field index of `RequestId` won't match a log event containing
  `requestId`.

  Log group-level field index policies created with `PutIndexPolicy` override
  account-level field index policies created with
  [PutAccountPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutAccountPolicy.html).
  If you use `PutIndexPolicy` to create a field index
  policy for a log group, that log group uses only that policy. The log group
  ignores any
  account-wide field index policy that you might have created.
  """
  @spec put_index_policy(map(), put_index_policy_request(), list()) ::
          {:ok, put_index_policy_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_index_policy_errors()}
  def put_index_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutIndexPolicy", input, options)
  end

  @doc """
  Creates an integration between CloudWatch Logs and another service in this
  account.

  Currently, only integrations with OpenSearch Service are supported, and
  currently you can have
  only one integration in your account.

  Integrating with OpenSearch Service makes it possible for you to create curated
  vended
  logs dashboards, powered by OpenSearch Service analytics. For more information,
  see [Vended log dashboards powered by Amazon OpenSearch
  Service](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatchLogs-OpenSearch-Dashboards.html).

  You can use this operation only to create a new integration. You can't modify an
  existing
  integration.
  """
  @spec put_integration(map(), put_integration_request(), list()) ::
          {:ok, put_integration_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_integration_errors()}
  def put_integration(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutIntegration", input, options)
  end

  @doc """
  Uploads a batch of log events to the specified log stream.

  The sequence token is now ignored in `PutLogEvents` actions.
  `PutLogEvents` actions are always accepted and never return
  `InvalidSequenceTokenException` or `DataAlreadyAcceptedException`
  even if the sequence token is not valid. You can use parallel `PutLogEvents`
  actions on the same log stream.

  The batch of events must satisfy the following constraints:

    *
  The maximum batch size is 1,048,576 bytes. This size is calculated as the sum of
  all event messages in UTF-8, plus 26 bytes for each log event.

    *
  Events more than 2 hours in the future are rejected while processing remaining
  valid events.

    *
  Events older than 14 days or preceding the log group's retention period are
  rejected while processing remaining valid events.

    *
  The log events in the batch must be in chronological order by their timestamp.
  The
  timestamp is the time that the event occurred, expressed as the number of
  milliseconds
  after `Jan 1, 1970 00:00:00 UTC`. (In Amazon Web Services Tools for PowerShell
  and the Amazon Web Services SDK for .NET, the timestamp is specified in .NET
  format:
  `yyyy-mm-ddThh:mm:ss`. For example, `2017-09-15T13:45:30`.)

    *
  A batch of log events in a single request must be in a chronological order.
  Otherwise, the operation fails.

    *
  Each log event can be no larger than 1 MB.

    *
  The maximum number of log events in a batch is 10,000.

    *
  For valid events (within 14 days in the past to 2 hours in future), the time
  span
  in a single batch cannot exceed 24 hours. Otherwise, the operation fails.

  The quota of five requests per second per log stream has been removed. Instead,
  `PutLogEvents` actions are throttled based on a per-second per-account quota.
  You can request an increase to the per-second throttling quota by using the
  Service Quotas service.

  If a call to `PutLogEvents` returns "UnrecognizedClientException" the most
  likely cause is a non-valid Amazon Web Services access key ID or secret key.
  """
  @spec put_log_events(map(), put_log_events_request(), list()) ::
          {:ok, put_log_events_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_log_events_errors()}
  def put_log_events(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutLogEvents", input, options)
  end

  @doc """
  Creates or updates a metric filter and associates it with the specified log
  group.

  With
  metric filters, you can configure rules to extract metric data from log events
  ingested
  through
  [PutLogEvents](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutLogEvents.html).   The maximum number of metric filters that can be associated with a log group is
  100.

  Using regular expressions in filter patterns is supported. For these filters,
  there is a
  quota of two regular expression patterns within a single filter pattern. There
  is also a quota
  of five regular expression patterns per log group. For more information about
  using regular
  expressions in filter patterns, see [ Filter pattern syntax for
  metric filters, subscription filters, filter log events, and Live
  Tail](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/FilterAndPatternSyntax.html).

  When you create a metric filter, you can also optionally assign a unit and
  dimensions to
  the metric that is created.

  Metrics extracted from log events are charged as custom metrics. To prevent
  unexpected
  high charges, do not specify high-cardinality fields such as `IPAddress` or
  `requestID` as dimensions. Each different value found for a dimension is
  treated as a separate metric and accrues charges as a separate custom metric.

  CloudWatch Logs might disable a metric filter if it generates 1,000 different
  name/value pairs for your specified dimensions within one hour.

  You can also set up a billing alarm to alert you if your charges are higher than
  expected. For more information, see [
  Creating a Billing Alarm to Monitor Your Estimated Amazon Web Services
  Charges](https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/monitor_estimated_charges_with_cloudwatch.html).
  """
  @spec put_metric_filter(map(), put_metric_filter_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_metric_filter_errors()}
  def put_metric_filter(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutMetricFilter", input, options)
  end

  @doc """
  Creates or updates a query definition for CloudWatch Logs Insights.

  For more information,
  see [Analyzing Log Data with CloudWatch Logs Insights](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/AnalyzingLogData.html).

  To update a query definition, specify its `queryDefinitionId` in your request.
  The values of `name`, `queryString`, and `logGroupNames` are
  changed to the values that you specify in your update operation. No current
  values are
  retained from the current query definition. For example, imagine updating a
  current query
  definition that includes log groups. If you don't specify the `logGroupNames`
  parameter in your update operation, the query definition changes to contain no
  log
  groups.

  You must have the `logs:PutQueryDefinition` permission to be able to perform
  this operation.
  """
  @spec put_query_definition(map(), put_query_definition_request(), list()) ::
          {:ok, put_query_definition_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_query_definition_errors()}
  def put_query_definition(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutQueryDefinition", input, options)
  end

  @doc """
  Creates or updates a resource policy allowing other Amazon Web Services services
  to put
  log events to this account, such as Amazon Route 53.

  An account can have up to 10 resource
  policies per Amazon Web Services Region.
  """
  @spec put_resource_policy(map(), put_resource_policy_request(), list()) ::
          {:ok, put_resource_policy_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_resource_policy_errors()}
  def put_resource_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutResourcePolicy", input, options)
  end

  @doc """
  Sets the retention of the specified log group.

  With a retention policy, you can
  configure the number of days for which to retain log events in the specified log
  group.

  CloudWatch Logs doesn't immediately delete log events when they reach their
  retention
  setting. It typically takes up to 72 hours after that before log events are
  deleted, but in
  rare situations might take longer.

  To illustrate, imagine that you change a log group to have a longer retention
  setting
  when it contains log events that are past the expiration date, but haven't been
  deleted.
  Those log events will take up to 72 hours to be deleted after the new retention
  date is
  reached. To make sure that log data is deleted permanently, keep a log group at
  its lower
  retention setting until 72 hours after the previous retention period ends.
  Alternatively,
  wait to change the retention setting until you confirm that the earlier log
  events are
  deleted.

  When log events reach their retention setting they are marked for deletion.
  After they
  are marked for deletion, they do not add to your archival storage costs anymore,
  even if
  they are not actually deleted until later. These log events marked for deletion
  are also not
  included when you use an API to retrieve the `storedBytes` value to see how many
  bytes a log group is storing.
  """
  @spec put_retention_policy(map(), put_retention_policy_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_retention_policy_errors()}
  def put_retention_policy(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutRetentionPolicy", input, options)
  end

  @doc """
  Creates or updates a subscription filter and associates it with the specified
  log
  group.

  With subscription filters, you can subscribe to a real-time stream of log events
  ingested through
  [PutLogEvents](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutLogEvents.html) and have them delivered to a specific destination. When log events are sent to
  the receiving
  service, they are Base64 encoded and compressed with the GZIP format.

  The following destinations are supported for subscription filters:

    *
  An Amazon Kinesis data stream belonging to the same account as the subscription
  filter, for same-account delivery.

    *
  A logical destination created with
  [PutDestination](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutDestination.html)
  that belongs to a different account, for cross-account delivery.
  We currently support Kinesis Data Streams and Firehose as logical
  destinations.

    *
  An Amazon Kinesis Data Firehose delivery stream that belongs to the same account
  as
  the subscription filter, for same-account delivery.

    *
  An Lambda function that belongs to the same account as the
  subscription filter, for same-account delivery.

  Each log group can have up to two subscription filters associated with it. If
  you are
  updating an existing filter, you must specify the correct name in `filterName`.

  Using regular expressions in filter patterns is supported. For these filters,
  there is a
  quotas of quota of two regular expression patterns within a single filter
  pattern. There is
  also a quota of five regular expression patterns per log group. For more
  information about
  using regular expressions in filter patterns, see [ Filter pattern syntax for metric filters, subscription filters, filter log events, and Live
  Tail](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/FilterAndPatternSyntax.html).

  To perform a `PutSubscriptionFilter` operation for any destination except a
  Lambda function, you must also have the `iam:PassRole`
  permission.
  """
  @spec put_subscription_filter(map(), put_subscription_filter_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_subscription_filter_errors()}
  def put_subscription_filter(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutSubscriptionFilter", input, options)
  end

  @doc """
  Creates or updates a *log transformer* for a single log group.

  You use
  log transformers to transform log events into a different format, making them
  easier for you
  to process and analyze. You can also transform logs from different sources into
  standardized
  formats that contains relevant, source-specific information.

  After you have created a transformer, CloudWatch Logs performs the
  transformations at
  the time of log ingestion. You can then refer to the transformed versions of the
  logs during
  operations such as querying with CloudWatch Logs Insights or creating metric
  filters or
  subscription filers.

  You can also use a transformer to copy metadata from metadata keys into the log
  events
  themselves. This metadata can include log group name, log stream name, account
  ID and
  Region.

  A transformer for a log group is a series of processors, where each processor
  applies one
  type of transformation to the log events ingested into this log group. The
  processors work one
  after another, in the order that you list them, like a pipeline. For more
  information about
  the available processors to use in a transformer, see [ Processors that you can use](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatch-Logs-Transformation.html#CloudWatch-Logs-Transformation-Processors).

  Having log events in standardized format enables visibility across your
  applications for
  your log analysis, reporting, and alarming needs. CloudWatch Logs provides
  transformation
  for common log types with out-of-the-box transformation templates for major
  Amazon Web Services
  log sources such as VPC flow logs, Lambda, and Amazon RDS. You can use
  pre-built transformation templates or create custom transformation policies.

  You can create transformers only for the log groups in the Standard log class.

  You can also set up a transformer at the account level. For more information,
  see
  [PutAccountPolicy](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutAccountPolicy.html).
  If there is both a log-group level transformer created with
  `PutTransformer` and an account-level transformer that could apply to the same
  log group, the log group uses only the log-group level transformer. It ignores
  the
  account-level transformer.
  """
  @spec put_transformer(map(), put_transformer_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, put_transformer_errors()}
  def put_transformer(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "PutTransformer", input, options)
  end

  @doc """
  Starts a Live Tail streaming session for one or more log groups.

  A Live Tail session
  returns a stream of log events that have been recently ingested in the log
  groups. For more
  information, see [Use Live Tail to view logs in near real
  time](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CloudWatchLogs_LiveTail.html).

  The response to this operation is a response stream, over which the server sends
  live log
  events and the client receives them.

  The following objects are sent over the stream:

    *
  A single
  [LiveTailSessionStart](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_LiveTailSessionStart.html) object is sent at the start of the session.

    *
  Every second, a
  [LiveTailSessionUpdate](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_LiveTailSessionUpdate.html)
  object is sent. Each of these objects contains an array
  of the actual log events.

  If no new log events were ingested in the past second, the
  `LiveTailSessionUpdate` object will contain an empty array.

  The array of log events contained in a `LiveTailSessionUpdate` can include
  as many as 500 log events. If the number of log events matching the request
  exceeds 500
  per second, the log events are sampled down to 500 log events to be included in
  each
  `LiveTailSessionUpdate` object.

  If your client consumes the log events slower than the server produces them,
  CloudWatch Logs buffers up to 10 `LiveTailSessionUpdate` events or 5000 log
  events, after which it starts dropping the oldest events.

    *
  A
  [SessionStreamingException](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_StartLiveTailResponseStream.html#CWL-Type-StartLiveTailResponseStream-SessionStreamingException) object is returned if an unknown error occurs on the
  server side.

    *
  A
  [SessionTimeoutException](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_StartLiveTailResponseStream.html#CWL-Type-StartLiveTailResponseStream-SessionTimeoutException)
  object is returned when the session times out, after it
  has been kept open for three hours.

  The `StartLiveTail` API routes requests to
  `streaming-logs.*Region*.amazonaws.com` using SDK host
  prefix injection. VPC endpoint support is not available for this API.

  You can end a session before it times out by closing the session stream or by
  closing
  the client that is receiving the stream. The session also ends if the
  established connection
  between the client and the server breaks.

  For examples of using an SDK to start a Live Tail session, see [ Start a Live Tail session using an Amazon Web Services
  SDK](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/example_cloudwatch-logs_StartLiveTail_section.html).
  """
  @spec start_live_tail(map(), start_live_tail_request(), list()) ::
          {:ok, start_live_tail_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, start_live_tail_errors()}
  def start_live_tail(%Client{} = client, input, options \\ []) do
    meta = metadata() |> Map.put_new(:host_prefix, "streaming-")

    Request.request_post(client, meta, "StartLiveTail", input, options)
  end

  @doc """
  Starts a query of one or more log groups using CloudWatch Logs Insights.

  You specify
  the log groups and time range to query and the query string to use.

  For more information, see [CloudWatch Logs Insights Query Syntax](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CWL_QuerySyntax.html).

  After you run a query using `StartQuery`, the query results are stored by
  CloudWatch Logs. You can use
  [GetQueryResults](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_GetQueryResults.html) to retrieve the results of a query, using the `queryId`
  that `StartQuery` returns.

  To specify the log groups to query, a `StartQuery` operation must include one
  of the following:

    
  Either exactly one of the following parameters: `logGroupName`,
  `logGroupNames`, or `logGroupIdentifiers`

    
  Or the `queryString` must include a `SOURCE` command to select
  log groups for the query. The `SOURCE` command can select log groups based on
  log group name prefix, account ID, and log class.

  For more information about the `SOURCE` command, see
  [SOURCE](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CWL_QuerySyntax-Source.html).

  If you have associated a KMS key with the query results in this
  account, then
  [StartQuery](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_StartQuery.html) uses
  that key to encrypt the results when it stores them. If no key is associated
  with query
  results, the query results are encrypted with the default CloudWatch Logs
  encryption
  method.

  Queries time out after 60 minutes of runtime. If your queries are timing out,
  reduce the
  time range being searched or partition your query into a number of queries.

  If you are using CloudWatch cross-account observability, you can use this
  operation
  in a monitoring account to start a query in a linked source account. For more
  information, see
  [CloudWatch cross-account
  observability](https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/CloudWatch-Unified-Cross-Account.html).
  For a cross-account `StartQuery`
  operation, the query definition must be defined in the monitoring account.

  You can have up to 30 concurrent CloudWatch Logs insights queries, including
  queries
  that have been added to dashboards.
  """
  @spec start_query(map(), start_query_request(), list()) ::
          {:ok, start_query_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, start_query_errors()}
  def start_query(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "StartQuery", input, options)
  end

  @doc """
  Stops a CloudWatch Logs Insights query that is in progress.

  If the query has already
  ended, the operation returns an error indicating that the specified query is not
  running.
  """
  @spec stop_query(map(), stop_query_request(), list()) ::
          {:ok, stop_query_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, stop_query_errors()}
  def stop_query(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "StopQuery", input, options)
  end

  @doc """

  The TagLogGroup operation is on the path to deprecation.

  We recommend that you use
  [TagResource](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_TagResource.html)  instead.

  Adds or updates the specified tags for the specified log group.

  To list the tags for a log group, use
  [ListTagsForResource](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_ListTagsForResource.html).
  To remove tags, use
  [UntagResource](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_UntagResource.html).   For more information about tags, see [Tag Log Groups in Amazon CloudWatch
  Logs](https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/Working-with-log-groups-and-streams.html#log-group-tagging)
  in the *Amazon CloudWatch Logs
  User Guide*.

  CloudWatch Logs doesn't support IAM policies that prevent users from assigning
  specified
  tags to log groups using the

  ```
  aws:Resource/*key-name*

  ```

  or
  `aws:TagKeys` condition keys. For more information about using tags to control
  access, see [Controlling access to Amazon Web Services resources using tags](https://docs.aws.amazon.com/IAM/latest/UserGuide/access_tags.html).
  """
  @spec tag_log_group(map(), tag_log_group_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, tag_log_group_errors()}
  def tag_log_group(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "TagLogGroup", input, options)
  end

  @doc """
  Assigns one or more tags (key-value pairs) to the specified CloudWatch Logs
  resource.

  Currently, the only CloudWatch Logs resources that can be tagged are log groups
  and
  destinations.

  Tags can help you organize and categorize your resources. You can also use them
  to scope
  user permissions by granting a user permission to access or change only
  resources with certain
  tag values.

  Tags don't have any semantic meaning to Amazon Web Services and are interpreted
  strictly as
  strings of characters.

  You can use the `TagResource` action with a resource that already has tags. If
  you specify a new tag key for the alarm, this tag is appended to the list of
  tags associated
  with the alarm. If you specify a tag key that is already associated with the
  alarm, the new
  tag value that you specify replaces the previous value for that tag.

  You can associate as many as 50 tags with a CloudWatch Logs resource.
  """
  @spec tag_resource(map(), tag_resource_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, tag_resource_errors()}
  def tag_resource(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "TagResource", input, options)
  end

  @doc """
  Tests the filter pattern of a metric filter against a sample of log event
  messages.

  You
  can use this operation to validate the correctness of a metric filter pattern.
  """
  @spec test_metric_filter(map(), test_metric_filter_request(), list()) ::
          {:ok, test_metric_filter_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, test_metric_filter_errors()}
  def test_metric_filter(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "TestMetricFilter", input, options)
  end

  @doc """
  Use this operation to test a log transformer.

  You enter the transformer configuration and
  a set of log events to test with. The operation responds with an array that
  includes the
  original log events and the transformed versions.
  """
  @spec test_transformer(map(), test_transformer_request(), list()) ::
          {:ok, test_transformer_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, test_transformer_errors()}
  def test_transformer(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "TestTransformer", input, options)
  end

  @doc """

  The UntagLogGroup operation is on the path to deprecation.

  We recommend that you use
  [UntagResource](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_UntagResource.html) instead.

  Removes the specified tags from the specified log group.

  To list the tags for a log group, use
  [ListTagsForResource](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_ListTagsForResource.html).
  To add tags, use
  [TagResource](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_TagResource.html).

  When using IAM policies to control tag management for CloudWatch Logs log
  groups, the
  condition keys `aws:Resource/key-name` and `aws:TagKeys` cannot be used
  to restrict which tags users can assign.
  """
  @spec untag_log_group(map(), untag_log_group_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, untag_log_group_errors()}
  def untag_log_group(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "UntagLogGroup", input, options)
  end

  @doc """
  Removes one or more tags from the specified resource.
  """
  @spec untag_resource(map(), untag_resource_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, untag_resource_errors()}
  def untag_resource(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "UntagResource", input, options)
  end

  @doc """
  Use this operation to *suppress* anomaly detection for a specified
  anomaly or pattern.

  If you suppress an anomaly, CloudWatch Logs won't report new
  occurrences of that anomaly and won't update that anomaly with new data. If you
  suppress a
  pattern, CloudWatch Logs won't report any anomalies related to that pattern.

  You must specify either `anomalyId` or `patternId`, but you can't
  specify both parameters in the same operation.

  If you have previously used this operation to suppress detection of a pattern or
  anomaly,
  you can use it again to cause CloudWatch Logs to end the suppression. To do
  this, use this
  operation and specify the anomaly or pattern to stop suppressing, and omit the
  `suppressionType` and `suppressionPeriod` parameters.
  """
  @spec update_anomaly(map(), update_anomaly_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, update_anomaly_errors()}
  def update_anomaly(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "UpdateAnomaly", input, options)
  end

  @doc """
  Use this operation to update the configuration of a
  [delivery](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_Delivery.html)
  to change
  either the S3 path pattern or the format of the delivered logs.

  You can't use this operation
  to change the source or destination of the delivery.
  """
  @spec update_delivery_configuration(map(), update_delivery_configuration_request(), list()) ::
          {:ok, update_delivery_configuration_response(), any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, update_delivery_configuration_errors()}
  def update_delivery_configuration(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "UpdateDeliveryConfiguration", input, options)
  end

  @doc """
  Updates an existing log anomaly detector.
  """
  @spec update_log_anomaly_detector(map(), update_log_anomaly_detector_request(), list()) ::
          {:ok, nil, any()}
          | {:error, {:unexpected_response, any()}}
          | {:error, term()}
          | {:error, update_log_anomaly_detector_errors()}
  def update_log_anomaly_detector(%Client{} = client, input, options \\ []) do
    meta = metadata()

    Request.request_post(client, meta, "UpdateLogAnomalyDetector", input, options)
  end
end