diff --git a/.bumpversion.cfg b/.bumpversion.cfg index 696b671002196..2622a65ffa2a8 100644 --- a/.bumpversion.cfg +++ b/.bumpversion.cfg @@ -1,5 +1,5 @@ [bumpversion] -current_version = 0.30.38-alpha +current_version = 0.30.39-alpha commit = False tag = False parse = (?P\d+)\.(?P\d+)\.(?P\d+)(\-[a-z]+)? diff --git a/.env b/.env index a997fd1b367bf..2cce69cf57e3e 100644 --- a/.env +++ b/.env @@ -1,4 +1,4 @@ -VERSION=0.30.38-alpha +VERSION=0.30.39-alpha # Airbyte Internal Job Database, see https://docs.airbyte.io/operator-guides/configuring-airbyte-db DATABASE_USER=docker diff --git a/airbyte-config/init/src/main/resources/seed/destination_specs.yaml b/airbyte-config/init/src/main/resources/seed/destination_specs.yaml index 6283f82deaa96..0e52d85a3e79f 100644 --- a/airbyte-config/init/src/main/resources/seed/destination_specs.yaml +++ b/airbyte-config/init/src/main/resources/seed/destination_specs.yaml @@ -9,78 +9,75 @@ title: "AzureBlobStorage Destination Spec" type: "object" required: - - "azure_blob_storage_account_name" - - "azure_blob_storage_account_key" - - "format" + - "azure_blob_storage_account_name" + - "azure_blob_storage_account_key" + - "format" additionalProperties: false properties: azure_blob_storage_endpoint_domain_name: title: "Endpoint Domain Name" type: "string" default: "blob.core.windows.net" - description: - "This is Azure Blob Storage endpoint domain name. Leave default\ + description: "This is Azure Blob Storage endpoint domain name. Leave default\ \ value (or leave it empty if run container from command line) to use\ \ Microsoft native from example." examples: - - "blob.core.windows.net" + - "blob.core.windows.net" azure_blob_storage_container_name: title: "Azure blob storage container (Bucket) Name" type: "string" - description: - "The name of the Azure blob storage container. If not exists\ + description: "The name of the Azure blob storage container. If not exists\ \ - will be created automatically. May be empty, then will be created\ \ automatically airbytecontainer+timestamp" examples: - - "airbytetescontainername" + - "airbytetescontainername" azure_blob_storage_account_name: title: "Azure Blob Storage account name" type: "string" description: "The account's name of the Azure Blob Storage." examples: - - "airbyte5storage" + - "airbyte5storage" azure_blob_storage_account_key: description: "The Azure blob storage account key." airbyte_secret: true type: "string" examples: - - "Z8ZkZpteggFx394vm+PJHnGTvdRncaYS+JhLKdj789YNmD+iyGTnG+PV+POiuYNhBg/ACS+LKjd%4FG3FHGN12Nd==" + - "Z8ZkZpteggFx394vm+PJHnGTvdRncaYS+JhLKdj789YNmD+iyGTnG+PV+POiuYNhBg/ACS+LKjd%4FG3FHGN12Nd==" format: title: "Output Format" type: "object" description: "Output data format" oneOf: - - title: "CSV: Comma-Separated Values" - required: - - "format_type" - - "flattening" - properties: - format_type: - type: "string" - const: "CSV" - flattening: - type: "string" - title: "Normalization (Flattening)" - description: - "Whether the input json data should be normalized (flattened)\ - \ in the output CSV. Please refer to docs for details." - default: "No flattening" - enum: - - "No flattening" - - "Root level flattening" - - title: "JSON Lines: newline-delimited JSON" - required: - - "format_type" - properties: - format_type: - type: "string" - const: "JSONL" + - title: "CSV: Comma-Separated Values" + required: + - "format_type" + - "flattening" + properties: + format_type: + type: "string" + const: "CSV" + flattening: + type: "string" + title: "Normalization (Flattening)" + description: "Whether the input json data should be normalized (flattened)\ + \ in the output CSV. Please refer to docs for details." + default: "No flattening" + enum: + - "No flattening" + - "Root level flattening" + - title: "JSON Lines: newline-delimited JSON" + required: + - "format_type" + properties: + format_type: + type: "string" + const: "JSONL" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-bigquery:0.5.0" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/bigquery" @@ -89,14 +86,13 @@ title: "BigQuery Destination Spec" type: "object" required: - - "project_id" - - "dataset_id" + - "project_id" + - "dataset_id" additionalProperties: true properties: big_query_client_buffer_size_mb: title: "Google BigQuery client chunk size" - description: - "Google BigQuery client's chunk(buffer) size (MIN=1, MAX =\ + description: "Google BigQuery client's chunk(buffer) size (MIN=1, MAX =\ \ 15) for each table. The default 15MiB value is used if not set explicitly.\ \ It's recommended to decrease value for big data sets migration for less\ \ HEAP memory consumption and avoiding crashes. For more details refer\ @@ -106,60 +102,56 @@ maximum: 15 default: 15 examples: - - "15" + - "15" project_id: type: "string" - description: - "The GCP project ID for the project containing the target BigQuery\ + description: "The GCP project ID for the project containing the target BigQuery\ \ dataset." title: "Project ID" dataset_id: type: "string" - description: - "Default BigQuery Dataset ID tables are replicated to if the\ + description: "Default BigQuery Dataset ID tables are replicated to if the\ \ source does not specify a namespace." title: "Default Dataset ID" dataset_location: type: "string" - description: - "The location of the dataset. Warning: Changes made after creation\ + description: "The location of the dataset. Warning: Changes made after creation\ \ will not be applied." title: "Dataset Location" default: "US" enum: - - "US" - - "EU" - - "asia-east1" - - "asia-east2" - - "asia-northeast1" - - "asia-northeast2" - - "asia-northeast3" - - "asia-south1" - - "asia-southeast1" - - "asia-southeast2" - - "australia-southeast1" - - "europe-central1" - - "europe-central2" - - "europe-north1" - - "europe-west1" - - "europe-west2" - - "europe-west3" - - "europe-west4" - - "europe-west5" - - "europe-west6" - - "northamerica-northeast1" - - "southamerica-east1" - - "us-central1" - - "us-east1" - - "us-east4" - - "us-west-1" - - "us-west-2" - - "us-west-3" - - "us-west-4" + - "US" + - "EU" + - "asia-east1" + - "asia-east2" + - "asia-northeast1" + - "asia-northeast2" + - "asia-northeast3" + - "asia-south1" + - "asia-southeast1" + - "asia-southeast2" + - "australia-southeast1" + - "europe-central1" + - "europe-central2" + - "europe-north1" + - "europe-west1" + - "europe-west2" + - "europe-west3" + - "europe-west4" + - "europe-west5" + - "europe-west6" + - "northamerica-northeast1" + - "southamerica-east1" + - "us-central1" + - "us-east1" + - "us-east4" + - "us-west-1" + - "us-west-2" + - "us-west-3" + - "us-west-4" credentials_json: type: "string" - description: - "The contents of the JSON service account key. Check out the\ + description: "The contents of the JSON service account key. Check out the\ \ docs if you need help generating this key. Default credentials will\ \ be used if this field is left empty." @@ -167,107 +159,101 @@ airbyte_secret: true transformation_priority: type: "string" - description: - "When running custom transformations or Basic normalization,\ + description: "When running custom transformations or Basic normalization,\ \ running queries on interactive mode can hit BQ limits, choosing batch\ \ will solve those limitss." title: "Transformation Query Run Type" default: "interactive" enum: - - "interactive" - - "batch" + - "interactive" + - "batch" loading_method: type: "object" title: "Loading Method" - description: - "Loading method used to send select the way data will be uploaded\ + description: "Loading method used to send select the way data will be uploaded\ \ to BigQuery." oneOf: - - title: "Standard Inserts" - additionalProperties: false - description: "Direct uploading using streams." - required: - - "method" - properties: - method: - type: "string" - const: "Standard" - - title: "GCS Staging" - additionalProperties: false - description: - "Writes large batches of records to a file, uploads the file\ - \ to GCS, then uses
COPY INTO table
to upload the file. Recommended\ - \ for large production workloads for better speed and scalability." - required: - - "method" - - "gcs_bucket_name" - - "gcs_bucket_path" - - "credential" - properties: - method: - type: "string" - const: "GCS Staging" - gcs_bucket_name: - title: "GCS Bucket Name" - type: "string" - description: "The name of the GCS bucket." - examples: - - "airbyte_sync" - gcs_bucket_path: - description: "Directory under the GCS bucket where data will be written." - type: "string" - examples: - - "data_sync/test" - keep_files_in_gcs-bucket: - type: "string" - description: - "This upload method is supposed to temporary store records\ - \ in GCS bucket. What do you want to do with data in GCS bucket\ - \ when migration has finished?" - title: "GCS tmp files afterward processing" - default: "Delete all tmp files from GCS" - enum: - - "Delete all tmp files from GCS" - - "Keep all tmp files in GCS" - credential: - title: "Credential" - type: "object" - oneOf: - - title: "HMAC key" - required: - - "credential_type" - - "hmac_key_access_id" - - "hmac_key_secret" - properties: - credential_type: - type: "string" - const: "HMAC_KEY" - hmac_key_access_id: - type: "string" - description: - "HMAC key access ID. When linked to a service account,\ - \ this ID is 61 characters long; when linked to a user account,\ - \ it is 24 characters long." - title: "HMAC Key Access ID" - airbyte_secret: true - examples: - - "1234567890abcdefghij1234" - hmac_key_secret: - type: "string" - description: - "The corresponding secret for the access ID. It\ - \ is a 40-character base-64 encoded string." - title: "HMAC Key Secret" - airbyte_secret: true - examples: - - "1234567890abcdefghij1234567890ABCDEFGHIJ" + - title: "Standard Inserts" + additionalProperties: false + description: "Direct uploading using streams." + required: + - "method" + properties: + method: + type: "string" + const: "Standard" + - title: "GCS Staging" + additionalProperties: false + description: "Writes large batches of records to a file, uploads the file\ + \ to GCS, then uses
COPY INTO table
to upload the file. Recommended\ + \ for large production workloads for better speed and scalability." + required: + - "method" + - "gcs_bucket_name" + - "gcs_bucket_path" + - "credential" + properties: + method: + type: "string" + const: "GCS Staging" + gcs_bucket_name: + title: "GCS Bucket Name" + type: "string" + description: "The name of the GCS bucket." + examples: + - "airbyte_sync" + gcs_bucket_path: + description: "Directory under the GCS bucket where data will be written." + type: "string" + examples: + - "data_sync/test" + keep_files_in_gcs-bucket: + type: "string" + description: "This upload method is supposed to temporary store records\ + \ in GCS bucket. What do you want to do with data in GCS bucket\ + \ when migration has finished?" + title: "GCS tmp files afterward processing" + default: "Delete all tmp files from GCS" + enum: + - "Delete all tmp files from GCS" + - "Keep all tmp files in GCS" + credential: + title: "Credential" + type: "object" + oneOf: + - title: "HMAC key" + required: + - "credential_type" + - "hmac_key_access_id" + - "hmac_key_secret" + properties: + credential_type: + type: "string" + const: "HMAC_KEY" + hmac_key_access_id: + type: "string" + description: "HMAC key access ID. When linked to a service account,\ + \ this ID is 61 characters long; when linked to a user account,\ + \ it is 24 characters long." + title: "HMAC Key Access ID" + airbyte_secret: true + examples: + - "1234567890abcdefghij1234" + hmac_key_secret: + type: "string" + description: "The corresponding secret for the access ID. It\ + \ is a 40-character base-64 encoded string." + title: "HMAC Key Secret" + airbyte_secret: true + examples: + - "1234567890abcdefghij1234567890ABCDEFGHIJ" supportsIncremental: true supportsNormalization: true supportsDBT: true supported_destination_sync_modes: - - "overwrite" - - "append" - - "append_dedup" + - "overwrite" + - "append" + - "append_dedup" - dockerImage: "airbyte/destination-bigquery-denormalized:0.1.10" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/bigquery" @@ -276,63 +262,59 @@ title: "BigQuery Denormalized Typed Struct Destination Spec" type: "object" required: - - "project_id" - - "dataset_id" + - "project_id" + - "dataset_id" additionalProperties: true properties: project_id: type: "string" - description: - "The GCP project ID for the project containing the target BigQuery\ + description: "The GCP project ID for the project containing the target BigQuery\ \ dataset." title: "Project ID" dataset_id: type: "string" - description: - "Default BigQuery Dataset ID tables are replicated to if the\ + description: "Default BigQuery Dataset ID tables are replicated to if the\ \ source does not specify a namespace." title: "Default Dataset ID" dataset_location: type: "string" - description: - "The location of the dataset. Warning: Changes made after creation\ + description: "The location of the dataset. Warning: Changes made after creation\ \ will not be applied." title: "Dataset Location" default: "US" enum: - - "US" - - "EU" - - "asia-east1" - - "asia-east2" - - "asia-northeast1" - - "asia-northeast2" - - "asia-northeast3" - - "asia-south1" - - "asia-southeast1" - - "asia-southeast2" - - "australia-southeast1" - - "europe-central1" - - "europe-central2" - - "europe-north1" - - "europe-west1" - - "europe-west2" - - "europe-west3" - - "europe-west4" - - "europe-west5" - - "europe-west6" - - "northamerica-northeast1" - - "southamerica-east1" - - "us-central1" - - "us-east1" - - "us-east4" - - "us-west-1" - - "us-west-2" - - "us-west-3" - - "us-west-4" + - "US" + - "EU" + - "asia-east1" + - "asia-east2" + - "asia-northeast1" + - "asia-northeast2" + - "asia-northeast3" + - "asia-south1" + - "asia-southeast1" + - "asia-southeast2" + - "australia-southeast1" + - "europe-central1" + - "europe-central2" + - "europe-north1" + - "europe-west1" + - "europe-west2" + - "europe-west3" + - "europe-west4" + - "europe-west5" + - "europe-west6" + - "northamerica-northeast1" + - "southamerica-east1" + - "us-central1" + - "us-east1" + - "us-east4" + - "us-west-1" + - "us-west-2" + - "us-west-3" + - "us-west-4" credentials_json: type: "string" - description: - "The contents of the JSON service account key. Check out the\ + description: "The contents of the JSON service account key. Check out the\ \ docs if you need help generating this key. Default credentials will\ \ be used if this field is left empty." @@ -342,8 +324,8 @@ supportsNormalization: false supportsDBT: true supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-cassandra:0.1.0" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/cassandra" @@ -352,11 +334,11 @@ title: "Cassandra Destination Spec" type: "object" required: - - "keyspace" - - "username" - - "password" - - "address" - - "port" + - "keyspace" + - "username" + - "password" + - "address" + - "port" additionalProperties: true properties: keyspace: @@ -380,7 +362,7 @@ description: "Address to connect to." type: "string" examples: - - "localhost,127.0.0.1" + - "localhost,127.0.0.1" order: 3 port: title: "Port" @@ -399,8 +381,7 @@ replication: title: "Replication factor" type: "integer" - description: - "Indicates to how many nodes the data should be replicated\ + description: "Indicates to how many nodes the data should be replicated\ \ to." default: 1 order: 6 @@ -408,8 +389,8 @@ supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-keen:0.2.0" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/keen" @@ -418,26 +399,25 @@ title: "Keen Spec" type: "object" required: - - "project_id" - - "api_key" + - "project_id" + - "api_key" additionalProperties: false properties: project_id: description: "Keen Project ID" type: "string" examples: - - "58b4acc22ba938934e888322e" + - "58b4acc22ba938934e888322e" api_key: title: "API Key" description: "Keen Master API key" type: "string" examples: - - "ABCDEFGHIJKLMNOPRSTUWXYZ" + - "ABCDEFGHIJKLMNOPRSTUWXYZ" airbyte_secret: true infer_timestamp: title: "Infer Timestamp" - description: - "Allow connector to guess keen.timestamp value based on the\ + description: "Allow connector to guess keen.timestamp value based on the\ \ streamed data" type: "boolean" default: true @@ -445,8 +425,8 @@ supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-dynamodb:0.1.0" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/dynamodb" @@ -455,81 +435,79 @@ title: "DynamoDB Destination Spec" type: "object" required: - - "dynamodb_table_name" - - "dynamodb_region" - - "access_key_id" - - "secret_access_key" + - "dynamodb_table_name" + - "dynamodb_region" + - "access_key_id" + - "secret_access_key" additionalProperties: false properties: dynamodb_endpoint: title: "Endpoint" type: "string" default: "" - description: - "This is your DynamoDB endpoint url.(if you are working with\ + description: "This is your DynamoDB endpoint url.(if you are working with\ \ AWS DynamoDB, just leave empty)." examples: - - "http://localhost:9000" + - "http://localhost:9000" dynamodb_table_name: title: "DynamoDB Table Name" type: "string" description: "The name of the DynamoDB table." examples: - - "airbyte_sync" + - "airbyte_sync" dynamodb_region: title: "DynamoDB Region" type: "string" default: "" description: "The region of the DynamoDB." enum: - - "" - - "us-east-1" - - "us-east-2" - - "us-west-1" - - "us-west-2" - - "af-south-1" - - "ap-east-1" - - "ap-south-1" - - "ap-northeast-1" - - "ap-northeast-2" - - "ap-northeast-3" - - "ap-southeast-1" - - "ap-southeast-2" - - "ca-central-1" - - "cn-north-1" - - "cn-northwest-1" - - "eu-central-1" - - "eu-north-1" - - "eu-south-1" - - "eu-west-1" - - "eu-west-2" - - "eu-west-3" - - "sa-east-1" - - "me-south-1" - - "us-gov-east-1" - - "us-gov-west-1" + - "" + - "us-east-1" + - "us-east-2" + - "us-west-1" + - "us-west-2" + - "af-south-1" + - "ap-east-1" + - "ap-south-1" + - "ap-northeast-1" + - "ap-northeast-2" + - "ap-northeast-3" + - "ap-southeast-1" + - "ap-southeast-2" + - "ca-central-1" + - "cn-north-1" + - "cn-northwest-1" + - "eu-central-1" + - "eu-north-1" + - "eu-south-1" + - "eu-west-1" + - "eu-west-2" + - "eu-west-3" + - "sa-east-1" + - "me-south-1" + - "us-gov-east-1" + - "us-gov-west-1" access_key_id: type: "string" - description: - "The access key id to access the DynamoDB. Airbyte requires\ + description: "The access key id to access the DynamoDB. Airbyte requires\ \ Read and Write permissions to the DynamoDB." title: "DynamoDB Key Id" airbyte_secret: true examples: - - "A012345678910EXAMPLE" + - "A012345678910EXAMPLE" secret_access_key: type: "string" description: "The corresponding secret to the access key id." title: "DynamoDB Access Key" airbyte_secret: true examples: - - "a012345678910ABCDEFGH/AbCdEfGhEXAMPLEKEY" + - "a012345678910ABCDEFGH/AbCdEfGhEXAMPLEKEY" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-elasticsearch:0.1.0" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/elasticsearch" @@ -538,7 +516,7 @@ title: "Elasticsearch Connection Configuration" type: "object" required: - - "endpoint" + - "endpoint" additionalProperties: false properties: endpoint: @@ -548,8 +526,7 @@ upsert: type: "boolean" title: "Upsert Records" - description: - "If a primary key identifier is defined in the source, an upsert\ + description: "If a primary key identifier is defined in the source, an upsert\ \ will be performed using the primary key value as the elasticsearch doc\ \ id. Does not support composite primary keys." default: true @@ -558,67 +535,64 @@ type: "object" description: "The type of authentication to be used" oneOf: - - title: "None" - additionalProperties: false - description: "No authentication will be used" - required: - - "method" - properties: - method: - type: "string" - const: "none" - - title: "Api Key/Secret" - additionalProperties: false - description: "Use a api key and secret combination to authenticate" - required: - - "method" - - "apiKeyId" - - "apiKeySecret" - properties: - method: - type: "string" - const: "secret" - apiKeyId: - title: "API Key ID" - description: - "The Key ID to used when accessing an enterprise Elasticsearch\ - \ instance." - type: "string" - apiKeySecret: - title: "API Key Secret" - description: "The secret associated with the API Key ID." - type: "string" - airbyte_secret: true - - title: "Username/Password" - additionalProperties: false - description: "Basic auth header with a username and password" - required: - - "method" - - "username" - - "password" - properties: - method: - type: "string" - const: "basic" - username: - title: "Username" - description: - "Basic auth username to access a secure Elasticsearch\ - \ server" - type: "string" - password: - title: "Password" - description: - "Basic auth password to access a secure Elasticsearch\ - \ server" - type: "string" - airbyte_secret: true + - title: "None" + additionalProperties: false + description: "No authentication will be used" + required: + - "method" + properties: + method: + type: "string" + const: "none" + - title: "Api Key/Secret" + additionalProperties: false + description: "Use a api key and secret combination to authenticate" + required: + - "method" + - "apiKeyId" + - "apiKeySecret" + properties: + method: + type: "string" + const: "secret" + apiKeyId: + title: "API Key ID" + description: "The Key ID to used when accessing an enterprise Elasticsearch\ + \ instance." + type: "string" + apiKeySecret: + title: "API Key Secret" + description: "The secret associated with the API Key ID." + type: "string" + airbyte_secret: true + - title: "Username/Password" + additionalProperties: false + description: "Basic auth header with a username and password" + required: + - "method" + - "username" + - "password" + properties: + method: + type: "string" + const: "basic" + username: + title: "Username" + description: "Basic auth username to access a secure Elasticsearch\ + \ server" + type: "string" + password: + title: "Password" + description: "Basic auth password to access a secure Elasticsearch\ + \ server" + type: "string" + airbyte_secret: true supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" supportsNamespaces: true - dockerImage: "airbyte/destination-gcs:0.1.3" spec: @@ -628,11 +602,11 @@ title: "GCS Destination Spec" type: "object" required: - - "gcs_bucket_name" - - "gcs_bucket_path" - - "gcs_bucket_region" - - "credential" - - "format" + - "gcs_bucket_name" + - "gcs_bucket_path" + - "gcs_bucket_region" + - "credential" + - "format" additionalProperties: false properties: gcs_bucket_name: @@ -640,341 +614,327 @@ type: "string" description: "The name of the GCS bucket." examples: - - "airbyte_sync" + - "airbyte_sync" gcs_bucket_path: description: "Directory under the GCS bucket where data will be written." type: "string" examples: - - "data_sync/test" + - "data_sync/test" gcs_bucket_region: title: "GCS Bucket Region" type: "string" default: "" description: "The region of the GCS bucket." enum: - - "" - - "-- North America --" - - "northamerica-northeast1" - - "us-central1" - - "us-east1" - - "us-east4" - - "us-west1" - - "us-west2" - - "us-west3" - - "us-west4" - - "-- South America --" - - "southamerica-east1" - - "-- Europe --" - - "europe-central2" - - "europe-north1" - - "europe-west1" - - "europe-west2" - - "europe-west3" - - "europe-west4" - - "europe-west6" - - "-- Asia --" - - "asia-east1" - - "asia-east2" - - "asia-northeast1" - - "asia-northeast2" - - "asia-northeast3" - - "asia-south1" - - "asia-south2" - - "asia-southeast1" - - "asia-southeast2" - - "-- Australia --" - - "australia-southeast1" - - "australia-southeast2" - - "-- Multi-regions --" - - "asia" - - "eu" - - "us" - - "-- Dual-regions --" - - "asia1" - - "eur4" - - "nam4" + - "" + - "-- North America --" + - "northamerica-northeast1" + - "us-central1" + - "us-east1" + - "us-east4" + - "us-west1" + - "us-west2" + - "us-west3" + - "us-west4" + - "-- South America --" + - "southamerica-east1" + - "-- Europe --" + - "europe-central2" + - "europe-north1" + - "europe-west1" + - "europe-west2" + - "europe-west3" + - "europe-west4" + - "europe-west6" + - "-- Asia --" + - "asia-east1" + - "asia-east2" + - "asia-northeast1" + - "asia-northeast2" + - "asia-northeast3" + - "asia-south1" + - "asia-south2" + - "asia-southeast1" + - "asia-southeast2" + - "-- Australia --" + - "australia-southeast1" + - "australia-southeast2" + - "-- Multi-regions --" + - "asia" + - "eu" + - "us" + - "-- Dual-regions --" + - "asia1" + - "eur4" + - "nam4" credential: title: "Credential" type: "object" oneOf: - - title: "HMAC key" - required: - - "credential_type" - - "hmac_key_access_id" - - "hmac_key_secret" - properties: - credential_type: - type: "string" - enum: - - "HMAC_KEY" - default: "HMAC_KEY" - hmac_key_access_id: - type: "string" - description: - "HMAC key access ID. When linked to a service account,\ - \ this ID is 61 characters long; when linked to a user account,\ - \ it is 24 characters long." - title: "HMAC Key Access ID" - airbyte_secret: true - examples: - - "1234567890abcdefghij1234" - hmac_key_secret: - type: "string" - description: - "The corresponding secret for the access ID. It is a\ - \ 40-character base-64 encoded string." - title: "HMAC Key Secret" - airbyte_secret: true - examples: - - "1234567890abcdefghij1234567890ABCDEFGHIJ" + - title: "HMAC key" + required: + - "credential_type" + - "hmac_key_access_id" + - "hmac_key_secret" + properties: + credential_type: + type: "string" + enum: + - "HMAC_KEY" + default: "HMAC_KEY" + hmac_key_access_id: + type: "string" + description: "HMAC key access ID. When linked to a service account,\ + \ this ID is 61 characters long; when linked to a user account,\ + \ it is 24 characters long." + title: "HMAC Key Access ID" + airbyte_secret: true + examples: + - "1234567890abcdefghij1234" + hmac_key_secret: + type: "string" + description: "The corresponding secret for the access ID. It is a\ + \ 40-character base-64 encoded string." + title: "HMAC Key Secret" + airbyte_secret: true + examples: + - "1234567890abcdefghij1234567890ABCDEFGHIJ" format: title: "Output Format" type: "object" description: "Output data format" oneOf: - - title: "Avro: Apache Avro" - required: - - "format_type" - - "compression_codec" - properties: - format_type: - type: "string" - enum: - - "Avro" - default: "Avro" - compression_codec: - title: "Compression Codec" - description: - "The compression algorithm used to compress data. Default\ - \ to no compression." - type: "object" - oneOf: - - title: "no compression" - required: - - "codec" - properties: - codec: - type: "string" - enum: - - "no compression" - default: "no compression" - - title: "Deflate" - required: - - "codec" - - "compression_level" - properties: - codec: - type: "string" - enum: - - "Deflate" - default: "Deflate" - compression_level: - title: "Deflate level" - description: - "0: no compression & fastest, 9: best compression\ - \ & slowest." - type: "integer" - default: 0 - minimum: 0 - maximum: 9 - - title: "bzip2" - required: - - "codec" - properties: - codec: - type: "string" - enum: - - "bzip2" - default: "bzip2" - - title: "xz" - required: - - "codec" - - "compression_level" - properties: - codec: - type: "string" - enum: - - "xz" - default: "xz" - compression_level: - title: "Compression level" - description: - "See here for details." - type: "integer" - default: 6 - minimum: 0 - maximum: 9 - - title: "zstandard" - required: - - "codec" - - "compression_level" - properties: - codec: - type: "string" - enum: - - "zstandard" - default: "zstandard" - compression_level: - title: "Compression level" - description: - "Negative levels are 'fast' modes akin to lz4 or\ - \ snappy, levels above 9 are generally for archival purposes,\ - \ and levels above 18 use a lot of memory." - type: "integer" - default: 3 - minimum: -5 - maximum: 22 - include_checksum: - title: "Include checksum" - description: "If true, include a checksum with each data block." - type: "boolean" - default: false - - title: "snappy" - required: - - "codec" - properties: - codec: - type: "string" - enum: - - "snappy" - default: "snappy" - part_size_mb: - title: "Block Size (MB) for GCS multipart upload" - description: - "This is the size of a \"Part\" being buffered in memory.\ - \ It limits the memory usage when writing. Larger values will allow\ - \ to upload a bigger files and improve the speed, but consumes9\ - \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." - type: "integer" - default: 5 - examples: - - 5 - - title: "CSV: Comma-Separated Values" - required: - - "format_type" - - "flattening" - properties: - format_type: - type: "string" - enum: - - "CSV" - default: "CSV" - flattening: - type: "string" - title: "Normalization (Flattening)" - description: - "Whether the input json data should be normalized (flattened)\ - \ in the output CSV. Please refer to docs for details." - default: "No flattening" - enum: - - "No flattening" - - "Root level flattening" - part_size_mb: - title: "Block Size (MB) for GCS multipart upload" - description: - "This is the size of a \"Part\" being buffered in memory.\ - \ It limits the memory usage when writing. Larger values will allow\ - \ to upload a bigger files and improve the speed, but consumes9\ - \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." - type: "integer" - default: 5 - examples: - - 5 - - title: "JSON Lines: newline-delimited JSON" - required: - - "format_type" - properties: - format_type: - type: "string" - enum: - - "JSONL" - default: "JSONL" - part_size_mb: - title: "Block Size (MB) for GCS multipart upload" - description: - "This is the size of a \"Part\" being buffered in memory.\ - \ It limits the memory usage when writing. Larger values will allow\ - \ to upload a bigger files and improve the speed, but consumes9\ - \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." - type: "integer" - default: 5 - examples: - - 5 - - title: "Parquet: Columnar Storage" - required: - - "format_type" - properties: - format_type: - type: "string" - enum: - - "Parquet" - default: "Parquet" - compression_codec: - title: "Compression Codec" - description: "The compression algorithm used to compress data pages." - type: "string" - enum: - - "UNCOMPRESSED" - - "SNAPPY" - - "GZIP" - - "LZO" - - "BROTLI" - - "LZ4" - - "ZSTD" - default: "UNCOMPRESSED" - block_size_mb: - title: "Block Size (Row Group Size) (MB)" - description: - "This is the size of a row group being buffered in memory.\ - \ It limits the memory usage when writing. Larger values will improve\ - \ the IO when reading, but consume more memory when writing. Default:\ - \ 128 MB." - type: "integer" - default: 128 - examples: - - 128 - max_padding_size_mb: - title: "Max Padding Size (MB)" - description: - "Maximum size allowed as padding to align row groups.\ - \ This is also the minimum size of a row group. Default: 8 MB." - type: "integer" - default: 8 - examples: - - 8 - page_size_kb: - title: "Page Size (KB)" - description: - "The page size is for compression. A block is composed\ - \ of pages. A page is the smallest unit that must be read fully\ - \ to access a single record. If this value is too small, the compression\ - \ will deteriorate. Default: 1024 KB." - type: "integer" - default: 1024 - examples: - - 1024 - dictionary_page_size_kb: - title: "Dictionary Page Size (KB)" - description: - "There is one dictionary page per column per row group\ - \ when dictionary encoding is used. The dictionary page size works\ - \ like the page size but for dictionary. Default: 1024 KB." - type: "integer" - default: 1024 - examples: - - 1024 - dictionary_encoding: - title: "Dictionary Encoding" - description: "Default: true." - type: "boolean" - default: true + - title: "Avro: Apache Avro" + required: + - "format_type" + - "compression_codec" + properties: + format_type: + type: "string" + enum: + - "Avro" + default: "Avro" + compression_codec: + title: "Compression Codec" + description: "The compression algorithm used to compress data. Default\ + \ to no compression." + type: "object" + oneOf: + - title: "no compression" + required: + - "codec" + properties: + codec: + type: "string" + enum: + - "no compression" + default: "no compression" + - title: "Deflate" + required: + - "codec" + - "compression_level" + properties: + codec: + type: "string" + enum: + - "Deflate" + default: "Deflate" + compression_level: + title: "Deflate level" + description: "0: no compression & fastest, 9: best compression\ + \ & slowest." + type: "integer" + default: 0 + minimum: 0 + maximum: 9 + - title: "bzip2" + required: + - "codec" + properties: + codec: + type: "string" + enum: + - "bzip2" + default: "bzip2" + - title: "xz" + required: + - "codec" + - "compression_level" + properties: + codec: + type: "string" + enum: + - "xz" + default: "xz" + compression_level: + title: "Compression level" + description: "See here for details." + type: "integer" + default: 6 + minimum: 0 + maximum: 9 + - title: "zstandard" + required: + - "codec" + - "compression_level" + properties: + codec: + type: "string" + enum: + - "zstandard" + default: "zstandard" + compression_level: + title: "Compression level" + description: "Negative levels are 'fast' modes akin to lz4 or\ + \ snappy, levels above 9 are generally for archival purposes,\ + \ and levels above 18 use a lot of memory." + type: "integer" + default: 3 + minimum: -5 + maximum: 22 + include_checksum: + title: "Include checksum" + description: "If true, include a checksum with each data block." + type: "boolean" + default: false + - title: "snappy" + required: + - "codec" + properties: + codec: + type: "string" + enum: + - "snappy" + default: "snappy" + part_size_mb: + title: "Block Size (MB) for GCS multipart upload" + description: "This is the size of a \"Part\" being buffered in memory.\ + \ It limits the memory usage when writing. Larger values will allow\ + \ to upload a bigger files and improve the speed, but consumes9\ + \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." + type: "integer" + default: 5 + examples: + - 5 + - title: "CSV: Comma-Separated Values" + required: + - "format_type" + - "flattening" + properties: + format_type: + type: "string" + enum: + - "CSV" + default: "CSV" + flattening: + type: "string" + title: "Normalization (Flattening)" + description: "Whether the input json data should be normalized (flattened)\ + \ in the output CSV. Please refer to docs for details." + default: "No flattening" + enum: + - "No flattening" + - "Root level flattening" + part_size_mb: + title: "Block Size (MB) for GCS multipart upload" + description: "This is the size of a \"Part\" being buffered in memory.\ + \ It limits the memory usage when writing. Larger values will allow\ + \ to upload a bigger files and improve the speed, but consumes9\ + \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." + type: "integer" + default: 5 + examples: + - 5 + - title: "JSON Lines: newline-delimited JSON" + required: + - "format_type" + properties: + format_type: + type: "string" + enum: + - "JSONL" + default: "JSONL" + part_size_mb: + title: "Block Size (MB) for GCS multipart upload" + description: "This is the size of a \"Part\" being buffered in memory.\ + \ It limits the memory usage when writing. Larger values will allow\ + \ to upload a bigger files and improve the speed, but consumes9\ + \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." + type: "integer" + default: 5 + examples: + - 5 + - title: "Parquet: Columnar Storage" + required: + - "format_type" + properties: + format_type: + type: "string" + enum: + - "Parquet" + default: "Parquet" + compression_codec: + title: "Compression Codec" + description: "The compression algorithm used to compress data pages." + type: "string" + enum: + - "UNCOMPRESSED" + - "SNAPPY" + - "GZIP" + - "LZO" + - "BROTLI" + - "LZ4" + - "ZSTD" + default: "UNCOMPRESSED" + block_size_mb: + title: "Block Size (Row Group Size) (MB)" + description: "This is the size of a row group being buffered in memory.\ + \ It limits the memory usage when writing. Larger values will improve\ + \ the IO when reading, but consume more memory when writing. Default:\ + \ 128 MB." + type: "integer" + default: 128 + examples: + - 128 + max_padding_size_mb: + title: "Max Padding Size (MB)" + description: "Maximum size allowed as padding to align row groups.\ + \ This is also the minimum size of a row group. Default: 8 MB." + type: "integer" + default: 8 + examples: + - 8 + page_size_kb: + title: "Page Size (KB)" + description: "The page size is for compression. A block is composed\ + \ of pages. A page is the smallest unit that must be read fully\ + \ to access a single record. If this value is too small, the compression\ + \ will deteriorate. Default: 1024 KB." + type: "integer" + default: 1024 + examples: + - 1024 + dictionary_page_size_kb: + title: "Dictionary Page Size (KB)" + description: "There is one dictionary page per column per row group\ + \ when dictionary encoding is used. The dictionary page size works\ + \ like the page size but for dictionary. Default: 1024 KB." + type: "integer" + default: 1024 + examples: + - 1024 + dictionary_encoding: + title: "Dictionary Encoding" + description: "Default: true." + type: "boolean" + default: true supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" $schema: "http://json-schema.org/draft-07/schema#" - dockerImage: "airbyte/destination-pubsub:0.1.1" spec: @@ -984,9 +944,9 @@ title: "Google PubSub Destination Spec" type: "object" required: - - "project_id" - - "topic_id" - - "credentials_json" + - "project_id" + - "topic_id" + - "credentials_json" additionalProperties: true properties: project_id: @@ -999,8 +959,7 @@ title: "PubSub Topic ID" credentials_json: type: "string" - description: - "The contents of the JSON service account key. Check out the\ + description: "The contents of the JSON service account key. Check out the\ \ docs if you need help generating this key." title: "Credentials JSON" @@ -1009,7 +968,7 @@ supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "append" + - "append" - dockerImage: "airbyte/destination-kafka:0.1.2" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/kafka" @@ -1018,32 +977,31 @@ title: "Kafka Destination Spec" type: "object" required: - - "bootstrap_servers" - - "topic_pattern" - - "protocol" - - "acks" - - "enable_idempotence" - - "compression_type" - - "batch_size" - - "linger_ms" - - "max_in_flight_requests_per_connection" - - "client_dns_lookup" - - "buffer_memory" - - "max_request_size" - - "retries" - - "socket_connection_setup_timeout_ms" - - "socket_connection_setup_timeout_max_ms" - - "max_block_ms" - - "request_timeout_ms" - - "delivery_timeout_ms" - - "send_buffer_bytes" - - "receive_buffer_bytes" + - "bootstrap_servers" + - "topic_pattern" + - "protocol" + - "acks" + - "enable_idempotence" + - "compression_type" + - "batch_size" + - "linger_ms" + - "max_in_flight_requests_per_connection" + - "client_dns_lookup" + - "buffer_memory" + - "max_request_size" + - "retries" + - "socket_connection_setup_timeout_ms" + - "socket_connection_setup_timeout_max_ms" + - "max_block_ms" + - "request_timeout_ms" + - "delivery_timeout_ms" + - "send_buffer_bytes" + - "receive_buffer_bytes" additionalProperties: true properties: bootstrap_servers: title: "Bootstrap servers" - description: - "A list of host/port pairs to use for establishing the initial\ + description: "A list of host/port pairs to use for establishing the initial\ \ connection to the Kafka cluster. The client will make use of all servers\ \ irrespective of which servers are specified here for bootstrapping—this\ \ list only impacts the initial hosts used to discover the full set of\ @@ -1054,24 +1012,23 @@ \ though, in case a server is down)." type: "string" examples: - - "kafka-broker1:9092,kafka-broker2:9092" + - "kafka-broker1:9092,kafka-broker2:9092" topic_pattern: title: "Topic pattern" - description: - "Topic pattern in which the records will be sent. You can use\ + description: "Topic pattern in which the records will be sent. You can use\ \ patterns like '{namespace}' and/or '{stream}' to send the message to\ \ a specific topic based on these values. Notice that the topic name will\ \ be transformed to a standard naming convention." type: "string" examples: - - "sample.topic" - - "{namespace}.{stream}.sample" + - "sample.topic" + - "{namespace}.{stream}.sample" test_topic: title: "Test topic" description: "Topic to test if Airbyte can produce messages." type: "string" examples: - - "test.topic" + - "test.topic" sync_producer: title: "Sync producer" description: "Wait synchronously until the record has been sent to Kafka." @@ -1082,99 +1039,92 @@ type: "object" description: "Protocol used to communicate with brokers." oneOf: - - title: "PLAINTEXT" - required: - - "security_protocol" - properties: - security_protocol: - type: "string" - enum: - - "PLAINTEXT" - default: "PLAINTEXT" - - title: "SASL PLAINTEXT" - required: - - "security_protocol" - - "sasl_mechanism" - - "sasl_jaas_config" - properties: - security_protocol: - type: "string" - enum: - - "SASL_PLAINTEXT" - default: "SASL_PLAINTEXT" - sasl_mechanism: - title: "SASL mechanism" - description: - "SASL mechanism used for client connections. This may\ - \ be any mechanism for which a security provider is available." - type: "string" - default: "PLAIN" - enum: - - "PLAIN" - sasl_jaas_config: - title: "SASL JAAS config" - description: - "JAAS login context parameters for SASL connections in\ - \ the format used by JAAS configuration files." - type: "string" - default: "" - airbyte_secret: true - - title: "SASL SSL" - required: - - "security_protocol" - - "sasl_mechanism" - - "sasl_jaas_config" - properties: - security_protocol: - type: "string" - enum: - - "SASL_SSL" - default: "SASL_SSL" - sasl_mechanism: - title: "SASL mechanism" - description: - "SASL mechanism used for client connections. This may\ - \ be any mechanism for which a security provider is available." - type: "string" - default: "GSSAPI" - enum: - - "GSSAPI" - - "OAUTHBEARER" - - "SCRAM-SHA-256" - sasl_jaas_config: - title: "SASL JAAS config" - description: - "JAAS login context parameters for SASL connections in\ - \ the format used by JAAS configuration files." - type: "string" - default: "" - airbyte_secret: true + - title: "PLAINTEXT" + required: + - "security_protocol" + properties: + security_protocol: + type: "string" + enum: + - "PLAINTEXT" + default: "PLAINTEXT" + - title: "SASL PLAINTEXT" + required: + - "security_protocol" + - "sasl_mechanism" + - "sasl_jaas_config" + properties: + security_protocol: + type: "string" + enum: + - "SASL_PLAINTEXT" + default: "SASL_PLAINTEXT" + sasl_mechanism: + title: "SASL mechanism" + description: "SASL mechanism used for client connections. This may\ + \ be any mechanism for which a security provider is available." + type: "string" + default: "PLAIN" + enum: + - "PLAIN" + sasl_jaas_config: + title: "SASL JAAS config" + description: "JAAS login context parameters for SASL connections in\ + \ the format used by JAAS configuration files." + type: "string" + default: "" + airbyte_secret: true + - title: "SASL SSL" + required: + - "security_protocol" + - "sasl_mechanism" + - "sasl_jaas_config" + properties: + security_protocol: + type: "string" + enum: + - "SASL_SSL" + default: "SASL_SSL" + sasl_mechanism: + title: "SASL mechanism" + description: "SASL mechanism used for client connections. This may\ + \ be any mechanism for which a security provider is available." + type: "string" + default: "GSSAPI" + enum: + - "GSSAPI" + - "OAUTHBEARER" + - "SCRAM-SHA-256" + sasl_jaas_config: + title: "SASL JAAS config" + description: "JAAS login context parameters for SASL connections in\ + \ the format used by JAAS configuration files." + type: "string" + default: "" + airbyte_secret: true client_id: title: "Client ID" - description: - "An id string to pass to the server when making requests. The\ + description: "An id string to pass to the server when making requests. The\ \ purpose of this is to be able to track the source of requests beyond\ \ just ip/port by allowing a logical application name to be included in\ \ server-side request logging." type: "string" examples: - - "airbyte-producer" + - "airbyte-producer" acks: title: "ACKs" - description: - "The number of acknowledgments the producer requires the leader\ + description: "The number of acknowledgments the producer requires the leader\ \ to have received before considering a request complete. This controls\ \ the durability of records that are sent." type: "string" default: "1" enum: - - "0" - - "1" - - "all" + - "0" + - "1" + - "all" enable_idempotence: title: "Enable idempotence" - description: - "When set to 'true', the producer will ensure that exactly\ + description: "When set to 'true', the producer will ensure that exactly\ \ one copy of each message is written in the stream. If 'false', producer\ \ retries due to broker failures, etc., may write duplicates of the retried\ \ message in the stream." @@ -1186,39 +1136,35 @@ type: "string" default: "none" enum: - - "none" - - "gzip" - - "snappy" - - "lz4" - - "zstd" + - "none" + - "gzip" + - "snappy" + - "lz4" + - "zstd" batch_size: title: "Batch size" - description: - "The producer will attempt to batch records together into fewer\ + description: "The producer will attempt to batch records together into fewer\ \ requests whenever multiple records are being sent to the same partition." type: "integer" examples: - - 16384 + - 16384 linger_ms: title: "Linger ms" - description: - "The producer groups together any records that arrive in between\ + description: "The producer groups together any records that arrive in between\ \ request transmissions into a single batched request." type: "string" examples: - - 0 + - 0 max_in_flight_requests_per_connection: title: "Max in flight requests per connection" - description: - "The maximum number of unacknowledged requests the client will\ + description: "The maximum number of unacknowledged requests the client will\ \ send on a single connection before blocking." type: "integer" examples: - - 5 + - 5 client_dns_lookup: title: "Client DNS lookup" - description: - "Controls how the client uses DNS lookups. If set to use_all_dns_ips,\ + description: "Controls how the client uses DNS lookups. If set to use_all_dns_ips,\ \ connect to each returned IP address in sequence until a successful connection\ \ is established. After a disconnection, the next IP is used. Once all\ \ IPs have been used once, the client resolves the IP(s) from the hostname\ @@ -1230,14 +1176,13 @@ type: "string" default: "use_all_dns_ips" enum: - - "default" - - "use_all_dns_ips" - - "resolve_canonical_bootstrap_servers_only" - - "use_all_dns_ips" + - "default" + - "use_all_dns_ips" + - "resolve_canonical_bootstrap_servers_only" + - "use_all_dns_ips" buffer_memory: title: "Buffer memory" - description: - "The total bytes of memory the producer can use to buffer records\ + description: "The total bytes of memory the producer can use to buffer records\ \ waiting to be sent to the server." type: "string" examples: 33554432 @@ -1246,80 +1191,72 @@ description: "The maximum size of a request in bytes." type: "integer" examples: - - 1048576 + - 1048576 retries: title: "Retries" - description: - "Setting a value greater than zero will cause the client to\ + description: "Setting a value greater than zero will cause the client to\ \ resend any record whose send fails with a potentially transient error." type: "integer" examples: - - 2147483647 + - 2147483647 socket_connection_setup_timeout_ms: title: "Socket connection setup timeout" - description: - "The amount of time the client will wait for the socket connection\ + description: "The amount of time the client will wait for the socket connection\ \ to be established." type: "string" examples: - - 10000 + - 10000 socket_connection_setup_timeout_max_ms: title: "Socket connection setup max timeout" - description: - "The maximum amount of time the client will wait for the socket\ + description: "The maximum amount of time the client will wait for the socket\ \ connection to be established. The connection setup timeout will increase\ \ exponentially for each consecutive connection failure up to this maximum." type: "string" examples: - - 30000 + - 30000 max_block_ms: title: "Max block ms" - description: - "The configuration controls how long the KafkaProducer's send(),\ + description: "The configuration controls how long the KafkaProducer's send(),\ \ partitionsFor(), initTransactions(), sendOffsetsToTransaction(), commitTransaction()\ \ and abortTransaction() methods will block." type: "string" examples: - - 60000 + - 60000 request_timeout_ms: title: "Request timeout" - description: - "The configuration controls the maximum amount of time the\ + description: "The configuration controls the maximum amount of time the\ \ client will wait for the response of a request. If the response is not\ \ received before the timeout elapses the client will resend the request\ \ if necessary or fail the request if retries are exhausted." type: "integer" examples: - - 30000 + - 30000 delivery_timeout_ms: title: "Delivery timeout" - description: - "An upper bound on the time to report success or failure after\ + description: "An upper bound on the time to report success or failure after\ \ a call to 'send()' returns." type: "integer" examples: - - 120000 + - 120000 send_buffer_bytes: title: "Send buffer bytes" - description: - "The size of the TCP send buffer (SO_SNDBUF) to use when sending\ + description: "The size of the TCP send buffer (SO_SNDBUF) to use when sending\ \ data. If the value is -1, the OS default will be used." type: "integer" examples: - - 131072 + - 131072 receive_buffer_bytes: title: "Receive buffer bytes" - description: - "The size of the TCP receive buffer (SO_RCVBUF) to use when\ + description: "The size of the TCP receive buffer (SO_RCVBUF) to use when\ \ reading data. If the value is -1, the OS default will be used." type: "integer" examples: - - 32768 + - 32768 supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "append" + - "append" - dockerImage: "airbyte/destination-csv:0.2.8" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/local-csv" @@ -1328,25 +1265,24 @@ title: "CSV Destination Spec" type: "object" required: - - "destination_path" + - "destination_path" additionalProperties: false properties: destination_path: - description: - "Path to the directory where csv files will be written. The\ + description: "Path to the directory where csv files will be written. The\ \ destination uses the local mount \"/local\" and any data files will\ \ be placed inside that local mount. For more information check out our\ \ docs" type: "string" examples: - - "/local" + - "/local" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-local-json:0.2.8" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/local-json" @@ -1355,24 +1291,23 @@ title: "Local Json Destination Spec" type: "object" required: - - "destination_path" + - "destination_path" additionalProperties: false properties: destination_path: - description: - "Path to the directory where json files will be written. The\ + description: "Path to the directory where json files will be written. The\ \ files will be placed inside that local mount. For more information check\ \ out our docs" type: "string" examples: - - "/json_data" + - "/json_data" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-mssql:0.1.11" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/mssql" @@ -1381,11 +1316,11 @@ title: "MS SQL Server Destination Spec" type: "object" required: - - "host" - - "port" - - "username" - - "database" - - "schema" + - "host" + - "port" + - "username" + - "database" + - "schema" additionalProperties: true properties: host: @@ -1401,7 +1336,7 @@ maximum: 65536 default: 1433 examples: - - "1433" + - "1433" order: 1 database: title: "DB Name" @@ -1410,13 +1345,12 @@ order: 2 schema: title: "Default Schema" - description: - "The default schema tables are written to if the source does\ + description: "The default schema tables are written to if the source does\ \ not specify a namespace. The usual value for this field is \"public\"\ ." type: "string" examples: - - "public" + - "public" default: "public" order: 3 username: @@ -1436,169 +1370,159 @@ description: "Encryption method to use when communicating with the database" order: 6 oneOf: - - title: "Unencrypted" - additionalProperties: false - description: "Data transfer will not be encrypted." - required: - - "ssl_method" - type: "object" - properties: - ssl_method: - type: "string" - enum: - - "unencrypted" - default: "unencrypted" - - title: "Encrypted (trust server certificate)" - additionalProperties: false - description: - "Use the cert provided by the server without verification.\ - \ (For testing purposes only!)" - required: - - "ssl_method" - type: "object" - properties: - ssl_method: - type: "string" - enum: - - "encrypted_trust_server_certificate" - default: "encrypted_trust_server_certificate" - - title: "Encrypted (verify certificate)" - additionalProperties: false - description: "Verify and use the cert provided by the server." - required: - - "ssl_method" - - "trustStoreName" - - "trustStorePassword" - type: "object" - properties: - ssl_method: - type: "string" - enum: - - "encrypted_verify_certificate" - default: "encrypted_verify_certificate" - hostNameInCertificate: - title: "Host Name In Certificate" - type: "string" - description: - "Specifies the host name of the server. The value of\ - \ this property must match the subject property of the certificate." - order: 7 + - title: "Unencrypted" + additionalProperties: false + description: "Data transfer will not be encrypted." + required: + - "ssl_method" + type: "object" + properties: + ssl_method: + type: "string" + enum: + - "unencrypted" + default: "unencrypted" + - title: "Encrypted (trust server certificate)" + additionalProperties: false + description: "Use the cert provided by the server without verification.\ + \ (For testing purposes only!)" + required: + - "ssl_method" + type: "object" + properties: + ssl_method: + type: "string" + enum: + - "encrypted_trust_server_certificate" + default: "encrypted_trust_server_certificate" + - title: "Encrypted (verify certificate)" + additionalProperties: false + description: "Verify and use the cert provided by the server." + required: + - "ssl_method" + - "trustStoreName" + - "trustStorePassword" + type: "object" + properties: + ssl_method: + type: "string" + enum: + - "encrypted_verify_certificate" + default: "encrypted_verify_certificate" + hostNameInCertificate: + title: "Host Name In Certificate" + type: "string" + description: "Specifies the host name of the server. The value of\ + \ this property must match the subject property of the certificate." + order: 7 tunnel_method: type: "object" title: "SSH Tunnel Method" - description: - "Whether to initiate an SSH tunnel before connecting to the\ + description: "Whether to initiate an SSH tunnel before connecting to the\ \ database, and if so, which kind of authentication to use." oneOf: - - title: "No Tunnel" - required: - - "tunnel_method" - properties: - tunnel_method: - description: "No ssh tunnel needed to connect to database" - type: "string" - const: "NO_TUNNEL" - order: 0 - - title: "SSH Key Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "ssh_key" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and ssh key" - type: "string" - const: "SSH_KEY_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host." - type: "string" - order: 3 - ssh_key: - title: "SSH Private Key" - description: - "OS-level user account ssh key credentials in RSA PEM\ - \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" - type: "string" - airbyte_secret: true - multiline: true - order: 4 - - title: "Password Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "tunnel_user_password" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and password authentication" - type: "string" - const: "SSH_PASSWORD_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host" - type: "string" - order: 3 - tunnel_user_password: - title: "Password" - description: "OS-level password for logging into the jump server host" - type: "string" - airbyte_secret: true - order: 4 + - title: "No Tunnel" + required: + - "tunnel_method" + properties: + tunnel_method: + description: "No ssh tunnel needed to connect to database" + type: "string" + const: "NO_TUNNEL" + order: 0 + - title: "SSH Key Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "ssh_key" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and ssh key" + type: "string" + const: "SSH_KEY_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host." + type: "string" + order: 3 + ssh_key: + title: "SSH Private Key" + description: "OS-level user account ssh key credentials in RSA PEM\ + \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" + type: "string" + airbyte_secret: true + multiline: true + order: 4 + - title: "Password Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "tunnel_user_password" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and password authentication" + type: "string" + const: "SSH_PASSWORD_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host" + type: "string" + order: 3 + tunnel_user_password: + title: "Password" + description: "OS-level password for logging into the jump server host" + type: "string" + airbyte_secret: true + order: 4 supportsIncremental: true supportsNormalization: true supportsDBT: true supported_destination_sync_modes: - - "overwrite" - - "append" - - "append_dedup" + - "overwrite" + - "append" + - "append_dedup" - dockerImage: "airbyte/destination-meilisearch:0.2.10" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/meilisearch" @@ -1607,7 +1531,7 @@ title: "MeiliSearch Destination Spec" type: "object" required: - - "host" + - "host" additionalProperties: true properties: host: @@ -1625,8 +1549,8 @@ supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-mongodb:0.1.2" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/mongodb" @@ -1635,94 +1559,91 @@ title: "MongoDB Destination Spec" type: "object" required: - - "database" - - "auth_type" + - "database" + - "auth_type" additionalProperties: true properties: instance_type: - description: - "MongoDb instance to connect to. For MongoDB Atlas and Replica\ + description: "MongoDb instance to connect to. For MongoDB Atlas and Replica\ \ Set TLS connection is used by default." title: "MongoDb instance type" type: "object" order: 0 oneOf: - - title: "Standalone MongoDb Instance" - required: - - "instance" - - "host" - - "port" - properties: - instance: - type: "string" - enum: - - "standalone" - default: "standalone" - host: - title: "Host" - type: "string" - description: "Host of a Mongo database to be replicated." - order: 0 - port: - title: "Port" - type: "integer" - description: "Port of a Mongo database to be replicated." - minimum: 0 - maximum: 65536 - default: 27017 - examples: - - "27017" - order: 1 - tls: - title: "TLS connection" - type: "boolean" - description: - "Indicates whether TLS encryption protocol will be used\ - \ to connect to MongoDB. It is recommended to use TLS connection\ - \ if possible. For more information see documentation." - default: false - order: 2 - - title: "Replica Set" - required: - - "instance" - - "server_addresses" - properties: - instance: - type: "string" - enum: - - "replica" - default: "replica" - server_addresses: - title: "Server addresses" - type: "string" - description: - "The members of a replica set. Please specify `host`:`port`\ - \ of each member seperated by comma." - examples: - - "host1:27017,host2:27017,host3:27017" - order: 0 - replica_set: - title: "Replica Set" - type: "string" - description: "A replica set name." - order: 1 - - title: "MongoDB Atlas" - additionalProperties: false - required: - - "instance" - - "cluster_url" - properties: - instance: - type: "string" - enum: - - "atlas" - default: "atlas" - cluster_url: - title: "Cluster URL" - type: "string" - description: "URL of a cluster to connect to." - order: 0 + - title: "Standalone MongoDb Instance" + required: + - "instance" + - "host" + - "port" + properties: + instance: + type: "string" + enum: + - "standalone" + default: "standalone" + host: + title: "Host" + type: "string" + description: "Host of a Mongo database to be replicated." + order: 0 + port: + title: "Port" + type: "integer" + description: "Port of a Mongo database to be replicated." + minimum: 0 + maximum: 65536 + default: 27017 + examples: + - "27017" + order: 1 + tls: + title: "TLS connection" + type: "boolean" + description: "Indicates whether TLS encryption protocol will be used\ + \ to connect to MongoDB. It is recommended to use TLS connection\ + \ if possible. For more information see documentation." + default: false + order: 2 + - title: "Replica Set" + required: + - "instance" + - "server_addresses" + properties: + instance: + type: "string" + enum: + - "replica" + default: "replica" + server_addresses: + title: "Server addresses" + type: "string" + description: "The members of a replica set. Please specify `host`:`port`\ + \ of each member seperated by comma." + examples: + - "host1:27017,host2:27017,host3:27017" + order: 0 + replica_set: + title: "Replica Set" + type: "string" + description: "A replica set name." + order: 1 + - title: "MongoDB Atlas" + additionalProperties: false + required: + - "instance" + - "cluster_url" + properties: + instance: + type: "string" + enum: + - "atlas" + default: "atlas" + cluster_url: + title: "Cluster URL" + type: "string" + description: "URL of a cluster to connect to." + order: 0 database: title: "DB Name" description: "Name of the database." @@ -1733,45 +1654,45 @@ type: "object" description: "Authorization type." oneOf: - - title: "None" - additionalProperties: false - description: "None." - required: - - "authorization" - type: "object" - properties: - authorization: - type: "string" - const: "none" - - title: "Login/Password" - additionalProperties: false - description: "Login/Password." - required: - - "authorization" - - "username" - - "password" - type: "object" - properties: - authorization: - type: "string" - const: "login/password" - username: - title: "User" - description: "Username to use to access the database." - type: "string" - order: 1 - password: - title: "Password" - description: "Password associated with the username." - type: "string" - airbyte_secret: true - order: 2 + - title: "None" + additionalProperties: false + description: "None." + required: + - "authorization" + type: "object" + properties: + authorization: + type: "string" + const: "none" + - title: "Login/Password" + additionalProperties: false + description: "Login/Password." + required: + - "authorization" + - "username" + - "password" + type: "object" + properties: + authorization: + type: "string" + const: "login/password" + username: + title: "User" + description: "Username to use to access the database." + type: "string" + order: 1 + password: + title: "Password" + description: "Password associated with the username." + type: "string" + airbyte_secret: true + order: 2 supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-mysql:0.1.14" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/mysql" @@ -1780,10 +1701,10 @@ title: "MySQL Destination Spec" type: "object" required: - - "host" - - "port" - - "username" - - "database" + - "host" + - "port" + - "username" + - "database" additionalProperties: true properties: host: @@ -1799,7 +1720,7 @@ maximum: 65536 default: 3306 examples: - - "3306" + - "3306" order: 1 database: title: "DB Name" @@ -1826,118 +1747,110 @@ tunnel_method: type: "object" title: "SSH Tunnel Method" - description: - "Whether to initiate an SSH tunnel before connecting to the\ + description: "Whether to initiate an SSH tunnel before connecting to the\ \ database, and if so, which kind of authentication to use." oneOf: - - title: "No Tunnel" - required: - - "tunnel_method" - properties: - tunnel_method: - description: "No ssh tunnel needed to connect to database" - type: "string" - const: "NO_TUNNEL" - order: 0 - - title: "SSH Key Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "ssh_key" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and ssh key" - type: "string" - const: "SSH_KEY_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host." - type: "string" - order: 3 - ssh_key: - title: "SSH Private Key" - description: - "OS-level user account ssh key credentials in RSA PEM\ - \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" - type: "string" - airbyte_secret: true - multiline: true - order: 4 - - title: "Password Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "tunnel_user_password" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and password authentication" - type: "string" - const: "SSH_PASSWORD_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host" - type: "string" - order: 3 - tunnel_user_password: - title: "Password" - description: "OS-level password for logging into the jump server host" - type: "string" - airbyte_secret: true - order: 4 + - title: "No Tunnel" + required: + - "tunnel_method" + properties: + tunnel_method: + description: "No ssh tunnel needed to connect to database" + type: "string" + const: "NO_TUNNEL" + order: 0 + - title: "SSH Key Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "ssh_key" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and ssh key" + type: "string" + const: "SSH_KEY_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host." + type: "string" + order: 3 + ssh_key: + title: "SSH Private Key" + description: "OS-level user account ssh key credentials in RSA PEM\ + \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" + type: "string" + airbyte_secret: true + multiline: true + order: 4 + - title: "Password Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "tunnel_user_password" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and password authentication" + type: "string" + const: "SSH_PASSWORD_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host" + type: "string" + order: 3 + tunnel_user_password: + title: "Password" + description: "OS-level password for logging into the jump server host" + type: "string" + airbyte_secret: true + order: 4 supportsIncremental: true supportsNormalization: true supportsDBT: true supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-oracle:0.1.12" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/oracle" @@ -1946,10 +1859,10 @@ title: "Oracle Destination Spec" type: "object" required: - - "host" - - "port" - - "username" - - "sid" + - "host" + - "port" + - "username" + - "sid" additionalProperties: true properties: host: @@ -1965,7 +1878,7 @@ maximum: 65536 default: 1521 examples: - - "1521" + - "1521" order: 1 sid: title: "SID" @@ -1974,8 +1887,7 @@ order: 2 username: title: "User" - description: - "Username to use to access the database. This user must have\ + description: "Username to use to access the database. This user must have\ \ CREATE USER privileges in the database." type: "string" order: 3 @@ -1987,15 +1899,14 @@ order: 4 schema: title: "Default Schema" - description: - "The default schema tables are written to if the source does\ + description: "The default schema tables are written to if the source does\ \ not specify a namespace. The usual value for this field is \"airbyte\"\ . In Oracle, schemas and users are the same thing, so the \"user\" parameter\ \ is used as the login credentials and this is used for the default Airbyte\ \ message schema." type: "string" examples: - - "airbyte" + - "airbyte" default: "airbyte" order: 5 encryption: @@ -2004,180 +1915,169 @@ description: "Encryption method to use when communicating with the database" order: 6 oneOf: - - title: "Unencrypted" - additionalProperties: false - description: "Data transfer will not be encrypted." - required: - - "encryption_method" - properties: - encryption_method: - type: "string" - const: "unencrypted" - enum: - - "unencrypted" - default: "unencrypted" - - title: "Native Network Ecryption (NNE)" - additionalProperties: false - description: - "Native network encryption gives you the ability to encrypt\ - \ database connections, without the configuration overhead of TCP/IP\ - \ and SSL/TLS and without the need to open and listen on different ports." - required: - - "encryption_method" - properties: - encryption_method: - type: "string" - const: "client_nne" - enum: - - "client_nne" - default: "client_nne" - encryption_algorithm: - type: "string" - description: - "This parameter defines the encryption algorithm to be\ - \ used" - title: "Encryption Algorithm" - default: "AES256" - enum: - - "AES256" - - "RC4_56" - - "3DES168" - - title: "TLS Encrypted (verify certificate)" - additionalProperties: false - description: "Verify and use the cert provided by the server." - required: - - "encryption_method" - - "ssl_certificate" - properties: - encryption_method: - type: "string" - const: "encrypted_verify_certificate" - enum: - - "encrypted_verify_certificate" - default: "encrypted_verify_certificate" - ssl_certificate: - title: "SSL PEM file" - description: - "Privacy Enhanced Mail (PEM) files are concatenated certificate\ - \ containers frequently used in certificate installations" - type: "string" - airbyte_secret: true - multiline: true + - title: "Unencrypted" + additionalProperties: false + description: "Data transfer will not be encrypted." + required: + - "encryption_method" + properties: + encryption_method: + type: "string" + const: "unencrypted" + enum: + - "unencrypted" + default: "unencrypted" + - title: "Native Network Ecryption (NNE)" + additionalProperties: false + description: "Native network encryption gives you the ability to encrypt\ + \ database connections, without the configuration overhead of TCP/IP\ + \ and SSL/TLS and without the need to open and listen on different ports." + required: + - "encryption_method" + properties: + encryption_method: + type: "string" + const: "client_nne" + enum: + - "client_nne" + default: "client_nne" + encryption_algorithm: + type: "string" + description: "This parameter defines the encryption algorithm to be\ + \ used" + title: "Encryption Algorithm" + default: "AES256" + enum: + - "AES256" + - "RC4_56" + - "3DES168" + - title: "TLS Encrypted (verify certificate)" + additionalProperties: false + description: "Verify and use the cert provided by the server." + required: + - "encryption_method" + - "ssl_certificate" + properties: + encryption_method: + type: "string" + const: "encrypted_verify_certificate" + enum: + - "encrypted_verify_certificate" + default: "encrypted_verify_certificate" + ssl_certificate: + title: "SSL PEM file" + description: "Privacy Enhanced Mail (PEM) files are concatenated certificate\ + \ containers frequently used in certificate installations" + type: "string" + airbyte_secret: true + multiline: true tunnel_method: type: "object" title: "SSH Tunnel Method" - description: - "Whether to initiate an SSH tunnel before connecting to the\ + description: "Whether to initiate an SSH tunnel before connecting to the\ \ database, and if so, which kind of authentication to use." oneOf: - - title: "No Tunnel" - required: - - "tunnel_method" - properties: - tunnel_method: - description: "No ssh tunnel needed to connect to database" - type: "string" - const: "NO_TUNNEL" - order: 0 - - title: "SSH Key Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "ssh_key" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and ssh key" - type: "string" - const: "SSH_KEY_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host." - type: "string" - order: 3 - ssh_key: - title: "SSH Private Key" - description: - "OS-level user account ssh key credentials in RSA PEM\ - \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" - type: "string" - airbyte_secret: true - multiline: true - order: 4 - - title: "Password Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "tunnel_user_password" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and password authentication" - type: "string" - const: "SSH_PASSWORD_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host" - type: "string" - order: 3 - tunnel_user_password: - title: "Password" - description: "OS-level password for logging into the jump server host" - type: "string" - airbyte_secret: true - order: 4 + - title: "No Tunnel" + required: + - "tunnel_method" + properties: + tunnel_method: + description: "No ssh tunnel needed to connect to database" + type: "string" + const: "NO_TUNNEL" + order: 0 + - title: "SSH Key Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "ssh_key" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and ssh key" + type: "string" + const: "SSH_KEY_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host." + type: "string" + order: 3 + ssh_key: + title: "SSH Private Key" + description: "OS-level user account ssh key credentials in RSA PEM\ + \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" + type: "string" + airbyte_secret: true + multiline: true + order: 4 + - title: "Password Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "tunnel_user_password" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and password authentication" + type: "string" + const: "SSH_PASSWORD_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host" + type: "string" + order: 3 + tunnel_user_password: + title: "Password" + description: "OS-level password for logging into the jump server host" + type: "string" + airbyte_secret: true + order: 4 supportsIncremental: true supportsNormalization: false supportsDBT: true supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-postgres:0.3.12" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/postgres" @@ -2186,11 +2086,11 @@ title: "Postgres Destination Spec" type: "object" required: - - "host" - - "port" - - "username" - - "database" - - "schema" + - "host" + - "port" + - "username" + - "database" + - "schema" additionalProperties: true properties: host: @@ -2206,7 +2106,7 @@ maximum: 65536 default: 5432 examples: - - "5432" + - "5432" order: 1 database: title: "DB Name" @@ -2215,13 +2115,12 @@ order: 2 schema: title: "Default Schema" - description: - "The default schema tables are written to if the source does\ + description: "The default schema tables are written to if the source does\ \ not specify a namespace. The usual value for this field is \"public\"\ ." type: "string" examples: - - "public" + - "public" default: "public" order: 3 username: @@ -2244,119 +2143,111 @@ tunnel_method: type: "object" title: "SSH Tunnel Method" - description: - "Whether to initiate an SSH tunnel before connecting to the\ + description: "Whether to initiate an SSH tunnel before connecting to the\ \ database, and if so, which kind of authentication to use." oneOf: - - title: "No Tunnel" - required: - - "tunnel_method" - properties: - tunnel_method: - description: "No ssh tunnel needed to connect to database" - type: "string" - const: "NO_TUNNEL" - order: 0 - - title: "SSH Key Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "ssh_key" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and ssh key" - type: "string" - const: "SSH_KEY_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host." - type: "string" - order: 3 - ssh_key: - title: "SSH Private Key" - description: - "OS-level user account ssh key credentials in RSA PEM\ - \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" - type: "string" - airbyte_secret: true - multiline: true - order: 4 - - title: "Password Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "tunnel_user_password" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and password authentication" - type: "string" - const: "SSH_PASSWORD_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host" - type: "string" - order: 3 - tunnel_user_password: - title: "Password" - description: "OS-level password for logging into the jump server host" - type: "string" - airbyte_secret: true - order: 4 + - title: "No Tunnel" + required: + - "tunnel_method" + properties: + tunnel_method: + description: "No ssh tunnel needed to connect to database" + type: "string" + const: "NO_TUNNEL" + order: 0 + - title: "SSH Key Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "ssh_key" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and ssh key" + type: "string" + const: "SSH_KEY_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host." + type: "string" + order: 3 + ssh_key: + title: "SSH Private Key" + description: "OS-level user account ssh key credentials in RSA PEM\ + \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" + type: "string" + airbyte_secret: true + multiline: true + order: 4 + - title: "Password Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "tunnel_user_password" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and password authentication" + type: "string" + const: "SSH_PASSWORD_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host" + type: "string" + order: 3 + tunnel_user_password: + title: "Password" + description: "OS-level password for logging into the jump server host" + type: "string" + airbyte_secret: true + order: 4 supportsIncremental: true supportsNormalization: true supportsDBT: true supported_destination_sync_modes: - - "overwrite" - - "append" - - "append_dedup" + - "overwrite" + - "append" + - "append_dedup" - dockerImage: "airbyte/destination-pulsar:0.1.0" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/pulsar" @@ -2365,30 +2256,29 @@ title: "Pulsar Destination Spec" type: "object" required: - - "brokers" - - "use_tls" - - "topic_type" - - "topic_tenant" - - "topic_namespace" - - "topic_pattern" - - "compression_type" - - "send_timeout_ms" - - "max_pending_messages" - - "max_pending_messages_across_partitions" - - "batching_enabled" - - "batching_max_messages" - - "batching_max_publish_delay" - - "block_if_queue_full" + - "brokers" + - "use_tls" + - "topic_type" + - "topic_tenant" + - "topic_namespace" + - "topic_pattern" + - "compression_type" + - "send_timeout_ms" + - "max_pending_messages" + - "max_pending_messages_across_partitions" + - "batching_enabled" + - "batching_max_messages" + - "batching_max_publish_delay" + - "block_if_queue_full" additionalProperties: true properties: brokers: title: "Pulsar brokers" - description: - "A list of host/port pairs to use for establishing the initial\ + description: "A list of host/port pairs to use for establishing the initial\ \ connection to the Pulsar cluster." type: "string" examples: - - "broker1:6650,broker2:6650" + - "broker1:6650,broker2:6650" use_tls: title: "Use TLS" description: "Whether to use TLS encryption on the connection." @@ -2396,8 +2286,7 @@ default: false topic_type: title: "Topic type" - description: - "It identifies type of topic. Pulsar supports two kind of topics:\ + description: "It identifies type of topic. Pulsar supports two kind of topics:\ \ persistent and non-persistent. In persistent topic, all messages are\ \ durably persisted on disk (that means on multiple disks unless the broker\ \ is standalone), whereas non-persistent topic does not persist message\ @@ -2405,52 +2294,48 @@ type: "string" default: "persistent" enum: - - "persistent" - - "non-persistent" + - "persistent" + - "non-persistent" topic_tenant: title: "Topic tenant" - description: - "The topic tenant within the instance. Tenants are essential\ + description: "The topic tenant within the instance. Tenants are essential\ \ to multi-tenancy in Pulsar, and spread across clusters." type: "string" default: "public" examples: - - "public" + - "public" topic_namespace: title: "Topic namespace" - description: - "The administrative unit of the topic, which acts as a grouping\ + description: "The administrative unit of the topic, which acts as a grouping\ \ mechanism for related topics. Most topic configuration is performed\ \ at the namespace level. Each tenant has one or multiple namespaces." type: "string" default: "default" examples: - - "default" + - "default" topic_pattern: title: "Topic pattern" - description: - "Topic pattern in which the records will be sent. You can use\ + description: "Topic pattern in which the records will be sent. You can use\ \ patterns like '{namespace}' and/or '{stream}' to send the message to\ \ a specific topic based on these values. Notice that the topic name will\ \ be transformed to a standard naming convention." type: "string" examples: - - "sample.topic" - - "{namespace}.{stream}.sample" + - "sample.topic" + - "{namespace}.{stream}.sample" topic_test: title: "Test topic" description: "Topic to test if Airbyte can produce messages." type: "string" examples: - - "test.topic" + - "test.topic" producer_name: title: "Producer name" - description: - "Name for the producer. If not filled, the system will generate\ + description: "Name for the producer. If not filled, the system will generate\ \ a globally unique name which can be accessed with." type: "string" examples: - - "airbyte-producer" + - "airbyte-producer" producer_sync: title: "Sync producer" description: "Wait synchronously until the record has been sent to Pulsar." @@ -2462,15 +2347,14 @@ type: "string" default: "NONE" enum: - - "NONE" - - "LZ4" - - "ZLIB" - - "ZSTD" - - "SNAPPY" + - "NONE" + - "LZ4" + - "ZLIB" + - "ZSTD" + - "SNAPPY" send_timeout_ms: title: "Message send timeout" - description: - "If a message is not acknowledged by a server before the send-timeout\ + description: "If a message is not acknowledged by a server before the send-timeout\ \ expires, an error occurs (in ms)." type: "integer" default: 30000 @@ -2486,8 +2370,7 @@ default: 50000 batching_enabled: title: "Enable batching" - description: - "Control whether automatic batching of messages is enabled\ + description: "Control whether automatic batching of messages is enabled\ \ for the producer." type: "boolean" default: true @@ -2498,15 +2381,13 @@ default: 1000 batching_max_publish_delay: title: "Batching max publish delay" - description: - " Time period in milliseconds within which the messages sent\ + description: " Time period in milliseconds within which the messages sent\ \ will be batched." type: "integer" default: 1 block_if_queue_full: title: "Block if queue is full" - description: - "If the send operation should block when the outgoing message\ + description: "If the send operation should block when the outgoing message\ \ queue is full." type: "boolean" default: false @@ -2514,7 +2395,7 @@ supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "append" + - "append" - dockerImage: "airbyte/destination-redshift:0.3.20" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/redshift" @@ -2523,17 +2404,16 @@ title: "Redshift Destination Spec" type: "object" required: - - "host" - - "port" - - "database" - - "username" - - "password" - - "schema" + - "host" + - "port" + - "database" + - "username" + - "password" + - "schema" additionalProperties: true properties: host: - description: - "Host Endpoint of the Redshift Cluster (must include the cluster-id,\ + description: "Host Endpoint of the Redshift Cluster (must include the cluster-id,\ \ region and end with .redshift.amazonaws.com)" type: "string" title: "Host" @@ -2544,7 +2424,7 @@ maximum: 65536 default: 5439 examples: - - "5439" + - "5439" title: "Port" username: description: "Username to use to access the database." @@ -2560,61 +2440,57 @@ type: "string" title: "Database" schema: - description: - "The default schema tables are written to if the source does\ + description: "The default schema tables are written to if the source does\ \ not specify a namespace. Unless specifically configured, the usual value\ \ for this field is \"public\"." type: "string" examples: - - "public" + - "public" default: "public" title: "Default Schema" s3_bucket_name: title: "S3 Bucket Name" type: "string" - description: - "The name of the staging S3 bucket to use if utilising a COPY\ + description: "The name of the staging S3 bucket to use if utilising a COPY\ \ strategy. COPY is recommended for production workloads for better speed\ \ and scalability. See AWS docs for more details." examples: - - "airbyte.staging" + - "airbyte.staging" s3_bucket_region: title: "S3 Bucket Region" type: "string" default: "" - description: - "The region of the S3 staging bucket to use if utilising a\ + description: "The region of the S3 staging bucket to use if utilising a\ \ copy strategy." enum: - - "" - - "us-east-1" - - "us-east-2" - - "us-west-1" - - "us-west-2" - - "af-south-1" - - "ap-east-1" - - "ap-south-1" - - "ap-northeast-1" - - "ap-northeast-2" - - "ap-northeast-3" - - "ap-southeast-1" - - "ap-southeast-2" - - "ca-central-1" - - "cn-north-1" - - "cn-northwest-1" - - "eu-central-1" - - "eu-north-1" - - "eu-south-1" - - "eu-west-1" - - "eu-west-2" - - "eu-west-3" - - "sa-east-1" - - "me-south-1" + - "" + - "us-east-1" + - "us-east-2" + - "us-west-1" + - "us-west-2" + - "af-south-1" + - "ap-east-1" + - "ap-south-1" + - "ap-northeast-1" + - "ap-northeast-2" + - "ap-northeast-3" + - "ap-southeast-1" + - "ap-southeast-2" + - "ca-central-1" + - "cn-north-1" + - "cn-northwest-1" + - "eu-central-1" + - "eu-north-1" + - "eu-south-1" + - "eu-west-1" + - "eu-west-2" + - "eu-west-3" + - "sa-east-1" + - "me-south-1" access_key_id: type: "string" - description: - "The Access Key Id granting allow one to access the above S3\ + description: "The Access Key Id granting allow one to access the above S3\ \ staging bucket. Airbyte requires Read and Write permissions to the given\ \ bucket." title: "S3 Key Id" @@ -2629,9 +2505,8 @@ minimum: 10 maximum: 100 examples: - - "10" - description: - "Optional. Increase this if syncing tables larger than 100GB.\ + - "10" + description: "Optional. Increase this if syncing tables larger than 100GB.\ \ Only relevant for COPY. Files are streamed to S3 in parts. This determines\ \ the size of each part, in MBs. As S3 has a limit of 10,000 parts per\ \ file, part size affects the table size. This is 10MB by default, resulting\ @@ -2643,9 +2518,9 @@ supportsNormalization: true supportsDBT: true supported_destination_sync_modes: - - "overwrite" - - "append" - - "append_dedup" + - "overwrite" + - "append" + - "append_dedup" - dockerImage: "airbyte/destination-s3:0.1.13" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/s3" @@ -2654,330 +2529,316 @@ title: "S3 Destination Spec" type: "object" required: - - "s3_bucket_name" - - "s3_bucket_path" - - "s3_bucket_region" - - "access_key_id" - - "secret_access_key" - - "format" + - "s3_bucket_name" + - "s3_bucket_path" + - "s3_bucket_region" + - "access_key_id" + - "secret_access_key" + - "format" additionalProperties: false properties: s3_endpoint: title: "Endpoint" type: "string" default: "" - description: - "This is your S3 endpoint url.(if you are working with AWS\ + description: "This is your S3 endpoint url.(if you are working with AWS\ \ S3, just leave empty)." examples: - - "http://localhost:9000" + - "http://localhost:9000" s3_bucket_name: title: "S3 Bucket Name" type: "string" description: "The name of the S3 bucket." examples: - - "airbyte_sync" + - "airbyte_sync" s3_bucket_path: description: "Directory under the S3 bucket where data will be written." type: "string" examples: - - "data_sync/test" + - "data_sync/test" s3_bucket_region: title: "S3 Bucket Region" type: "string" default: "" description: "The region of the S3 bucket." enum: - - "" - - "us-east-1" - - "us-east-2" - - "us-west-1" - - "us-west-2" - - "af-south-1" - - "ap-east-1" - - "ap-south-1" - - "ap-northeast-1" - - "ap-northeast-2" - - "ap-northeast-3" - - "ap-southeast-1" - - "ap-southeast-2" - - "ca-central-1" - - "cn-north-1" - - "cn-northwest-1" - - "eu-central-1" - - "eu-north-1" - - "eu-south-1" - - "eu-west-1" - - "eu-west-2" - - "eu-west-3" - - "sa-east-1" - - "me-south-1" - - "us-gov-east-1" - - "us-gov-west-1" + - "" + - "us-east-1" + - "us-east-2" + - "us-west-1" + - "us-west-2" + - "af-south-1" + - "ap-east-1" + - "ap-south-1" + - "ap-northeast-1" + - "ap-northeast-2" + - "ap-northeast-3" + - "ap-southeast-1" + - "ap-southeast-2" + - "ca-central-1" + - "cn-north-1" + - "cn-northwest-1" + - "eu-central-1" + - "eu-north-1" + - "eu-south-1" + - "eu-west-1" + - "eu-west-2" + - "eu-west-3" + - "sa-east-1" + - "me-south-1" + - "us-gov-east-1" + - "us-gov-west-1" access_key_id: type: "string" - description: - "The access key id to access the S3 bucket. Airbyte requires\ + description: "The access key id to access the S3 bucket. Airbyte requires\ \ Read and Write permissions to the given bucket." title: "S3 Key Id" airbyte_secret: true examples: - - "A012345678910EXAMPLE" + - "A012345678910EXAMPLE" secret_access_key: type: "string" description: "The corresponding secret to the access key id." title: "S3 Access Key" airbyte_secret: true examples: - - "a012345678910ABCDEFGH/AbCdEfGhEXAMPLEKEY" + - "a012345678910ABCDEFGH/AbCdEfGhEXAMPLEKEY" format: title: "Output Format" type: "object" description: "Output data format" oneOf: - - title: "Avro: Apache Avro" - required: - - "format_type" - - "compression_codec" - properties: - format_type: - type: "string" - enum: - - "Avro" - default: "Avro" - compression_codec: - title: "Compression Codec" - description: - "The compression algorithm used to compress data. Default\ - \ to no compression." - type: "object" - oneOf: - - title: "no compression" - required: - - "codec" - properties: - codec: - type: "string" - enum: - - "no compression" - default: "no compression" - - title: "Deflate" - required: - - "codec" - - "compression_level" - properties: - codec: - type: "string" - enum: - - "Deflate" - default: "Deflate" - compression_level: - title: "Deflate level" - description: - "0: no compression & fastest, 9: best compression\ - \ & slowest." - type: "integer" - default: 0 - minimum: 0 - maximum: 9 - - title: "bzip2" - required: - - "codec" - properties: - codec: - type: "string" - enum: - - "bzip2" - default: "bzip2" - - title: "xz" - required: - - "codec" - - "compression_level" - properties: - codec: - type: "string" - enum: - - "xz" - default: "xz" - compression_level: - title: "Compression level" - description: - "See here for details." - type: "integer" - default: 6 - minimum: 0 - maximum: 9 - - title: "zstandard" - required: - - "codec" - - "compression_level" - properties: - codec: - type: "string" - enum: - - "zstandard" - default: "zstandard" - compression_level: - title: "Compression level" - description: - "Negative levels are 'fast' modes akin to lz4 or\ - \ snappy, levels above 9 are generally for archival purposes,\ - \ and levels above 18 use a lot of memory." - type: "integer" - default: 3 - minimum: -5 - maximum: 22 - include_checksum: - title: "Include checksum" - description: "If true, include a checksum with each data block." - type: "boolean" - default: false - - title: "snappy" - required: - - "codec" - properties: - codec: - type: "string" - enum: - - "snappy" - default: "snappy" - part_size_mb: - title: "Block Size (MB) for Amazon S3 multipart upload" - description: - "This is the size of a \"Part\" being buffered in memory.\ - \ It limits the memory usage when writing. Larger values will allow\ - \ to upload a bigger files and improve the speed, but consumes9\ - \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." - type: "integer" - default: 5 - examples: - - 5 - - title: "CSV: Comma-Separated Values" - required: - - "format_type" - - "flattening" - properties: - format_type: - type: "string" - enum: - - "CSV" - default: "CSV" - flattening: - type: "string" - title: "Normalization (Flattening)" - description: - "Whether the input json data should be normalized (flattened)\ - \ in the output CSV. Please refer to docs for details." - default: "No flattening" - enum: - - "No flattening" - - "Root level flattening" - part_size_mb: - title: "Block Size (MB) for Amazon S3 multipart upload" - description: - "This is the size of a \"Part\" being buffered in memory.\ - \ It limits the memory usage when writing. Larger values will allow\ - \ to upload a bigger files and improve the speed, but consumes9\ - \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." - type: "integer" - default: 5 - examples: - - 5 - - title: "JSON Lines: newline-delimited JSON" - required: - - "format_type" - properties: - format_type: - type: "string" - enum: - - "JSONL" - default: "JSONL" - part_size_mb: - title: "Block Size (MB) for Amazon S3 multipart upload" - description: - "This is the size of a \"Part\" being buffered in memory.\ - \ It limits the memory usage when writing. Larger values will allow\ - \ to upload a bigger files and improve the speed, but consumes9\ - \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." - type: "integer" - default: 5 - examples: - - 5 - - title: "Parquet: Columnar Storage" - required: - - "format_type" - properties: - format_type: - type: "string" - enum: - - "Parquet" - default: "Parquet" - compression_codec: - title: "Compression Codec" - description: "The compression algorithm used to compress data pages." - type: "string" - enum: - - "UNCOMPRESSED" - - "SNAPPY" - - "GZIP" - - "LZO" - - "BROTLI" - - "LZ4" - - "ZSTD" - default: "UNCOMPRESSED" - block_size_mb: - title: "Block Size (Row Group Size) (MB)" - description: - "This is the size of a row group being buffered in memory.\ - \ It limits the memory usage when writing. Larger values will improve\ - \ the IO when reading, but consume more memory when writing. Default:\ - \ 128 MB." - type: "integer" - default: 128 - examples: - - 128 - max_padding_size_mb: - title: "Max Padding Size (MB)" - description: - "Maximum size allowed as padding to align row groups.\ - \ This is also the minimum size of a row group. Default: 8 MB." - type: "integer" - default: 8 - examples: - - 8 - page_size_kb: - title: "Page Size (KB)" - description: - "The page size is for compression. A block is composed\ - \ of pages. A page is the smallest unit that must be read fully\ - \ to access a single record. If this value is too small, the compression\ - \ will deteriorate. Default: 1024 KB." - type: "integer" - default: 1024 - examples: - - 1024 - dictionary_page_size_kb: - title: "Dictionary Page Size (KB)" - description: - "There is one dictionary page per column per row group\ - \ when dictionary encoding is used. The dictionary page size works\ - \ like the page size but for dictionary. Default: 1024 KB." - type: "integer" - default: 1024 - examples: - - 1024 - dictionary_encoding: - title: "Dictionary Encoding" - description: "Default: true." - type: "boolean" - default: true + - title: "Avro: Apache Avro" + required: + - "format_type" + - "compression_codec" + properties: + format_type: + type: "string" + enum: + - "Avro" + default: "Avro" + compression_codec: + title: "Compression Codec" + description: "The compression algorithm used to compress data. Default\ + \ to no compression." + type: "object" + oneOf: + - title: "no compression" + required: + - "codec" + properties: + codec: + type: "string" + enum: + - "no compression" + default: "no compression" + - title: "Deflate" + required: + - "codec" + - "compression_level" + properties: + codec: + type: "string" + enum: + - "Deflate" + default: "Deflate" + compression_level: + title: "Deflate level" + description: "0: no compression & fastest, 9: best compression\ + \ & slowest." + type: "integer" + default: 0 + minimum: 0 + maximum: 9 + - title: "bzip2" + required: + - "codec" + properties: + codec: + type: "string" + enum: + - "bzip2" + default: "bzip2" + - title: "xz" + required: + - "codec" + - "compression_level" + properties: + codec: + type: "string" + enum: + - "xz" + default: "xz" + compression_level: + title: "Compression level" + description: "See here for details." + type: "integer" + default: 6 + minimum: 0 + maximum: 9 + - title: "zstandard" + required: + - "codec" + - "compression_level" + properties: + codec: + type: "string" + enum: + - "zstandard" + default: "zstandard" + compression_level: + title: "Compression level" + description: "Negative levels are 'fast' modes akin to lz4 or\ + \ snappy, levels above 9 are generally for archival purposes,\ + \ and levels above 18 use a lot of memory." + type: "integer" + default: 3 + minimum: -5 + maximum: 22 + include_checksum: + title: "Include checksum" + description: "If true, include a checksum with each data block." + type: "boolean" + default: false + - title: "snappy" + required: + - "codec" + properties: + codec: + type: "string" + enum: + - "snappy" + default: "snappy" + part_size_mb: + title: "Block Size (MB) for Amazon S3 multipart upload" + description: "This is the size of a \"Part\" being buffered in memory.\ + \ It limits the memory usage when writing. Larger values will allow\ + \ to upload a bigger files and improve the speed, but consumes9\ + \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." + type: "integer" + default: 5 + examples: + - 5 + - title: "CSV: Comma-Separated Values" + required: + - "format_type" + - "flattening" + properties: + format_type: + type: "string" + enum: + - "CSV" + default: "CSV" + flattening: + type: "string" + title: "Normalization (Flattening)" + description: "Whether the input json data should be normalized (flattened)\ + \ in the output CSV. Please refer to docs for details." + default: "No flattening" + enum: + - "No flattening" + - "Root level flattening" + part_size_mb: + title: "Block Size (MB) for Amazon S3 multipart upload" + description: "This is the size of a \"Part\" being buffered in memory.\ + \ It limits the memory usage when writing. Larger values will allow\ + \ to upload a bigger files and improve the speed, but consumes9\ + \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." + type: "integer" + default: 5 + examples: + - 5 + - title: "JSON Lines: newline-delimited JSON" + required: + - "format_type" + properties: + format_type: + type: "string" + enum: + - "JSONL" + default: "JSONL" + part_size_mb: + title: "Block Size (MB) for Amazon S3 multipart upload" + description: "This is the size of a \"Part\" being buffered in memory.\ + \ It limits the memory usage when writing. Larger values will allow\ + \ to upload a bigger files and improve the speed, but consumes9\ + \ more memory. Allowed values: min=5MB, max=525MB Default: 5MB." + type: "integer" + default: 5 + examples: + - 5 + - title: "Parquet: Columnar Storage" + required: + - "format_type" + properties: + format_type: + type: "string" + enum: + - "Parquet" + default: "Parquet" + compression_codec: + title: "Compression Codec" + description: "The compression algorithm used to compress data pages." + type: "string" + enum: + - "UNCOMPRESSED" + - "SNAPPY" + - "GZIP" + - "LZO" + - "BROTLI" + - "LZ4" + - "ZSTD" + default: "UNCOMPRESSED" + block_size_mb: + title: "Block Size (Row Group Size) (MB)" + description: "This is the size of a row group being buffered in memory.\ + \ It limits the memory usage when writing. Larger values will improve\ + \ the IO when reading, but consume more memory when writing. Default:\ + \ 128 MB." + type: "integer" + default: 128 + examples: + - 128 + max_padding_size_mb: + title: "Max Padding Size (MB)" + description: "Maximum size allowed as padding to align row groups.\ + \ This is also the minimum size of a row group. Default: 8 MB." + type: "integer" + default: 8 + examples: + - 8 + page_size_kb: + title: "Page Size (KB)" + description: "The page size is for compression. A block is composed\ + \ of pages. A page is the smallest unit that must be read fully\ + \ to access a single record. If this value is too small, the compression\ + \ will deteriorate. Default: 1024 KB." + type: "integer" + default: 1024 + examples: + - 1024 + dictionary_page_size_kb: + title: "Dictionary Page Size (KB)" + description: "There is one dictionary page per column per row group\ + \ when dictionary encoding is used. The dictionary page size works\ + \ like the page size but for dictionary. Default: 1024 KB." + type: "integer" + default: 1024 + examples: + - 1024 + dictionary_encoding: + title: "Dictionary Encoding" + description: "Default: true." + type: "boolean" + default: true supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" + - "overwrite" + - "append" - dockerImage: "airbyte/destination-snowflake:0.3.17" spec: documentationUrl: "https://docs.airbyte.io/integrations/destinations/snowflake" @@ -2986,58 +2847,56 @@ title: "Snowflake Destination Spec" type: "object" required: - - "host" - - "role" - - "warehouse" - - "database" - - "schema" - - "username" - - "password" + - "host" + - "role" + - "warehouse" + - "database" + - "schema" + - "username" + - "password" additionalProperties: true properties: host: - description: - "Host domain of the snowflake instance (must include the account,\ + description: "Host domain of the snowflake instance (must include the account,\ \ region, cloud environment, and end with snowflakecomputing.com)." examples: - - "accountname.us-east-2.aws.snowflakecomputing.com" + - "accountname.us-east-2.aws.snowflakecomputing.com" type: "string" title: "Host" order: 0 role: description: "The role you created for Airbyte to access Snowflake." examples: - - "AIRBYTE_ROLE" + - "AIRBYTE_ROLE" type: "string" title: "Role" order: 1 warehouse: description: "The warehouse you created for Airbyte to sync data into." examples: - - "AIRBYTE_WAREHOUSE" + - "AIRBYTE_WAREHOUSE" type: "string" title: "Warehouse" order: 2 database: description: "The database you created for Airbyte to sync data into." examples: - - "AIRBYTE_DATABASE" + - "AIRBYTE_DATABASE" type: "string" title: "Database" order: 3 schema: - description: - "The default Snowflake schema tables are written to if the\ + description: "The default Snowflake schema tables are written to if the\ \ source does not specify a namespace." examples: - - "AIRBYTE_SCHEMA" + - "AIRBYTE_SCHEMA" type: "string" title: "Default Schema" order: 4 username: description: "The username you created to allow Airbyte to access the database." examples: - - "AIRBYTE_USER" + - "AIRBYTE_USER" type: "string" title: "Username" order: 5 @@ -3053,148 +2912,140 @@ description: "Loading method used to send data to Snowflake." order: 7 oneOf: - - title: "Standard Inserts" - additionalProperties: false - description: - "Uses
INSERT
statements to send batches of records\ - \ to Snowflake. Easiest (no setup) but not recommended for large production\ - \ workloads due to slow speed." - required: - - "method" - properties: - method: - type: "string" - enum: - - "Standard" - default: "Standard" - - title: "AWS S3 Staging" - additionalProperties: false - description: - "Writes large batches of records to a file, uploads the file\ - \ to S3, then uses
COPY INTO table
to upload the file. Recommended\ - \ for large production workloads for better speed and scalability." - required: - - "method" - - "s3_bucket_name" - - "access_key_id" - - "secret_access_key" - properties: - method: - type: "string" - enum: - - "S3 Staging" - default: "S3 Staging" - order: 0 - s3_bucket_name: - title: "S3 Bucket Name" - type: "string" - description: - "The name of the staging S3 bucket. Airbyte will write\ - \ files to this bucket and read them via
COPY
statements\ - \ on Snowflake." - examples: - - "airbyte.staging" - order: 1 - s3_bucket_region: - title: "S3 Bucket Region" - type: "string" - default: "" - description: - "The region of the S3 staging bucket to use if utilising\ - \ a copy strategy." - enum: - - "" - - "us-east-1" - - "us-east-2" - - "us-west-1" - - "us-west-2" - - "af-south-1" - - "ap-east-1" - - "ap-south-1" - - "ap-northeast-1" - - "ap-northeast-2" - - "ap-northeast-3" - - "ap-southeast-1" - - "ap-southeast-2" - - "ca-central-1" - - "cn-north-1" - - "cn-northwest-1" - - "eu-central-1" - - "eu-west-1" - - "eu-west-2" - - "eu-west-3" - - "eu-south-1" - - "eu-north-1" - - "sa-east-1" - - "me-south-1" - order: 2 - access_key_id: - type: "string" - description: - "The Access Key Id granting allow one to access the above\ - \ S3 staging bucket. Airbyte requires Read and Write permissions\ - \ to the given bucket." - title: "S3 Key Id" - airbyte_secret: true - order: 3 - secret_access_key: - type: "string" - description: "The corresponding secret to the above access key id." - title: "S3 Access Key" - airbyte_secret: true - order: 4 - - title: "GCS Staging" - additionalProperties: false - description: - "Writes large batches of records to a file, uploads the file\ - \ to GCS, then uses
COPY INTO table
to upload the file. Recommended\ - \ for large production workloads for better speed and scalability." - required: - - "method" - - "project_id" - - "bucket_name" - - "credentials_json" - properties: - method: - type: "string" - enum: - - "GCS Staging" - default: "GCS Staging" - order: 0 - project_id: - title: "GCP Project ID" - type: "string" - description: "The name of the GCP project ID for your credentials." - examples: - - "my-project" - order: 1 - bucket_name: - title: "GCS Bucket Name" - type: "string" - description: - "The name of the staging GCS bucket. Airbyte will write\ - \ files to this bucket and read them via
COPY
statements\ - \ on Snowflake." - examples: - - "airbyte-staging" - order: 2 - credentials_json: - title: "Google Application Credentials" - type: "string" - description: - "The contents of the JSON key file that has read/write\ - \ permissions to the staging GCS bucket. You will separately need\ - \ to grant bucket access to your Snowflake GCP service account.\ - \ See the GCP docs for more information on how to generate a JSON key\ - \ for your service account." - airbyte_secret: true - multiline: true - order: 3 + - title: "Standard Inserts" + additionalProperties: false + description: "Uses
INSERT
statements to send batches of records\ + \ to Snowflake. Easiest (no setup) but not recommended for large production\ + \ workloads due to slow speed." + required: + - "method" + properties: + method: + type: "string" + enum: + - "Standard" + default: "Standard" + - title: "AWS S3 Staging" + additionalProperties: false + description: "Writes large batches of records to a file, uploads the file\ + \ to S3, then uses
COPY INTO table
to upload the file. Recommended\ + \ for large production workloads for better speed and scalability." + required: + - "method" + - "s3_bucket_name" + - "access_key_id" + - "secret_access_key" + properties: + method: + type: "string" + enum: + - "S3 Staging" + default: "S3 Staging" + order: 0 + s3_bucket_name: + title: "S3 Bucket Name" + type: "string" + description: "The name of the staging S3 bucket. Airbyte will write\ + \ files to this bucket and read them via
COPY
statements\ + \ on Snowflake." + examples: + - "airbyte.staging" + order: 1 + s3_bucket_region: + title: "S3 Bucket Region" + type: "string" + default: "" + description: "The region of the S3 staging bucket to use if utilising\ + \ a copy strategy." + enum: + - "" + - "us-east-1" + - "us-east-2" + - "us-west-1" + - "us-west-2" + - "af-south-1" + - "ap-east-1" + - "ap-south-1" + - "ap-northeast-1" + - "ap-northeast-2" + - "ap-northeast-3" + - "ap-southeast-1" + - "ap-southeast-2" + - "ca-central-1" + - "cn-north-1" + - "cn-northwest-1" + - "eu-central-1" + - "eu-west-1" + - "eu-west-2" + - "eu-west-3" + - "eu-south-1" + - "eu-north-1" + - "sa-east-1" + - "me-south-1" + order: 2 + access_key_id: + type: "string" + description: "The Access Key Id granting allow one to access the above\ + \ S3 staging bucket. Airbyte requires Read and Write permissions\ + \ to the given bucket." + title: "S3 Key Id" + airbyte_secret: true + order: 3 + secret_access_key: + type: "string" + description: "The corresponding secret to the above access key id." + title: "S3 Access Key" + airbyte_secret: true + order: 4 + - title: "GCS Staging" + additionalProperties: false + description: "Writes large batches of records to a file, uploads the file\ + \ to GCS, then uses
COPY INTO table
to upload the file. Recommended\ + \ for large production workloads for better speed and scalability." + required: + - "method" + - "project_id" + - "bucket_name" + - "credentials_json" + properties: + method: + type: "string" + enum: + - "GCS Staging" + default: "GCS Staging" + order: 0 + project_id: + title: "GCP Project ID" + type: "string" + description: "The name of the GCP project ID for your credentials." + examples: + - "my-project" + order: 1 + bucket_name: + title: "GCS Bucket Name" + type: "string" + description: "The name of the staging GCS bucket. Airbyte will write\ + \ files to this bucket and read them via
COPY
statements\ + \ on Snowflake." + examples: + - "airbyte-staging" + order: 2 + credentials_json: + title: "Google Application Credentials" + type: "string" + description: "The contents of the JSON key file that has read/write\ + \ permissions to the staging GCS bucket. You will separately need\ + \ to grant bucket access to your Snowflake GCP service account.\ + \ See the GCP docs for more information on how to generate a JSON key\ + \ for your service account." + airbyte_secret: true + multiline: true + order: 3 supportsIncremental: true supportsNormalization: true supportsDBT: true supported_destination_sync_modes: - - "overwrite" - - "append" - - "append_dedup" + - "overwrite" + - "append" + - "append_dedup" diff --git a/airbyte-config/init/src/main/resources/seed/source_specs.yaml b/airbyte-config/init/src/main/resources/seed/source_specs.yaml index e358dbf93d03f..539bbb8e7c74b 100644 --- a/airbyte-config/init/src/main/resources/seed/source_specs.yaml +++ b/airbyte-config/init/src/main/resources/seed/source_specs.yaml @@ -9,37 +9,33 @@ title: "Aws CloudTrail Spec" type: "object" required: - - "aws_key_id" - - "aws_secret_key" - - "aws_region_name" - - "start_date" + - "aws_key_id" + - "aws_secret_key" + - "aws_region_name" + - "start_date" additionalProperties: true properties: aws_key_id: type: "string" - description: - "Specifies an AWS access key associated with an IAM user or\ + description: "Specifies an AWS access key associated with an IAM user or\ \ role." airbyte_secret: true aws_secret_key: type: "string" - description: - "Specifies the secret key associated with the access key. This\ + description: "Specifies the secret key associated with the access key. This\ \ is essentially the 'password' for the access key." airbyte_secret: true aws_region_name: type: "string" - description: - "The default AWS Region to use, for example, us-west-1 or us-west-2.\ + description: "The default AWS Region to use, for example, us-west-1 or us-west-2.\ \ When specifying a Region inline during client initialization, this property\ \ is named region_name." start_date: type: "string" - description: - "The date you would like to replicate data. Data in ClouTraid\ + description: "The date you would like to replicate data. Data in ClouTraid\ \ is available for last 90 days only. Format: YYYY-MM-DD." examples: - - "2021-01-01" + - "2021-01-01" default: "1970-01-01" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" supportsNormalization: false @@ -54,46 +50,41 @@ properties: client_id: title: "Client Id" - description: - "Oauth client id How to create your Login with Amazon" name: "Client ID" type: "string" client_secret: title: "Client Secret" - description: - "Oauth client secret How to create your Login with Amazon" name: "Client secret" airbyte_secret: true type: "string" scope: title: "Scope" - description: - "By default its advertising::campaign_management, but customers\ + description: "By default its advertising::campaign_management, but customers\ \ may need to set scope to cpc_advertising:campaign_management." default: "advertising::campaign_management" name: "Client scope" examples: - - "cpc_advertising:campaign_management" + - "cpc_advertising:campaign_management" type: "string" refresh_token: title: "Refresh Token" - description: - "Oauth 2.0 refresh_token, read details here" name: "Oauth refresh token" airbyte_secret: true type: "string" start_date: title: "Start Date" - description: - "Start date for collectiong reports, should not be more than\ + description: "Start date for collectiong reports, should not be more than\ \ 60 days in past. In YYYY-MM-DD format" name: "Start date" examples: - - "2022-10-10" - - "2022-10-22" + - "2022-10-10" + - "2022-10-22" type: "string" region: description: "Region to pull data from (EU/NA/FE/SANDBOX)" @@ -101,10 +92,10 @@ name: "Region" title: "AmazonAdsRegion" enum: - - "NA" - - "EU" - - "FE" - - "SANDBOX" + - "NA" + - "EU" + - "FE" + - "SANDBOX" type: "string" profiles: title: "Profiles" @@ -114,9 +105,9 @@ items: type: "integer" required: - - "client_id" - - "client_secret" - - "refresh_token" + - "client_id" + - "client_secret" + - "refresh_token" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -130,17 +121,15 @@ properties: replication_start_date: title: "Replication Start Date" - description: - "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ \ data before this date will not be replicated." pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2017-01-25T00:00:00Z" + - "2017-01-25T00:00:00Z" type: "string" refresh_token: title: "Refresh Token" - description: - "The refresh token used obtained via authorization (can be\ + description: "The refresh token used obtained via authorization (can be\ \ passed to the client instead)" airbyte_secret: true type: "string" @@ -173,78 +162,78 @@ title: "AWSEnvironment" description: "An enumeration." enum: - - "PRODUCTION" - - "SANDBOX" + - "PRODUCTION" + - "SANDBOX" type: "string" region: title: "AWSRegion" description: "An enumeration." enum: - - "AE" - - "DE" - - "PL" - - "EG" - - "ES" - - "FR" - - "IN" - - "IT" - - "NL" - - "SA" - - "SE" - - "TR" - - "UK" - - "AU" - - "JP" - - "SG" - - "US" - - "BR" - - "CA" - - "MX" - - "GB" + - "AE" + - "DE" + - "PL" + - "EG" + - "ES" + - "FR" + - "IN" + - "IT" + - "NL" + - "SA" + - "SE" + - "TR" + - "UK" + - "AU" + - "JP" + - "SG" + - "US" + - "BR" + - "CA" + - "MX" + - "GB" type: "string" required: - - "replication_start_date" - - "refresh_token" - - "lwa_app_id" - - "lwa_client_secret" - - "aws_access_key" - - "aws_secret_key" - - "role_arn" - - "aws_environment" - - "region" + - "replication_start_date" + - "refresh_token" + - "lwa_app_id" + - "lwa_client_secret" + - "aws_access_key" + - "aws_secret_key" + - "role_arn" + - "aws_environment" + - "region" definitions: AWSEnvironment: title: "AWSEnvironment" description: "An enumeration." enum: - - "PRODUCTION" - - "SANDBOX" + - "PRODUCTION" + - "SANDBOX" type: "string" AWSRegion: title: "AWSRegion" description: "An enumeration." enum: - - "AE" - - "DE" - - "PL" - - "EG" - - "ES" - - "FR" - - "IN" - - "IT" - - "NL" - - "SA" - - "SE" - - "TR" - - "UK" - - "AU" - - "JP" - - "SG" - - "US" - - "BR" - - "CA" - - "MX" - - "GB" + - "AE" + - "DE" + - "PL" + - "EG" + - "ES" + - "FR" + - "IN" + - "IT" + - "NL" + - "SA" + - "SE" + - "TR" + - "UK" + - "AU" + - "JP" + - "SG" + - "US" + - "BR" + - "CA" + - "MX" + - "GB" type: "string" supportsNormalization: false supportsDBT: false @@ -257,31 +246,28 @@ title: "Amplitude Spec" type: "object" required: - - "api_key" - - "secret_key" - - "start_date" + - "api_key" + - "secret_key" + - "start_date" additionalProperties: false properties: api_key: type: "string" - description: - "This is the project’s API key, used for calling Amplitude’\ + description: "This is the project’s API key, used for calling Amplitude’\ s APIs" airbyte_secret: true secret_key: type: "string" - description: - "This is the project's secret key, which is also used for calling\ + description: "This is the project's secret key, which is also used for calling\ \ Amplitude’s APIs" airbyte_secret: true start_date: type: "string" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" - description: - "UTC date and time in the format 2021-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2021-01-25T00:00:00Z. Any\ \ data before this date will not be replicated." examples: - - "2021-01-25T00:00:00Z" + - "2021-01-25T00:00:00Z" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -293,7 +279,7 @@ title: "Apify Dataset Spec" type: "object" required: - - "datasetId" + - "datasetId" additionalProperties: false properties: datasetId: @@ -301,8 +287,7 @@ description: "ID of the dataset you would like to load to Airbyte." clean: type: "boolean" - description: - "If set to true, only clean items will be downloaded from the\ + description: "If set to true, only clean items will be downloaded from the\ \ dataset. See description of what clean means in Apify API docs. If not sure, set clean to false." supportsNormalization: false @@ -316,29 +301,26 @@ title: "Source Appstore Singer Spec" type: "object" required: - - "key_id" - - "private_key" - - "issuer_id" - - "vendor" - - "start_date" + - "key_id" + - "private_key" + - "issuer_id" + - "vendor" + - "start_date" additionalProperties: false properties: key_id: type: "string" - description: - "Key_id is the API key you use to connect to appstore's API." private_key: type: "string" - description: - "Private_key is the contents of the key file you use to connect to appstore's API." airbyte_secret: true multiline: true issuer_id: type: "string" - description: - "Issuer_id is used to generate the credentials to connect to appstore's\ \ API." vendor: @@ -348,7 +330,7 @@ type: "string" description: "Date from which to start pulling data." examples: - - "2020-11-16T00:00:00Z" + - "2020-11-16T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" supportsNormalization: false supportsDBT: false @@ -367,50 +349,49 @@ description: "Choose how to authenticate to Github" type: "object" oneOf: - - type: "object" - title: "Authenticate with Personal Access Token" - required: - - "personal_access_token" - properties: - option_title: - type: "string" - title: "Credentials title" - description: "PAT Credentials" - const: "PAT Credentials" - personal_access_token: - type: "string" - title: "Personal Access Token" - description: - "Asana Personal Access Token (generate yours here)." - airbyte_secret: true - - type: "object" - title: "Authenticate via Asana (Oauth)" - required: - - "client_id" - - "client_secret" - - "refresh_token" - properties: - option_title: - type: "string" - title: "Credentials title" - description: "OAuth Credentials" - const: "OAuth Credentials" - client_id: - type: "string" - title: "" - description: "" - airbyte_secret: true - client_secret: - type: "string" - title: "" - description: "" - airbyte_secret: true - refresh_token: - type: "string" - title: "" - description: "" - airbyte_secret: true + - type: "object" + title: "Authenticate with Personal Access Token" + required: + - "personal_access_token" + properties: + option_title: + type: "string" + title: "Credentials title" + description: "PAT Credentials" + const: "PAT Credentials" + personal_access_token: + type: "string" + title: "Personal Access Token" + description: "Asana Personal Access Token (generate yours here)." + airbyte_secret: true + - type: "object" + title: "Authenticate via Asana (Oauth)" + required: + - "client_id" + - "client_secret" + - "refresh_token" + properties: + option_title: + type: "string" + title: "Credentials title" + description: "OAuth Credentials" + const: "OAuth Credentials" + client_id: + type: "string" + title: "" + description: "" + airbyte_secret: true + client_secret: + type: "string" + title: "" + description: "" + airbyte_secret: true + refresh_token: + type: "string" + title: "" + description: "" + airbyte_secret: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -418,13 +399,13 @@ auth_type: "oauth2.0" oauth2Specification: rootObject: - - "credentials" - - "1" + - "credentials" + - "1" oauthFlowInitParameters: - - - "client_id" - - - "client_secret" + - - "client_id" + - - "client_secret" oauthFlowOutputParameters: - - - "refresh_token" + - - "refresh_token" - dockerImage: "airbyte/source-bamboo-hr:0.1.0" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/bamboo-hr" @@ -433,8 +414,8 @@ title: "Bamboo HR Spec" type: "object" required: - - "subdomain" - - "api_key" + - "subdomain" + - "api_key" additionalProperties: false properties: subdomain: @@ -455,21 +436,20 @@ title: "BigCommerce Source CDK Specifications" type: "object" required: - - "start_date" - - "store_hash" - - "access_token" + - "start_date" + - "store_hash" + - "access_token" additionalProperties: false properties: start_date: type: "string" description: "The date you would like to replicate data. Format: YYYY-MM-DD." examples: - - "2021-01-01" + - "2021-01-01" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" store_hash: type: "string" - description: - "The hash code of the store. For https://api.bigcommerce.com/stores/HASH_CODE/v3/,\ + description: "The hash code of the store. For https://api.bigcommerce.com/stores/HASH_CODE/v3/,\ \ The store's hash code is 'HASH_CODE'." access_token: type: "string" @@ -486,14 +466,13 @@ title: "BigQuery Source Spec" type: "object" required: - - "project_id" - - "credentials_json" + - "project_id" + - "credentials_json" additionalProperties: false properties: project_id: type: "string" - description: - "The GCP project ID for the project containing the target BigQuery\ + description: "The GCP project ID for the project containing the target BigQuery\ \ dataset." title: "Project ID" dataset_id: @@ -502,8 +481,7 @@ title: "Default Dataset ID" credentials_json: type: "string" - description: - "The contents of the JSON service account key. Check out the\ + description: "The contents of the JSON service account key. Check out the\ \ docs\ \ if you need help generating this key." title: "Credentials JSON" @@ -513,9 +491,9 @@ supportsDBT: true supported_destination_sync_modes: [] supported_sync_modes: - - "overwrite" - - "append" - - "append_dedup" + - "overwrite" + - "append" + - "append_dedup" - dockerImage: "airbyte/source-bing-ads:0.1.1" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/bing-ads" @@ -524,18 +502,18 @@ title: "Bing Ads Spec" type: "object" required: - - "accounts" - - "client_id" - - "client_secret" - - "customer_id" - - "developer_token" - - "refresh_token" - - "user_id" - - "reports_start_date" - - "hourly_reports" - - "daily_reports" - - "weekly_reports" - - "monthly_reports" + - "accounts" + - "client_id" + - "client_secret" + - "customer_id" + - "developer_token" + - "refresh_token" + - "user_id" + - "reports_start_date" + - "hourly_reports" + - "daily_reports" + - "weekly_reports" + - "monthly_reports" additionalProperties: false properties: accounts: @@ -543,36 +521,36 @@ type: "object" description: "Account selection strategy." oneOf: - - title: "All accounts assigned to your user" - additionalProperties: false - description: "Fetch data for all available accounts." - required: - - "selection_strategy" - properties: - selection_strategy: - type: "string" - enum: - - "all" - const: "all" - - title: "Subset of your accounts" - additionalProperties: false - description: "Fetch data for subset of account ids." - required: - - "ids" - - "selection_strategy" - properties: - selection_strategy: + - title: "All accounts assigned to your user" + additionalProperties: false + description: "Fetch data for all available accounts." + required: + - "selection_strategy" + properties: + selection_strategy: + type: "string" + enum: + - "all" + const: "all" + - title: "Subset of your accounts" + additionalProperties: false + description: "Fetch data for subset of account ids." + required: + - "ids" + - "selection_strategy" + properties: + selection_strategy: + type: "string" + enum: + - "subset" + const: "subset" + ids: + type: "array" + description: "List of accounts from which data will be fetched." + items: type: "string" - enum: - - "subset" - const: "subset" - ids: - type: "array" - description: "List of accounts from which data will be fetched." - items: - type: "string" - minItems: 1 - uniqueItems: true + minItems: 1 + uniqueItems: true client_id: type: "string" description: "ID of your Microsoft Advertising client application." @@ -590,8 +568,7 @@ airbyte_secret: true refresh_token: type: "string" - description: - "The long-lived Refresh token received via grant_type=refresh_token\ + description: "The long-lived Refresh token received via grant_type=refresh_token\ \ request." airbyte_secret: true user_id: @@ -601,8 +578,7 @@ type: "string" format: "date" default: "2020-01-01" - description: - "From which date perform initial sync for report related streams.\ + description: "From which date perform initial sync for report related streams.\ \ In YYYY-MM-DD format" hourly_reports: title: "Hourly reports" @@ -617,8 +593,7 @@ weekly_reports: title: "Weekly reports" type: "boolean" - description: - "The report data will be aggregated by each week running from\ + description: "The report data will be aggregated by each week running from\ \ Sunday through Saturday." default: false monthly_reports: @@ -638,8 +613,7 @@ properties: merchant_id: title: "Merchant Id" - description: - "Merchant ID is the unique identifier for entire gateway account." name: "Merchant ID" type: "string" @@ -656,40 +630,39 @@ type: "string" start_date: title: "Start Date" - description: - "The date from which you'd like to replicate data for Braintree\ + description: "The date from which you'd like to replicate data for Braintree\ \ API for UTC timezone, All data generated after this date will be replicated." name: "Start date" examples: - - "2020" - - "2020-12-30" - - "2020-11-22 20:20:05" + - "2020" + - "2020-12-30" + - "2020-11-22 20:20:05" type: "string" format: "date-time" environment: description: "Environment specifies where the data will come from." name: "Environment" examples: - - "sandbox" - - "production" - - "qa" - - "development" + - "sandbox" + - "production" + - "qa" + - "development" allOf: - - $ref: "#/definitions/Environment" + - $ref: "#/definitions/Environment" required: - - "merchant_id" - - "public_key" - - "private_key" - - "environment" + - "merchant_id" + - "public_key" + - "private_key" + - "environment" definitions: Environment: title: "Environment" description: "An enumeration." enum: - - "Development" - - "Sandbox" - - "Qa" - - "Production" + - "Development" + - "Sandbox" + - "Qa" + - "Production" type: "string" supportsNormalization: false supportsDBT: false @@ -702,21 +675,19 @@ title: "Cart Spec" type: "object" required: - - "access_token" - - "start_date" - - "store_name" + - "access_token" + - "start_date" + - "store_name" additionalProperties: true properties: access_token: type: "string" airbyte_secret: true - description: - "API Key. See the docs for information on how to generate this key." store_name: type: "string" - description: - "Store name. All API URLs start with https://[mystorename.com]/api/v1/,\ + description: "Store name. All API URLs start with https://[mystorename.com]/api/v1/,\ \ where [mystorename.com] is the domain name of your store." start_date: title: "Start Date" @@ -724,7 +695,7 @@ description: "The date from which you'd like to replicate the data" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2021-01-01T00:00:00Z" + - "2021-01-01T00:00:00Z" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -736,10 +707,10 @@ title: "Chargebee Spec" type: "object" required: - - "site" - - "site_api_key" - - "start_date" - - "product_catalog" + - "site" + - "site_api_key" + - "start_date" + - "product_catalog" additionalProperties: false properties: site: @@ -747,33 +718,31 @@ title: "Site" description: "The site prefix for your Chargebee instance." examples: - - "airbyte-test" + - "airbyte-test" site_api_key: type: "string" title: "API Key" description: "The API key from your Chargebee instance." examples: - - "test_3yzfanAXF66USdWC9wQcM555DQJkSYoppu" + - "test_3yzfanAXF66USdWC9wQcM555DQJkSYoppu" airbyte_secret: true start_date: type: "string" title: "Start Date" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" - description: - "UTC date and time in the format 2021-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2021-01-25T00:00:00Z. Any\ \ data before this date will not be replicated." examples: - - "2021-01-25T00:00:00Z" + - "2021-01-25T00:00:00Z" product_catalog: title: "Product Catalog" type: "string" - description: - "Product Catalog version of your Chargebee site. Instructions\ + description: "Product Catalog version of your Chargebee site. Instructions\ \ on how to find your version you may find here under `API Version` section." enum: - - "1.0" - - "2.0" + - "1.0" + - "2.0" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -785,10 +754,10 @@ title: "ClickHouse Source Spec" type: "object" required: - - "host" - - "port" - - "database" - - "username" + - "host" + - "port" + - "database" + - "username" additionalProperties: false properties: host: @@ -801,12 +770,12 @@ maximum: 65536 default: 8123 examples: - - "8123" + - "8123" database: description: "Name of the database." type: "string" examples: - - "default" + - "default" username: description: "Username to use to access the database." type: "string" @@ -822,112 +791,104 @@ tunnel_method: type: "object" title: "SSH Tunnel Method" - description: - "Whether to initiate an SSH tunnel before connecting to the\ + description: "Whether to initiate an SSH tunnel before connecting to the\ \ database, and if so, which kind of authentication to use." oneOf: - - title: "No Tunnel" - required: - - "tunnel_method" - properties: - tunnel_method: - description: "No ssh tunnel needed to connect to database" - type: "string" - const: "NO_TUNNEL" - order: 0 - - title: "SSH Key Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "ssh_key" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and ssh key" - type: "string" - const: "SSH_KEY_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host." - type: "string" - order: 3 - ssh_key: - title: "SSH Private Key" - description: - "OS-level user account ssh key credentials in RSA PEM\ - \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" - type: "string" - airbyte_secret: true - multiline: true - order: 4 - - title: "Password Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "tunnel_user_password" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and password authentication" - type: "string" - const: "SSH_PASSWORD_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host" - type: "string" - order: 3 - tunnel_user_password: - title: "Password" - description: "OS-level password for logging into the jump server host" - type: "string" - airbyte_secret: true - order: 4 + - title: "No Tunnel" + required: + - "tunnel_method" + properties: + tunnel_method: + description: "No ssh tunnel needed to connect to database" + type: "string" + const: "NO_TUNNEL" + order: 0 + - title: "SSH Key Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "ssh_key" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and ssh key" + type: "string" + const: "SSH_KEY_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host." + type: "string" + order: 3 + ssh_key: + title: "SSH Private Key" + description: "OS-level user account ssh key credentials in RSA PEM\ + \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" + type: "string" + airbyte_secret: true + multiline: true + order: 4 + - title: "Password Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "tunnel_user_password" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and password authentication" + type: "string" + const: "SSH_PASSWORD_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host" + type: "string" + order: 3 + tunnel_user_password: + title: "Password" + description: "OS-level password for logging into the jump server host" + type: "string" + airbyte_secret: true + order: 4 supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -939,22 +900,20 @@ title: "Close.com Spec" type: "object" required: - - "api_key" + - "api_key" additionalProperties: false properties: api_key: type: "string" - description: - "Close.com API key (usually starts with 'api_'; find yours\ + description: "Close.com API key (usually starts with 'api_'; find yours\ \ here)." airbyte_secret: true start_date: type: "string" - description: - "The start date to sync data. Leave blank for full sync. Format:\ + description: "The start date to sync data. Leave blank for full sync. Format:\ \ YYYY-MM-DD." examples: - - "2021-01-01" + - "2021-01-01" default: "2021-01-01" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" supportsNormalization: false @@ -968,10 +927,10 @@ title: "Cockroach Source Spec" type: "object" required: - - "host" - - "port" - - "database" - - "username" + - "host" + - "port" + - "database" + - "username" additionalProperties: false properties: host: @@ -987,7 +946,7 @@ maximum: 65536 default: 5432 examples: - - "5432" + - "5432" order: 1 database: title: "DB Name" @@ -1022,31 +981,31 @@ title: "Commercetools Source CDK Specifications" type: "object" required: - - "region" - - "start_date" - - "host" - - "project_key" - - "client_id" - - "client_secret" + - "region" + - "start_date" + - "host" + - "project_key" + - "client_id" + - "client_secret" additionalProperties: false properties: region: type: "string" description: "The region of the platform." examples: - - "us-central1" - - "australia-southeast1" + - "us-central1" + - "australia-southeast1" host: type: "string" enum: - - "gcp" - - "aws" + - "gcp" + - "aws" description: "The cloud provider your shop is hosted. See: https://docs.commercetools.com/api/authorization" start_date: type: "string" description: "The date you would like to replicate data. Format: YYYY-MM-DD." examples: - - "2021-01-01" + - "2021-01-01" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" project_key: type: "string" @@ -1070,17 +1029,16 @@ title: "Delighted Spec" type: "object" required: - - "since" - - "api_key" + - "since" + - "api_key" additionalProperties: false properties: since: type: "integer" - description: - "An Unix timestamp to retrieve records created on or after\ + description: "An Unix timestamp to retrieve records created on or after\ \ this time." examples: - - 1625328167 + - 1625328167 api_key: type: "string" description: "A Delighted API key." @@ -1096,8 +1054,8 @@ title: "Dixa Spec" type: "object" required: - - "api_token" - - "start_date" + - "api_token" + - "start_date" additionalProperties: false properties: api_token: @@ -1109,14 +1067,14 @@ description: "The connector pulls records updated from this date onwards." pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" examples: - - "YYYY-MM-DD" + - "YYYY-MM-DD" batch_size: type: "integer" description: "Number of days to batch into one request. Max 31." pattern: "^[0-9]{1,2}$" examples: - - 1 - - 31 + - 1 + - 31 default: 31 supportsNormalization: false supportsDBT: false @@ -1129,13 +1087,12 @@ title: "Drift Spec" type: "object" required: - - "access_token" + - "access_token" additionalProperties: false properties: access_token: type: "string" - description: - "Drift Access Token. See the docs for more information on how to generate this key." airbyte_secret: true supportsNormalization: false @@ -1149,8 +1106,8 @@ title: "exchangeratesapi.io Source Spec" type: "object" required: - - "start_date" - - "access_key" + - "start_date" + - "access_key" additionalProperties: false properties: start_date: @@ -1158,22 +1115,20 @@ description: "Start getting data from that date." pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" examples: - - "YYYY-MM-DD" + - "YYYY-MM-DD" access_key: type: "string" - description: - "Your API Access Key. See here. The key is case sensitive." airbyte_secret: true base: type: "string" - description: - "ISO reference currency. See here. Free plan doesn't support Source Currency Switching, default\ \ base currency is EUR" examples: - - "EUR" - - "USD" + - "EUR" + - "USD" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -1187,39 +1142,35 @@ properties: account_id: title: "Account Id" - description: - "The Facebook Ad account ID to use when pulling data from the\ + description: "The Facebook Ad account ID to use when pulling data from the\ \ Facebook Marketing API." type: "string" access_token: title: "Access Token" - description: - "The value of the access token generated. See the docs\ \ for more information" airbyte_secret: true type: "string" start_date: title: "Start Date" - description: - "The date from which you'd like to replicate data for AdCreatives\ + description: "The date from which you'd like to replicate data for AdCreatives\ \ and AdInsights APIs, in the format YYYY-MM-DDT00:00:00Z. All data generated\ \ after this date will be replicated." pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2017-01-25T00:00:00Z" + - "2017-01-25T00:00:00Z" type: "string" format: "date-time" end_date: title: "End Date" - description: - "The date until which you'd like to replicate data for AdCreatives\ + description: "The date until which you'd like to replicate data for AdCreatives\ \ and AdInsights APIs, in the format YYYY-MM-DDT00:00:00Z. All data generated\ \ between start_date and this date will be replicated. Not setting this\ \ option will result in always syncing the latest data." pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2017-01-26T00:00:00Z" + - "2017-01-26T00:00:00Z" type: "string" format: "date-time" include_deleted: @@ -1236,8 +1187,7 @@ type: "integer" insights_days_per_job: title: "Insights Days Per Job" - description: - "Number of days to sync in one job. The more data you have\ + description: "Number of days to sync in one job. The more data you have\ \ - the smaller you want this parameter to be." default: 7 minimum: 1 @@ -1245,8 +1195,7 @@ type: "integer" custom_insights: title: "Custom Insights" - description: - "A list wich contains insights entries, each entry must have\ + description: "A list wich contains insights entries, each entry must have\ \ a name and can contains fields, breakdowns or action_breakdowns)" type: "array" items: @@ -1279,11 +1228,11 @@ items: type: "string" required: - - "name" + - "name" required: - - "account_id" - - "access_token" - - "start_date" + - "account_id" + - "access_token" + - "start_date" definitions: InsightConfig: title: "InsightConfig" @@ -1315,19 +1264,19 @@ items: type: "string" required: - - "name" + - "name" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "append" + - "append" authSpecification: auth_type: "oauth2.0" oauth2Specification: rootObject: [] oauthFlowInitParameters: [] oauthFlowOutputParameters: - - - "access_token" + - - "access_token" - dockerImage: "airbyte/source-facebook-pages:0.1.3" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/facebook-pages" @@ -1336,8 +1285,8 @@ title: "Facebook Pages Spec" type: "object" required: - - "access_token" - - "page_id" + - "access_token" + - "page_id" additionalProperties: true properties: access_token: @@ -1358,7 +1307,7 @@ rootObject: [] oauthFlowInitParameters: [] oauthFlowOutputParameters: - - - "access_token" + - - "access_token" - dockerImage: "airbyte/source-file:0.2.6" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/file" @@ -1368,37 +1317,34 @@ type: "object" additionalProperties: false required: - - "dataset_name" - - "format" - - "url" - - "provider" + - "dataset_name" + - "format" + - "url" + - "provider" properties: dataset_name: type: "string" - description: - "Name of the final table where to replicate this file (should\ + description: "Name of the final table where to replicate this file (should\ \ include only letters, numbers dash and underscores)" format: type: "string" enum: - - "csv" - - "json" - - "jsonl" - - "excel" - - "feather" - - "parquet" + - "csv" + - "json" + - "jsonl" + - "excel" + - "feather" + - "parquet" default: "csv" - description: - "File Format of the file to be replicated (Warning: some format\ + description: "File Format of the file to be replicated (Warning: some format\ \ may be experimental, please refer to docs)." reader_options: type: "string" - description: - "This should be a valid JSON string used by each reader/parser\ + description: "This should be a valid JSON string used by each reader/parser\ \ to provide additional options and tune its behavior" examples: - - "{}" - - "{'sep': ' '}" + - "{}" + - "{'sep': ' '}" url: type: "string" description: "URL path to access the file to be replicated" @@ -1407,164 +1353,157 @@ description: "Storage Provider or Location of the file(s) to be replicated." default: "Public Web" oneOf: - - title: "HTTPS: Public Web" - required: - - "storage" - properties: - storage: - type: "string" - enum: - - "HTTPS" - default: "HTTPS" - - title: "GCS: Google Cloud Storage" - required: - - "storage" - properties: - storage: - type: "string" - enum: - - "GCS" - default: "GCS" - service_account_json: - type: "string" - description: - "In order to access private Buckets stored on Google\ - \ Cloud, this connector would need a service account json credentials\ - \ with the proper permissions as described here. Please generate the credentials.json\ - \ file and copy/paste its content to this field (expecting JSON\ - \ formats). If accessing publicly available data, this field is\ - \ not necessary." - - title: "S3: Amazon Web Services" - required: - - "storage" - properties: - storage: - type: "string" - enum: - - "S3" - default: "S3" - aws_access_key_id: - type: "string" - description: - "In order to access private Buckets stored on AWS S3,\ - \ this connector would need credentials with the proper permissions.\ - \ If accessing publicly available data, this field is not necessary." - aws_secret_access_key: - type: "string" - description: - "In order to access private Buckets stored on AWS S3,\ - \ this connector would need credentials with the proper permissions.\ - \ If accessing publicly available data, this field is not necessary." - airbyte_secret: true - - title: "AzBlob: Azure Blob Storage" - required: - - "storage" - - "storage_account" - properties: - storage: - type: "string" - enum: - - "AzBlob" - default: "AzBlob" - storage_account: - type: "string" - description: - "The globally unique name of the storage account that\ - \ the desired blob sits within. See here for more details." - sas_token: - type: "string" - description: - "To access Azure Blob Storage, this connector would need\ - \ credentials with the proper permissions. One option is a SAS (Shared\ - \ Access Signature) token. If accessing publicly available data,\ - \ this field is not necessary." - airbyte_secret: true - shared_key: - type: "string" - description: - "To access Azure Blob Storage, this connector would need\ - \ credentials with the proper permissions. One option is a storage\ - \ account shared key (aka account key or access key). If accessing\ - \ publicly available data, this field is not necessary." - airbyte_secret: true - - title: "SSH: Secure Shell" - required: - - "storage" - - "user" - - "host" - properties: - storage: - type: "string" - enum: - - "SSH" - default: "SSH" - user: - type: "string" - password: - type: "string" - airbyte_secret: true - host: - type: "string" - port: - type: "string" - default: "22" - - title: "SCP: Secure copy protocol" - required: - - "storage" - - "user" - - "host" - properties: - storage: - type: "string" - enum: - - "SCP" - default: "SCP" - user: - type: "string" - password: - type: "string" - airbyte_secret: true - host: - type: "string" - port: - type: "string" - default: "22" - - title: "SFTP: Secure File Transfer Protocol" - required: - - "storage" - - "user" - - "host" - properties: - storage: - type: "string" - enum: - - "SFTP" - default: "SFTP" - user: - type: "string" - password: - type: "string" - airbyte_secret: true - host: - type: "string" - port: - type: "string" - default: "22" - - title: "Local Filesystem (limited)" - required: - - "storage" - properties: - storage: - type: "string" - description: - "WARNING: Note that local storage URL available for read\ - \ must start with the local mount \"/local/\" at the moment until\ - \ we implement more advanced docker mounting options..." - enum: - - "local" - default: "local" + - title: "HTTPS: Public Web" + required: + - "storage" + properties: + storage: + type: "string" + enum: + - "HTTPS" + default: "HTTPS" + - title: "GCS: Google Cloud Storage" + required: + - "storage" + properties: + storage: + type: "string" + enum: + - "GCS" + default: "GCS" + service_account_json: + type: "string" + description: "In order to access private Buckets stored on Google\ + \ Cloud, this connector would need a service account json credentials\ + \ with the proper permissions as described here. Please generate the credentials.json\ + \ file and copy/paste its content to this field (expecting JSON\ + \ formats). If accessing publicly available data, this field is\ + \ not necessary." + - title: "S3: Amazon Web Services" + required: + - "storage" + properties: + storage: + type: "string" + enum: + - "S3" + default: "S3" + aws_access_key_id: + type: "string" + description: "In order to access private Buckets stored on AWS S3,\ + \ this connector would need credentials with the proper permissions.\ + \ If accessing publicly available data, this field is not necessary." + aws_secret_access_key: + type: "string" + description: "In order to access private Buckets stored on AWS S3,\ + \ this connector would need credentials with the proper permissions.\ + \ If accessing publicly available data, this field is not necessary." + airbyte_secret: true + - title: "AzBlob: Azure Blob Storage" + required: + - "storage" + - "storage_account" + properties: + storage: + type: "string" + enum: + - "AzBlob" + default: "AzBlob" + storage_account: + type: "string" + description: "The globally unique name of the storage account that\ + \ the desired blob sits within. See here for more details." + sas_token: + type: "string" + description: "To access Azure Blob Storage, this connector would need\ + \ credentials with the proper permissions. One option is a SAS (Shared\ + \ Access Signature) token. If accessing publicly available data,\ + \ this field is not necessary." + airbyte_secret: true + shared_key: + type: "string" + description: "To access Azure Blob Storage, this connector would need\ + \ credentials with the proper permissions. One option is a storage\ + \ account shared key (aka account key or access key). If accessing\ + \ publicly available data, this field is not necessary." + airbyte_secret: true + - title: "SSH: Secure Shell" + required: + - "storage" + - "user" + - "host" + properties: + storage: + type: "string" + enum: + - "SSH" + default: "SSH" + user: + type: "string" + password: + type: "string" + airbyte_secret: true + host: + type: "string" + port: + type: "string" + default: "22" + - title: "SCP: Secure copy protocol" + required: + - "storage" + - "user" + - "host" + properties: + storage: + type: "string" + enum: + - "SCP" + default: "SCP" + user: + type: "string" + password: + type: "string" + airbyte_secret: true + host: + type: "string" + port: + type: "string" + default: "22" + - title: "SFTP: Secure File Transfer Protocol" + required: + - "storage" + - "user" + - "host" + properties: + storage: + type: "string" + enum: + - "SFTP" + default: "SFTP" + user: + type: "string" + password: + type: "string" + airbyte_secret: true + host: + type: "string" + port: + type: "string" + default: "22" + - title: "Local Filesystem (limited)" + required: + - "storage" + properties: + storage: + type: "string" + description: "WARNING: Note that local storage URL available for read\ + \ must start with the local mount \"/local/\" at the moment until\ + \ we implement more advanced docker mounting options..." + enum: + - "local" + default: "local" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -1576,28 +1515,26 @@ title: "Freshdesk Spec" type: "object" required: - - "domain" - - "api_key" + - "domain" + - "api_key" additionalProperties: false properties: domain: type: "string" description: "Freshdesk domain" examples: - - "myaccount.freshdesk.com" + - "myaccount.freshdesk.com" pattern: - - "^[a-zA-Z0-9._-]*\\.freshdesk\\.com$" + - "^[a-zA-Z0-9._-]*\\.freshdesk\\.com$" api_key: type: "string" - description: - "Freshdesk API Key. See the docs for more information on how to obtain this key." airbyte_secret: true requests_per_minute: title: "Requests per minute" type: "integer" - description: - "Number of requests per minute that this source allowed to\ + description: "Number of requests per minute that this source allowed to\ \ use." start_date: title: "Start date" @@ -1605,7 +1542,7 @@ format: "date-time" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2020-12-01T00:00:00Z" + - "2020-12-01T00:00:00Z" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -1617,19 +1554,18 @@ title: "Freshsales Spec" type: "object" required: - - "domain_name" - - "api_key" + - "domain_name" + - "api_key" additionalProperties: false properties: domain_name: type: "string" description: "Freshsales domain" examples: - - "mydomain.myfreshworks.com" + - "mydomain.myfreshworks.com" api_key: type: "string" - description: - "Your API Access Key. See here. The key is case sensitive." airbyte_secret: true supportsNormalization: false @@ -1643,31 +1579,29 @@ title: "Freshservice Spec" type: "object" required: - - "domain_name" - - "api_key" - - "start_date" + - "domain_name" + - "api_key" + - "start_date" additionalProperties: false properties: domain_name: type: "string" description: "Freshservice domain" examples: - - "mydomain.freshservice.com" + - "mydomain.freshservice.com" api_key: title: "Api Key" type: "string" - description: - "Your API Access Key. See here. The key is case sensitive." airbyte_secret: true start_date: title: "Replication Start Date" type: "string" - description: - "UTC date and time in the format 2020-10-01T00:00:00Z. Any\ + description: "UTC date and time in the format 2020-10-01T00:00:00Z. Any\ \ data before this date will not be replicated." examples: - - "2020-10-01T00:00:00Z" + - "2020-10-01T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" supportsNormalization: false supportsDBT: false @@ -1680,8 +1614,8 @@ title: "Github Source Spec" type: "object" required: - - "start_date" - - "repository" + - "start_date" + - "repository" additionalProperties: true properties: credentials: @@ -1689,68 +1623,64 @@ description: "Choose how to authenticate to Github" type: "object" oneOf: - - type: "object" - title: "Authenticate via Github (Oauth)" - required: - - "access_token" - properties: - option_title: - type: "string" - title: "Credentials title" - description: "OAuth Credentials" - const: "OAuth Credentials" - access_token: - type: "string" - title: "Access Token" - description: "Oauth access token" - airbyte_secret: true - - type: "object" - title: "Authenticate with Personal Access Token" - required: - - "personal_access_token" - properties: - option_title: - type: "string" - title: "Credentials title" - description: "PAT Credentials" - const: "PAT Credentials" - personal_access_token: - type: "string" - title: "Personal Access Tokens" - description: - "Log into Github and then generate a personal access token. To load balance your API quota consumption\ - \ across multiple API tokens, input multiple tokens separated with\ - \ \",\"" - airbyte_secret: true + - type: "object" + title: "Authenticate via Github (Oauth)" + required: + - "access_token" + properties: + option_title: + type: "string" + title: "Credentials title" + description: "OAuth Credentials" + const: "OAuth Credentials" + access_token: + type: "string" + title: "Access Token" + description: "Oauth access token" + airbyte_secret: true + - type: "object" + title: "Authenticate with Personal Access Token" + required: + - "personal_access_token" + properties: + option_title: + type: "string" + title: "Credentials title" + description: "PAT Credentials" + const: "PAT Credentials" + personal_access_token: + type: "string" + title: "Personal Access Tokens" + description: "Log into Github and then generate a personal access token. To load balance your API quota consumption\ + \ across multiple API tokens, input multiple tokens separated with\ + \ \",\"" + airbyte_secret: true repository: type: "string" examples: - - "airbytehq/airbyte" - - "airbytehq/*" + - "airbytehq/airbyte" + - "airbytehq/*" title: "Github repositories" - description: - "Space-delimited list of GitHub repositories/organizations,\ + description: "Space-delimited list of GitHub repositories/organizations,\ \ e.g. `airbytehq/airbyte` for single repository and `airbytehq/*` for\ \ get all repositories from organization" start_date: type: "string" title: "Start date" - description: - "The date from which you'd like to replicate data for GitHub\ + description: "The date from which you'd like to replicate data for GitHub\ \ in the format YYYY-MM-DDT00:00:00Z. All data generated after this date\ \ will be replicated. Note that it will be used only in the following\ \ incremental streams: comments, commits and issues." examples: - - "2021-03-01T00:00:00Z" + - "2021-03-01T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" branch: type: "string" title: "Branch" examples: - - "airbytehq/airbyte/master" - description: - "Space-delimited list of GitHub repository branches to pull\ + - "airbytehq/airbyte/master" + description: "Space-delimited list of GitHub repository branches to pull\ \ commits for, e.g. `airbytehq/airbyte/master`. If no branches are specified\ \ for a repository, the default branch will be pulled." supportsNormalization: false @@ -1760,11 +1690,11 @@ auth_type: "oauth2.0" oauth2Specification: rootObject: - - "credentials" - - "0" + - "credentials" + - "0" oauthFlowInitParameters: [] oauthFlowOutputParameters: - - - "access_token" + - - "access_token" - dockerImage: "airbyte/source-gitlab:0.1.2" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/gitlab" @@ -1773,42 +1703,39 @@ title: "Source Gitlab Singer Spec" type: "object" required: - - "api_url" - - "private_token" - - "start_date" + - "api_url" + - "private_token" + - "start_date" additionalProperties: false properties: api_url: type: "string" examples: - - "gitlab.com" + - "gitlab.com" description: "Please enter your basic URL from Gitlab instance" private_token: type: "string" - description: - "Log into your Gitlab account and then generate a personal\ + description: "Log into your Gitlab account and then generate a personal\ \ Access Token." airbyte_secret: true groups: type: "string" examples: - - "airbyte.io" + - "airbyte.io" description: "Space-delimited list of groups. e.g. airbyte.io" projects: type: "string" examples: - - "airbyte.io/documentation" - description: - "Space-delimited list of projects. e.g. airbyte.io/documentation\ + - "airbyte.io/documentation" + description: "Space-delimited list of projects. e.g. airbyte.io/documentation\ \ meltano/tap-gitlab" start_date: type: "string" - description: - "The date from which you'd like to replicate data for Gitlab\ + description: "The date from which you'd like to replicate data for Gitlab\ \ API, in the format YYYY-MM-DDT00:00:00Z. All data generated after this\ \ date will be replicated." examples: - - "2021-03-01T00:00:00Z" + - "2021-03-01T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" supportsNormalization: false supportsDBT: false @@ -1821,48 +1748,44 @@ title: "Google Ads Spec" type: "object" required: - - "credentials" - - "start_date" - - "customer_id" + - "credentials" + - "start_date" + - "customer_id" additionalProperties: true properties: credentials: type: "object" title: "Google Credentials" required: - - "developer_token" - - "client_id" - - "client_secret" - - "refresh_token" + - "developer_token" + - "client_id" + - "client_secret" + - "refresh_token" properties: developer_token: type: "string" title: "Developer Token" - description: - "Developer token granted by Google to use their APIs. More\ + description: "Developer token granted by Google to use their APIs. More\ \ instruction on how to find this value in our docs" airbyte_secret: true client_id: type: "string" title: "Client Id" - description: - "Google client id. More instruction on how to find this\ + description: "Google client id. More instruction on how to find this\ \ value in our docs" client_secret: type: "string" title: "Client Secret" - description: - "Google client secret. More instruction on how to find\ + description: "Google client secret. More instruction on how to find\ \ this value in our docs" airbyte_secret: true access_token: type: "string" title: "Access Token" - description: - "Access token generated using developer_token, oauth_client_id,\ + description: "Access token generated using developer_token, oauth_client_id,\ \ and oauth_client_secret. More instruction on how to find this value\ \ in our docs" @@ -1870,8 +1793,7 @@ refresh_token: type: "string" title: "Refresh Token" - description: - "Refresh token generated using developer_token, oauth_client_id,\ + description: "Refresh token generated using developer_token, oauth_client_id,\ \ and oauth_client_secret. More instruction on how to find this value\ \ in our docs" @@ -1879,16 +1801,14 @@ customer_id: title: "Customer Id" type: "string" - description: - "Customer id must be specified as a 10-digit number without\ + description: "Customer id must be specified as a 10-digit number without\ \ dashes. More instruction on how to find this value in our docs" login_customer_id: type: "string" title: "Login Customer ID" - description: - "If your access to the customer account is through a manager\ + description: "If your access to the customer account is through a manager\ \ account, this field is required and must be set to the customer ID of\ \ the manager account (10-digit number without dashes). More information\ \ about this field you can see Google Analytics Account Explorer." airbyte_secret: true @@ -1973,88 +1889,84 @@ title: "Start Date" description: "A date in the format YYYY-MM-DD." examples: - - "2020-06-01" + - "2020-06-01" window_in_days: type: "integer" - description: - "The amount of days for each data-chunk begining from start_date.\ + description: "The amount of days for each data-chunk begining from start_date.\ \ Bigger the value - faster the fetch. (Min=1, as for a Day; Max=364,\ \ as for a Year)." examples: - - 30 - - 60 - - 90 - - 120 - - 200 - - 364 + - 30 + - 60 + - 90 + - 120 + - 200 + - 364 default: 90 custom_reports: title: "Custom Reports" type: "string" - description: - "A JSON array describing the custom reports you want to sync\ + description: "A JSON array describing the custom reports you want to sync\ \ from GA. Check out the docs to get more information about this field." credentials: type: "object" oneOf: - - title: "Authenticate via Google (Oauth)" - type: "object" - required: - - "client_id" - - "client_secret" - - "refresh_token" - properties: - auth_type: - type: "string" - const: "Client" - enum: - - "Client" - default: "Client" - order: 0 - client_id: - title: "Client ID" - type: "string" - description: "The Client ID of your developer application" - airbyte_secret: true - client_secret: - title: "Client Secret" - type: "string" - description: "The client secret of your developer application" - airbyte_secret: true - refresh_token: - title: "Refresh Token" - type: "string" - description: - "A refresh token generated using the above client ID\ - \ and secret" - airbyte_secret: true - access_token: - title: "Access Token" - type: "string" - description: - "A access token generated using the above client ID,\ - \ secret and refresh_token" - airbyte_secret: true - - type: "object" - title: "Service Account Key Authentication" - required: - - "credentials_json" - properties: - auth_type: - type: "string" - const: "Service" - enum: - - "Service" - default: "Service" - order: 0 - credentials_json: - type: "string" - description: "The JSON key of the service account to use for authorization" - examples: - - "{ \"type\": \"service_account\", \"project_id\": YOUR_PROJECT_ID,\ - \ \"private_key_id\": YOUR_PRIVATE_KEY, ... }" - airbyte_secret: true + - title: "Authenticate via Google (Oauth)" + type: "object" + required: + - "client_id" + - "client_secret" + - "refresh_token" + properties: + auth_type: + type: "string" + const: "Client" + enum: + - "Client" + default: "Client" + order: 0 + client_id: + title: "Client ID" + type: "string" + description: "The Client ID of your developer application" + airbyte_secret: true + client_secret: + title: "Client Secret" + type: "string" + description: "The client secret of your developer application" + airbyte_secret: true + refresh_token: + title: "Refresh Token" + type: "string" + description: "A refresh token generated using the above client ID\ + \ and secret" + airbyte_secret: true + access_token: + title: "Access Token" + type: "string" + description: "A access token generated using the above client ID,\ + \ secret and refresh_token" + airbyte_secret: true + - type: "object" + title: "Service Account Key Authentication" + required: + - "credentials_json" + properties: + auth_type: + type: "string" + const: "Service" + enum: + - "Service" + default: "Service" + order: 0 + credentials_json: + type: "string" + description: "The JSON key of the service account to use for authorization" + examples: + - "{ \"type\": \"service_account\", \"project_id\": YOUR_PROJECT_ID,\ + \ \"private_key_id\": YOUR_PRIVATE_KEY, ... }" + airbyte_secret: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -2062,14 +1974,14 @@ auth_type: "oauth2.0" oauth2Specification: rootObject: - - "credentials" - - "0" + - "credentials" + - "0" oauthFlowInitParameters: - - - "client_id" - - - "client_secret" + - - "client_id" + - - "client_secret" oauthFlowOutputParameters: - - - "access_token" - - - "refresh_token" + - - "access_token" + - - "refresh_token" - dockerImage: "airbyte/source-google-directory:0.1.8" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/google-directory" @@ -2082,79 +1994,74 @@ properties: credentials: title: "Google Credentials" - description: - "Google APIs use the OAuth 2.0 protocol for authentication\ + description: "Google APIs use the OAuth 2.0 protocol for authentication\ \ and authorization. The Source supports Web server application and Service accounts scenarios" type: "object" oneOf: - - title: "Sign in via Google (Oauth)" - description: - "For these scenario user only needs to give permission to\ - \ read Google Directory data" - type: "object" - required: - - "client_id" - - "client_secret" - - "refresh_token" - properties: - credentials_title: - type: "string" - title: "Credentials title" - description: "Authentication scenario" - const: "Web server app" - enum: - - "Web server app" - default: "Web server app" - order: 0 - client_id: - title: "Client ID" - type: "string" - description: "The client ID of developer application" - airbyte_secret: true - client_secret: - title: "Client secret" - type: "string" - description: "The client secret of developer application" - airbyte_secret: true - refresh_token: - title: "Refresh Token" - type: "string" - description: "The token for obtaining new access token" - airbyte_secret: true - - title: "Service account Key" - description: - "For these scenario user should obtain service account's\ - \ credentials from the Google API Console and provide delegated email" - type: "object" - required: - - "credentials_json" - - "email" - properties: - credentials_title: - type: "string" - title: "Credentials title" - description: "Authentication scenario" - const: "Service accounts" - enum: - - "Service accounts" - default: "Service accounts" - order: 0 - credentials_json: - type: "string" - title: "Credentials JSON" - description: - "The contents of the JSON service account key. See the\ - \ docs for more information on how to generate this key." - airbyte_secret: true - email: - type: "string" - title: "Email" - description: - "The email of the user, which has permissions to access\ - \ the Google Workspace Admin APIs." + - title: "Sign in via Google (Oauth)" + description: "For these scenario user only needs to give permission to\ + \ read Google Directory data" + type: "object" + required: + - "client_id" + - "client_secret" + - "refresh_token" + properties: + credentials_title: + type: "string" + title: "Credentials title" + description: "Authentication scenario" + const: "Web server app" + enum: + - "Web server app" + default: "Web server app" + order: 0 + client_id: + title: "Client ID" + type: "string" + description: "The client ID of developer application" + airbyte_secret: true + client_secret: + title: "Client secret" + type: "string" + description: "The client secret of developer application" + airbyte_secret: true + refresh_token: + title: "Refresh Token" + type: "string" + description: "The token for obtaining new access token" + airbyte_secret: true + - title: "Service account Key" + description: "For these scenario user should obtain service account's\ + \ credentials from the Google API Console and provide delegated email" + type: "object" + required: + - "credentials_json" + - "email" + properties: + credentials_title: + type: "string" + title: "Credentials title" + description: "Authentication scenario" + const: "Service accounts" + enum: + - "Service accounts" + default: "Service accounts" + order: 0 + credentials_json: + type: "string" + title: "Credentials JSON" + description: "The contents of the JSON service account key. See the\ + \ docs for more information on how to generate this key." + airbyte_secret: true + email: + type: "string" + title: "Email" + description: "The email of the user, which has permissions to access\ + \ the Google Workspace Admin APIs." supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -2162,13 +2069,13 @@ auth_type: "oauth2.0" oauth2Specification: rootObject: - - "credentials" - - "0" + - "credentials" + - "0" oauthFlowInitParameters: - - - "client_id" - - - "client_secret" + - - "client_id" + - - "client_secret" oauthFlowOutputParameters: - - - "refresh_token" + - - "refresh_token" - dockerImage: "airbyte/source-google-search-console:0.1.6" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/google-search-console" @@ -2178,106 +2085,100 @@ type: "object" additionalProperties: false required: - - "site_urls" - - "start_date" - - "authorization" + - "site_urls" + - "start_date" + - "authorization" properties: site_urls: type: "array" items: type: "string" - description: - "Website URLs property; do not include the domain-level property\ + description: "Website URLs property; do not include the domain-level property\ \ in the list" examples: - - "https://example1.com" - - "https://example2.com" + - "https://example1.com" + - "https://example2.com" start_date: type: "string" - description: - "The date from which you'd like to replicate data in the format\ + description: "The date from which you'd like to replicate data in the format\ \ YYYY-MM-DD." examples: - - "2021-01-01" + - "2021-01-01" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" end_date: type: "string" - description: - "The date from which you'd like to replicate data in the format\ + description: "The date from which you'd like to replicate data in the format\ \ YYYY-MM-DD. Must be greater or equal start_date field" examples: - - "2021-12-12" + - "2021-12-12" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" authorization: type: "object" title: "Authentication Type" oneOf: - - title: "Authenticate via Google (Oauth)" - type: "object" - required: - - "auth_type" - - "client_id" - - "client_secret" - - "refresh_token" - properties: - auth_type: - type: "string" - const: "Client" - enum: - - "Client" - default: "Client" - order: 0 - client_id: - title: "Client ID" - type: "string" - description: "The Client ID of your developer application" - airbyte_secret: true - client_secret: - title: "Client Secret" - type: "string" - description: "The client secret of your developer application" - airbyte_secret: true - access_token: - title: "Access Token" - type: "string" - description: - "An access token generated using the above client ID\ - \ and secret" - airbyte_secret: true - refresh_token: - title: "Refresh Token" - type: "string" - description: - "A refresh token generated using the above client ID\ - \ and secret" - airbyte_secret: true - - type: "object" - title: "Service Account Key Authentication" - required: - - "auth_type" - - "service_account_info" - - "email" - properties: - auth_type: - type: "string" - const: "Service" - enum: - - "Service" - default: "Service" - order: 0 - service_account_info: - title: "Service Account JSON Key" - type: "string" - description: "The JSON key of the service account to use for authorization" - examples: - - "{ \"type\": \"service_account\", \"project_id\": YOUR_PROJECT_ID,\ - \ \"private_key_id\": YOUR_PRIVATE_KEY, ... }" - email: - title: "Admin Email" - type: "string" - description: - "The email of the user which has permissions to access\ - \ the Google Workspace Admin APIs." + - title: "Authenticate via Google (Oauth)" + type: "object" + required: + - "auth_type" + - "client_id" + - "client_secret" + - "refresh_token" + properties: + auth_type: + type: "string" + const: "Client" + enum: + - "Client" + default: "Client" + order: 0 + client_id: + title: "Client ID" + type: "string" + description: "The Client ID of your developer application" + airbyte_secret: true + client_secret: + title: "Client Secret" + type: "string" + description: "The client secret of your developer application" + airbyte_secret: true + access_token: + title: "Access Token" + type: "string" + description: "An access token generated using the above client ID\ + \ and secret" + airbyte_secret: true + refresh_token: + title: "Refresh Token" + type: "string" + description: "A refresh token generated using the above client ID\ + \ and secret" + airbyte_secret: true + - type: "object" + title: "Service Account Key Authentication" + required: + - "auth_type" + - "service_account_info" + - "email" + properties: + auth_type: + type: "string" + const: "Service" + enum: + - "Service" + default: "Service" + order: 0 + service_account_info: + title: "Service Account JSON Key" + type: "string" + description: "The JSON key of the service account to use for authorization" + examples: + - "{ \"type\": \"service_account\", \"project_id\": YOUR_PROJECT_ID,\ + \ \"private_key_id\": YOUR_PRIVATE_KEY, ... }" + email: + title: "Admin Email" + type: "string" + description: "The email of the user which has permissions to access\ + \ the Google Workspace Admin APIs." supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -2285,14 +2186,14 @@ auth_type: "oauth2.0" oauth2Specification: rootObject: - - "authorization" - - "0" + - "authorization" + - "0" oauthFlowInitParameters: - - - "client_id" - - - "client_secret" + - - "client_id" + - - "client_secret" oauthFlowOutputParameters: - - - "access_token" - - - "refresh_token" + - - "access_token" + - - "refresh_token" - dockerImage: "airbyte/source-google-sheets:0.2.6" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/google-sheets" @@ -2301,7 +2202,7 @@ title: "Stripe Source Spec" type: "object" required: - - "spreadsheet_id" + - "spreadsheet_id" additionalProperties: true properties: spreadsheet_id: @@ -2310,49 +2211,48 @@ credentials: type: "object" oneOf: - - title: "Authenticate via Google (Oauth)" - type: "object" - required: - - "auth_type" - - "client_id" - - "client_secret" - - "refresh_token" - properties: - auth_type: - type: "string" - const: "Client" - client_id: - title: "Client ID" - type: "string" - description: "The Client ID of your developer application" - airbyte_secret: true - client_secret: - title: "Client Secret" - type: "string" - description: "The client secret of your developer application" - airbyte_secret: true - refresh_token: - title: "Refresh Token" - type: "string" - description: - "A refresh token generated using the above client ID\ - \ and secret" - airbyte_secret: true - - title: "Service Account Key Authentication" - type: "object" - required: - - "auth_type" - - "service_account_info" - properties: - auth_type: - type: "string" - const: "Service" - service_account_info: - type: "string" - description: "The JSON key of the service account to use for authorization" - examples: - - "{ \"type\": \"service_account\", \"project_id\": YOUR_PROJECT_ID,\ - \ \"private_key_id\": YOUR_PRIVATE_KEY, ... }" + - title: "Authenticate via Google (Oauth)" + type: "object" + required: + - "auth_type" + - "client_id" + - "client_secret" + - "refresh_token" + properties: + auth_type: + type: "string" + const: "Client" + client_id: + title: "Client ID" + type: "string" + description: "The Client ID of your developer application" + airbyte_secret: true + client_secret: + title: "Client Secret" + type: "string" + description: "The client secret of your developer application" + airbyte_secret: true + refresh_token: + title: "Refresh Token" + type: "string" + description: "A refresh token generated using the above client ID\ + \ and secret" + airbyte_secret: true + - title: "Service Account Key Authentication" + type: "object" + required: + - "auth_type" + - "service_account_info" + properties: + auth_type: + type: "string" + const: "Service" + service_account_info: + type: "string" + description: "The JSON key of the service account to use for authorization" + examples: + - "{ \"type\": \"service_account\", \"project_id\": YOUR_PROJECT_ID,\ + \ \"private_key_id\": YOUR_PRIVATE_KEY, ... }" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -2360,13 +2260,13 @@ auth_type: "oauth2.0" oauth2Specification: rootObject: - - "credentials" - - 0 + - "credentials" + - 0 oauthFlowInitParameters: - - - "client_id" - - - "client_secret" + - - "client_id" + - - "client_secret" oauthFlowOutputParameters: - - - "refresh_token" + - - "refresh_token" - dockerImage: "airbyte/source-google-workspace-admin-reports:0.1.6" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/google-workspace-admin-reports" @@ -2375,28 +2275,25 @@ title: "Google Directory Spec" type: "object" required: - - "credentials_json" - - "email" + - "credentials_json" + - "email" additionalProperties: false properties: credentials_json: type: "string" - description: - "The contents of the JSON service account key. See the docs for more information on how to generate this key." airbyte_secret: true email: type: "string" - description: - "The email of the user, which has permissions to access the\ + description: "The email of the user, which has permissions to access the\ \ Google Workspace Admin APIs." lookback: type: "integer" minimum: 0 maximum: 180 - description: - "Sets the range of time shown in the report. Reports API allows\ + description: "Sets the range of time shown in the report. Reports API allows\ \ from up to 180 days ago. " supportsNormalization: false supportsDBT: false @@ -2409,13 +2306,12 @@ title: "Greenhouse Spec" type: "object" required: - - "api_key" + - "api_key" additionalProperties: false properties: api_key: type: "string" - description: - "Greenhouse API Key. See the docs for more information on how to generate this key." airbyte_secret: true supportsNormalization: false @@ -2429,9 +2325,9 @@ title: "Harvest Spec" type: "object" required: - - "api_token" - - "account_id" - - "replication_start_date" + - "api_token" + - "account_id" + - "replication_start_date" additionalProperties: false properties: api_token: @@ -2441,25 +2337,23 @@ type: "string" account_id: title: "Account ID" - description: - "Harvest account ID. Required for all Harvest requests in pair\ + description: "Harvest account ID. Required for all Harvest requests in pair\ \ with API Key" airbyte_secret: true type: "string" replication_start_date: title: "Replication Start Date" - description: - "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ \ data before this date will not be replicated." pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2017-01-25T00:00:00Z" + - "2017-01-25T00:00:00Z" type: "string" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "append" + - "append" - dockerImage: "airbyte/source-hubspot:0.1.24" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/hubspot" @@ -2468,89 +2362,84 @@ title: "HubSpot Source Spec" type: "object" required: - - "start_date" - - "credentials" + - "start_date" + - "credentials" additionalProperties: false properties: start_date: type: "string" title: "Replication start date" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" - description: - "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ \ data before this date will not be replicated." examples: - - "2017-01-25T00:00:00Z" + - "2017-01-25T00:00:00Z" credentials: title: "Authentication mechanism" description: "Choose either to provide the API key or the OAuth2.0 credentials" type: "object" oneOf: - - type: "object" - title: "Authenticate via HubSpot (Oauth)" - required: - - "client_id" - - "client_secret" + - type: "object" + title: "Authenticate via HubSpot (Oauth)" + required: + - "client_id" + - "client_secret" + - "refresh_token" + - "credentials_title" + properties: + credentials_title: + type: "string" + title: "Credentials title" + description: "Name of the credentials set" + const: "OAuth Credentials" + enum: + - "OAuth Credentials" + default: "OAuth Credentials" + order: 0 + client_id: + title: "Client ID" + description: "HubSpot client_id. See our docs if you need help finding this id." + type: "string" + examples: + - "123456789000" + client_secret: + title: "Client Secret" + description: "HubSpot client_secret. See our docs if you need help finding this secret." + type: "string" + examples: + - "secret" + airbyte_secret: true + refresh_token: + title: "Refresh token" + description: "HubSpot refresh_token. See our docs if you need help generating the token." + type: "string" + examples: - "refresh_token" - - "credentials_title" - properties: - credentials_title: - type: "string" - title: "Credentials title" - description: "Name of the credentials set" - const: "OAuth Credentials" - enum: - - "OAuth Credentials" - default: "OAuth Credentials" - order: 0 - client_id: - title: "Client ID" - description: - "HubSpot client_id. See our docs if you need help finding this id." - type: "string" - examples: - - "123456789000" - client_secret: - title: "Client Secret" - description: - "HubSpot client_secret. See our docs if you need help finding this secret." - type: "string" - examples: - - "secret" - airbyte_secret: true - refresh_token: - title: "Refresh token" - description: - "HubSpot refresh_token. See our docs if you need help generating the token." - type: "string" - examples: - - "refresh_token" - airbyte_secret: true - - type: "object" - title: "API key" - required: - - "api_key" - - "credentials_title" - properties: - credentials_title: - type: "string" - title: "Credentials title" - description: "Name of the credentials set" - const: "API Key Credentials" - enum: - - "API Key Credentials" - default: "API Key Credentials" - order: 0 - api_key: - title: "API key" - description: - "HubSpot API Key. See our docs if you need help finding this key." - type: "string" - airbyte_secret: true + airbyte_secret: true + - type: "object" + title: "API key" + required: + - "api_key" + - "credentials_title" + properties: + credentials_title: + type: "string" + title: "Credentials title" + description: "Name of the credentials set" + const: "API Key Credentials" + enum: + - "API Key Credentials" + default: "API Key Credentials" + order: 0 + api_key: + title: "API key" + description: "HubSpot API Key. See our docs if you need help finding this key." + type: "string" + airbyte_secret: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -2558,13 +2447,13 @@ auth_type: "oauth2.0" oauth2Specification: rootObject: - - "credentials" - - "0" + - "credentials" + - "0" oauthFlowInitParameters: - - - "client_id" - - - "client_secret" + - - "client_id" + - - "client_secret" oauthFlowOutputParameters: - - - "refresh_token" + - - "refresh_token" - dockerImage: "airbyte/source-db2:0.1.2" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/db2" @@ -2573,12 +2462,12 @@ title: "IBM Db2 Source Spec" type: "object" required: - - "host" - - "port" - - "db" - - "username" - - "password" - - "encryption" + - "host" + - "port" + - "db" + - "username" + - "password" + - "encryption" additionalProperties: false properties: host: @@ -2592,13 +2481,13 @@ maximum: 65536 default: 8123 examples: - - "8123" + - "8123" order: 1 db: description: "Name of the database." type: "string" examples: - - "default" + - "default" order: 2 username: description: "Username to use to access the database." @@ -2615,46 +2504,44 @@ description: "Encryption method to use when communicating with the database" order: 5 oneOf: - - title: "Unencrypted" - additionalProperties: false - description: "Data transfer will not be encrypted." - required: - - "encryption_method" - properties: - encryption_method: - type: "string" - const: "unencrypted" - enum: - - "unencrypted" - default: "unencrypted" - - title: "TLS Encrypted (verify certificate)" - additionalProperties: false - description: "Verify and use the cert provided by the server." - required: - - "encryption_method" - - "ssl_certificate" - properties: - encryption_method: - type: "string" - const: "encrypted_verify_certificate" - enum: - - "encrypted_verify_certificate" - default: "encrypted_verify_certificate" - ssl_certificate: - title: "SSL PEM file" - description: - "Privacy Enhanced Mail (PEM) files are concatenated certificate\ - \ containers frequently used in certificate installations" - type: "string" - airbyte_secret: true - multiline: true - key_store_password: - title: - "Key Store Password. This field is optional. If you do not\ - \ fill in this field, the password will be randomly generated." - description: "Key Store Password" - type: "string" - airbyte_secret: true + - title: "Unencrypted" + additionalProperties: false + description: "Data transfer will not be encrypted." + required: + - "encryption_method" + properties: + encryption_method: + type: "string" + const: "unencrypted" + enum: + - "unencrypted" + default: "unencrypted" + - title: "TLS Encrypted (verify certificate)" + additionalProperties: false + description: "Verify and use the cert provided by the server." + required: + - "encryption_method" + - "ssl_certificate" + properties: + encryption_method: + type: "string" + const: "encrypted_verify_certificate" + enum: + - "encrypted_verify_certificate" + default: "encrypted_verify_certificate" + ssl_certificate: + title: "SSL PEM file" + description: "Privacy Enhanced Mail (PEM) files are concatenated certificate\ + \ containers frequently used in certificate installations" + type: "string" + airbyte_secret: true + multiline: true + key_store_password: + title: "Key Store Password. This field is optional. If you do not\ + \ fill in this field, the password will be randomly generated." + description: "Key Store Password" + type: "string" + airbyte_secret: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -2668,38 +2555,36 @@ properties: start_date: title: "Start Date" - description: - "The date from which you'd like to replicate data for User\ + description: "The date from which you'd like to replicate data for User\ \ Insights, in the format YYYY-MM-DDT00:00:00Z. All data generated after\ \ this date will be replicated." pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2017-01-25T00:00:00Z" + - "2017-01-25T00:00:00Z" type: "string" format: "date-time" access_token: title: "Access Token" - description: - "The value of the access token generated. See the docs for\ \ more information" airbyte_secret: true type: "string" required: - - "start_date" - - "access_token" + - "start_date" + - "access_token" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "append" + - "append" authSpecification: auth_type: "oauth2.0" oauth2Specification: rootObject: [] oauthFlowInitParameters: [] oauthFlowOutputParameters: - - - "access_token" + - - "access_token" - dockerImage: "airbyte/source-intercom:0.1.8" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/intercom" @@ -2708,24 +2593,22 @@ title: "Source Intercom Spec" type: "object" required: - - "start_date" - - "access_token" + - "start_date" + - "access_token" additionalProperties: true properties: start_date: type: "string" - description: - "The date from which you'd like to replicate data for Intercom\ + description: "The date from which you'd like to replicate data for Intercom\ \ API, in the format YYYY-MM-DDT00:00:00Z. All data generated after this\ \ date will be replicated." examples: - - "2020-11-16T00:00:00Z" + - "2020-11-16T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" access_token: title: "Access Token" type: "string" - description: - "Access token generated either from an oauth flow or from the\ + description: "Access token generated either from an oauth flow or from the\ \ Intercom Developer dashboard. See the docs for more information on how to obtain this key manually." airbyte_secret: true @@ -2738,7 +2621,7 @@ rootObject: [] oauthFlowInitParameters: [] oauthFlowOutputParameters: - - - "access_token" + - - "access_token" - dockerImage: "airbyte/source-iterable:0.1.12" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/iterable" @@ -2747,23 +2630,21 @@ title: "Iterable Spec" type: "object" required: - - "start_date" - - "api_key" + - "start_date" + - "api_key" additionalProperties: false properties: start_date: type: "string" - description: - "The date from which you'd like to replicate data for Iterable,\ + description: "The date from which you'd like to replicate data for Iterable,\ \ in the format YYYY-MM-DDT00:00:00Z. All data generated after this date\ \ will be replicated." examples: - - "2021-04-01T00:00:00Z" + - "2021-04-01T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" api_key: type: "string" - description: - "Iterable API Key. See the docs for more information on how to obtain this key." airbyte_secret: true supportsNormalization: false @@ -2777,21 +2658,20 @@ title: "Jira Spec" type: "object" required: - - "api_token" - - "domain" - - "email" + - "api_token" + - "domain" + - "email" additionalProperties: true properties: api_token: type: "string" - description: - "Jira API Token. See the docs for more information on how to generate this key." airbyte_secret: true domain: type: "string" examples: - - "domainname.atlassian.net" + - "domainname.atlassian.net" pattern: "^[a-zA-Z0-9._-]*\\.atlassian\\.net$" description: "Domain for your Jira account, e.g. airbyteio.atlassian.net" email: @@ -2803,33 +2683,30 @@ items: type: "string" examples: - - "PROJ1" - - "PROJ2" - description: - "Comma-separated list of Jira project keys to replicate data\ + - "PROJ1" + - "PROJ2" + description: "Comma-separated list of Jira project keys to replicate data\ \ for" start_date: type: "string" title: "Start Date" - description: - "The date from which you'd like to replicate data for Jira\ + description: "The date from which you'd like to replicate data for Jira\ \ in the format YYYY-MM-DDT00:00:00Z. All data generated after this date\ \ will be replicated. Note that it will be used only in the following\ \ incremental streams: issues." examples: - - "2021-03-01T00:00:00Z" + - "2021-03-01T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" additional_fields: type: "array" title: "Additional Fields" items: type: "string" - description: - "Comma-separated list of additional fields to include in replicating\ + description: "Comma-separated list of additional fields to include in replicating\ \ issues" examples: - - "Field A" - - "Field B" + - "Field A" + - "Field B" expand_issue_changelog: type: "boolean" title: "Expand Issue Changelog" @@ -2846,15 +2723,14 @@ title: "Kafka Source Spec" type: "object" required: - - "bootstrap_servers" - - "subscription" - - "protocol" + - "bootstrap_servers" + - "subscription" + - "protocol" additionalProperties: false properties: bootstrap_servers: title: "Bootstrap servers" - description: - "A list of host/port pairs to use for establishing the initial\ + description: "A list of host/port pairs to use for establishing the initial\ \ connection to the Kafka cluster. The client will make use of all servers\ \ irrespective of which servers are specified here for bootstrapping—this\ \ list only impacts the initial hosts used to discover the full set of\ @@ -2865,69 +2741,66 @@ \ though, in case a server is down)." type: "string" examples: - - "kafka-broker1:9092,kafka-broker2:9092" + - "kafka-broker1:9092,kafka-broker2:9092" subscription: title: "Subscribe method" type: "object" - description: - "You can choose to manually assign a list of partitions, or\ + description: "You can choose to manually assign a list of partitions, or\ \ subscribe to all topics matching specified pattern to get dynamically\ \ assigned partitions" oneOf: - - title: "Manually assign a list of partitions" - required: - - "subscription_type" - - "topic_partitions" - properties: - subscription_type: - description: - "Manually assign a list of partitions to this consumer.\ - \ This interface does not allow for incremental assignment and will\ - \ replace the previous assignment (if there is one).\nIf the given\ - \ list of topic partitions is empty, it is treated the same as unsubscribe()." - type: "string" - const: "assign" - enum: - - "assign" - default: "assign" - topic_partitions: - title: "List of topic:partition pairs" - type: "string" - examples: - - "sample.topic:0, sample.topic:1" - - title: "Subscribe to all topics matching specified pattern" - required: - - "subscription_type" - - "topic_pattern" - properties: - subscription_type: - description: "Topic pattern from which the records will be read." - type: "string" - const: "subscribe" - enum: - - "subscribe" - default: "subscribe" - topic_pattern: - title: "Topic pattern" - type: "string" - examples: - - "sample.topic" + - title: "Manually assign a list of partitions" + required: + - "subscription_type" + - "topic_partitions" + properties: + subscription_type: + description: "Manually assign a list of partitions to this consumer.\ + \ This interface does not allow for incremental assignment and will\ + \ replace the previous assignment (if there is one).\nIf the given\ + \ list of topic partitions is empty, it is treated the same as unsubscribe()." + type: "string" + const: "assign" + enum: + - "assign" + default: "assign" + topic_partitions: + title: "List of topic:partition pairs" + type: "string" + examples: + - "sample.topic:0, sample.topic:1" + - title: "Subscribe to all topics matching specified pattern" + required: + - "subscription_type" + - "topic_pattern" + properties: + subscription_type: + description: "Topic pattern from which the records will be read." + type: "string" + const: "subscribe" + enum: + - "subscribe" + default: "subscribe" + topic_pattern: + title: "Topic pattern" + type: "string" + examples: + - "sample.topic" test_topic: title: "Test topic" description: "Topic to test if Airbyte can consume messages." type: "string" examples: - - "test.topic" + - "test.topic" group_id: title: "Group ID" description: "Group id." type: "string" examples: - - "group.id" + - "group.id" max_poll_records: title: "Max poll records" - description: - "The maximum number of records returned in a single call to\ + description: "The maximum number of records returned in a single call to\ \ poll(). Note, that max_poll_records does not impact the underlying fetching\ \ behavior. The consumer will cache the records from each fetch request\ \ and returns them incrementally from each poll." @@ -2938,101 +2811,93 @@ type: "object" description: "Protocol used to communicate with brokers." oneOf: - - title: "PLAINTEXT" - required: - - "security_protocol" - properties: - security_protocol: - type: "string" - enum: - - "PLAINTEXT" - default: "PLAINTEXT" - - title: "SASL PLAINTEXT" - required: - - "security_protocol" - - "sasl_mechanism" - - "sasl_jaas_config" - properties: - security_protocol: - type: "string" - enum: - - "SASL_PLAINTEXT" - default: "SASL_PLAINTEXT" - sasl_mechanism: - title: "SASL mechanism" - description: - "SASL mechanism used for client connections. This may\ - \ be any mechanism for which a security provider is available." - type: "string" - default: "PLAIN" - enum: - - "PLAIN" - sasl_jaas_config: - title: "SASL JAAS config" - description: - "JAAS login context parameters for SASL connections in\ - \ the format used by JAAS configuration files." - type: "string" - default: "" - airbyte_secret: true - - title: "SASL SSL" - required: - - "security_protocol" - - "sasl_mechanism" - - "sasl_jaas_config" - properties: - security_protocol: - type: "string" - enum: - - "SASL_SSL" - default: "SASL_SSL" - sasl_mechanism: - title: "SASL mechanism" - description: - "SASL mechanism used for client connections. This may\ - \ be any mechanism for which a security provider is available." - type: "string" - default: "GSSAPI" - enum: - - "GSSAPI" - - "OAUTHBEARER" - - "SCRAM-SHA-256" - sasl_jaas_config: - title: "SASL JAAS config" - description: - "JAAS login context parameters for SASL connections in\ - \ the format used by JAAS configuration files." - type: "string" - default: "" - airbyte_secret: true + - title: "PLAINTEXT" + required: + - "security_protocol" + properties: + security_protocol: + type: "string" + enum: + - "PLAINTEXT" + default: "PLAINTEXT" + - title: "SASL PLAINTEXT" + required: + - "security_protocol" + - "sasl_mechanism" + - "sasl_jaas_config" + properties: + security_protocol: + type: "string" + enum: + - "SASL_PLAINTEXT" + default: "SASL_PLAINTEXT" + sasl_mechanism: + title: "SASL mechanism" + description: "SASL mechanism used for client connections. This may\ + \ be any mechanism for which a security provider is available." + type: "string" + default: "PLAIN" + enum: + - "PLAIN" + sasl_jaas_config: + title: "SASL JAAS config" + description: "JAAS login context parameters for SASL connections in\ + \ the format used by JAAS configuration files." + type: "string" + default: "" + airbyte_secret: true + - title: "SASL SSL" + required: + - "security_protocol" + - "sasl_mechanism" + - "sasl_jaas_config" + properties: + security_protocol: + type: "string" + enum: + - "SASL_SSL" + default: "SASL_SSL" + sasl_mechanism: + title: "SASL mechanism" + description: "SASL mechanism used for client connections. This may\ + \ be any mechanism for which a security provider is available." + type: "string" + default: "GSSAPI" + enum: + - "GSSAPI" + - "OAUTHBEARER" + - "SCRAM-SHA-256" + sasl_jaas_config: + title: "SASL JAAS config" + description: "JAAS login context parameters for SASL connections in\ + \ the format used by JAAS configuration files." + type: "string" + default: "" + airbyte_secret: true client_id: title: "Client ID" - description: - "An id string to pass to the server when making requests. The\ + description: "An id string to pass to the server when making requests. The\ \ purpose of this is to be able to track the source of requests beyond\ \ just ip/port by allowing a logical application name to be included in\ \ server-side request logging." type: "string" examples: - - "airbyte-consumer" + - "airbyte-consumer" enable_auto_commit: title: "Enable auto commit" - description: - "If true the consumer's offset will be periodically committed\ + description: "If true the consumer's offset will be periodically committed\ \ in the background." type: "boolean" default: true auto_commit_interval_ms: title: "Auto commit interval ms" - description: - "The frequency in milliseconds that the consumer offsets are\ + description: "The frequency in milliseconds that the consumer offsets are\ \ auto-committed to Kafka if enable.auto.commit is set to true." type: "integer" default: 5000 client_dns_lookup: title: "Client DNS lookup" - description: - "Controls how the client uses DNS lookups. If set to use_all_dns_ips,\ + description: "Controls how the client uses DNS lookups. If set to use_all_dns_ips,\ \ connect to each returned IP address in sequence until a successful connection\ \ is established. After a disconnection, the next IP is used. Once all\ \ IPs have been used once, the client resolves the IP(s) from the hostname\ @@ -3044,21 +2909,19 @@ type: "string" default: "use_all_dns_ips" enum: - - "default" - - "use_all_dns_ips" - - "resolve_canonical_bootstrap_servers_only" + - "default" + - "use_all_dns_ips" + - "resolve_canonical_bootstrap_servers_only" retry_backoff_ms: title: "Retry backoff ms" - description: - "The amount of time to wait before attempting to retry a failed\ + description: "The amount of time to wait before attempting to retry a failed\ \ request to a given topic partition. This avoids repeatedly sending requests\ \ in a tight loop under some failure scenarios." type: "integer" default: 100 request_timeout_ms: title: "Request timeout ms" - description: - "The configuration controls the maximum amount of time the\ + description: "The configuration controls the maximum amount of time the\ \ client will wait for the response of a request. If the response is not\ \ received before the timeout elapses the client will resend the request\ \ if necessary or fail the request if retries are exhausted." @@ -3066,15 +2929,13 @@ default: 30000 receive_buffer_bytes: title: "Receive buffer bytes" - description: - "The size of the TCP receive buffer (SO_RCVBUF) to use when\ + description: "The size of the TCP receive buffer (SO_RCVBUF) to use when\ \ reading data. If the value is -1, the OS default will be used." type: "integer" default: 32768 auto_offset_reset: title: "Auto offset reset" - description: - "What to do when there is no initial offset in Kafka or if\ + description: "What to do when there is no initial offset in Kafka or if\ \ the current offset does not exist any more on the server - earliest:\ \ automatically reset the offset to the earliest offset, latest: automatically\ \ reset the offset to the latest offset, none: throw exception to the\ @@ -3083,13 +2944,12 @@ type: "string" default: "latest" enum: - - "latest" - - "earliest" - - "none" + - "latest" + - "earliest" + - "none" repeated_calls: title: "Repeated calls" - description: - "The number of repeated calls to poll() if no messages were\ + description: "The number of repeated calls to poll() if no messages were\ \ received." type: "integer" default: 3 @@ -3098,7 +2958,7 @@ supportsDBT: false supported_destination_sync_modes: [] supported_source_sync_modes: - - "append" + - "append" - dockerImage: "airbyte/source-klaviyo:0.1.2" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/klaviyo" @@ -3109,28 +2969,26 @@ properties: api_key: title: "Api Key" - description: - "Klaviyo API Key. See our docs if you need help finding this key." airbyte_secret: true type: "string" start_date: title: "Start Date" - description: - "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ \ data before this date will not be replicated." pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2017-01-25T00:00:00Z" + - "2017-01-25T00:00:00Z" type: "string" required: - - "api_key" - - "start_date" + - "api_key" + - "start_date" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "append" + - "append" - dockerImage: "airbyte/source-lever-hiring:0.1.0" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/lever-hiring" @@ -3141,21 +2999,18 @@ properties: client_id: title: "Client Id" - description: - "The client application id as provided when registering the\ + description: "The client application id as provided when registering the\ \ application with Lever." type: "string" client_secret: title: "Client Secret" - description: - "The application secret as provided when registering the application\ + description: "The application secret as provided when registering the application\ \ with Lever." airbyte_secret: true type: "string" refresh_token: title: "Refresh Token" - description: - "The refresh token your application will need to submit to\ + description: "The refresh token your application will need to submit to\ \ get a new access token after it's expired." type: "string" environment: @@ -3163,23 +3018,22 @@ description: "Sandbox or Production environment." default: "Production" enum: - - "Sandbox" - - "Production" + - "Sandbox" + - "Production" type: "string" start_date: title: "Start Date" - description: - "UTC date and time in the format 2019-02-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2019-02-25T00:00:00Z. Any\ \ data before this date will not be replicated." pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2021-04-25T00:00:00Z" + - "2021-04-25T00:00:00Z" type: "string" required: - - "client_id" - - "client_secret" - - "refresh_token" - - "start_date" + - "client_id" + - "client_secret" + - "refresh_token" + - "start_date" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -3188,9 +3042,9 @@ oauth2Specification: rootObject: [] oauthFlowInitParameters: - - - "client_id" - - - "client_secret" - - - "refresh_token" + - - "client_id" + - - "client_secret" + - - "refresh_token" oauthFlowOutputParameters: [] - dockerImage: "airbyte/source-linkedin-ads:0.1.2" spec: @@ -3200,19 +3054,18 @@ title: "Linkedin Ads Spec" type: "object" required: - - "start_date" - - "access_token" + - "start_date" + - "access_token" additionalProperties: false properties: start_date: type: "string" title: "Start Date" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" - description: - "Date in the format 2020-09-17. Any data before this date will\ + description: "Date in the format 2020-09-17. Any data before this date will\ \ not be replicated." examples: - - "2021-05-17" + - "2021-05-17" access_token: type: "string" title: "Access Token" @@ -3221,8 +3074,7 @@ account_ids: title: "Account IDs" type: "array" - description: - "Specify the Account IDs separated by space, from which to\ + description: "Specify the Account IDs separated by space, from which to\ \ pull the data. Leave empty to pull from all associated accounts." items: type: "integer" @@ -3238,25 +3090,23 @@ title: "Looker Spec" type: "object" required: - - "domain" - - "client_id" - - "client_secret" + - "domain" + - "client_id" + - "client_secret" additionalProperties: false properties: domain: type: "string" examples: - - "domainname.looker.com" - - "looker.clientname.com" - - "123.123.124.123:8000" - description: - "Domain for your Looker account, e.g. airbyte.cloud.looker.com,looker.[clientname].com,IP\ + - "domainname.looker.com" + - "looker.clientname.com" + - "123.123.124.123:8000" + description: "Domain for your Looker account, e.g. airbyte.cloud.looker.com,looker.[clientname].com,IP\ \ address" client_id: title: "Client ID" type: "string" - description: - "The Client ID is first part of an API3 key that is specific\ + description: "The Client ID is first part of an API3 key that is specific\ \ to each Looker user. See the docs for more information on how to generate this key." client_secret: @@ -3281,8 +3131,8 @@ title: "Mailchimp Spec" type: "object" required: - - "username" - - "apikey" + - "username" + - "apikey" additionalProperties: false properties: username: @@ -3291,8 +3141,7 @@ apikey: type: "string" airbyte_secret: true - description: - "API Key. See the docs for information on how to generate this key." supportsNormalization: false supportsDBT: false @@ -3305,51 +3154,46 @@ title: "Source Marketo Spec" type: "object" required: - - "domain_url" - - "client_id" - - "client_secret" - - "start_date" + - "domain_url" + - "client_id" + - "client_secret" + - "start_date" additionalProperties: false properties: domain_url: type: "string" - description: - "Your Marketo Base URL. See the docs for info on how to obtain this." examples: - - "https://000-AAA-000.mktorest.com" + - "https://000-AAA-000.mktorest.com" airbyte_secret: true client_id: type: "string" - description: - "Your Marketo client_id. See the docs for info on how to obtain this." airbyte_secret: true client_secret: type: "string" - description: - "Your Marketo client secret. See the docs for info on how to obtain this." airbyte_secret: true start_date: type: "string" - description: - "Data generated in Marketo after this date will be replicated.\ + description: "Data generated in Marketo after this date will be replicated.\ \ This date must be specified in the format YYYY-MM-DDT00:00:00Z." examples: - - "2020-09-25T00:00:00Z" + - "2020-09-25T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" window_in_days: type: "integer" - description: - "The amount of days for each data-chunk begining from start_date.\ + description: "The amount of days for each data-chunk begining from start_date.\ \ (Min=1, as for a Day; Max=30, as for a Month)." examples: - - 1 - - 5 - - 10 - - 15 - - 30 + - 1 + - 5 + - 10 + - 15 + - 30 default: 30 supportsNormalization: false supportsDBT: false @@ -3362,10 +3206,10 @@ title: "MSSQL Source Spec" type: "object" required: - - "host" - - "port" - - "database" - - "username" + - "host" + - "port" + - "database" + - "username" additionalProperties: false properties: host: @@ -3377,12 +3221,12 @@ minimum: 0 maximum: 65536 examples: - - "1433" + - "1433" database: description: "Name of the database." type: "string" examples: - - "master" + - "master" username: description: "Username to use to access the database." type: "string" @@ -3396,175 +3240,164 @@ description: "Encryption method to use when communicating with the database" order: 6 oneOf: - - title: "Unencrypted" - additionalProperties: false - description: "Data transfer will not be encrypted." - required: - - "ssl_method" - properties: - ssl_method: - type: "string" - const: "unencrypted" - enum: - - "unencrypted" - default: "unencrypted" - - title: "Encrypted (trust server certificate)" - additionalProperties: false - description: - "Use the cert provided by the server without verification.\ - \ (For testing purposes only!)" - required: - - "ssl_method" - properties: - ssl_method: - type: "string" - const: "encrypted_trust_server_certificate" - enum: - - "encrypted_trust_server_certificate" - default: "encrypted_trust_server_certificate" - - title: "Encrypted (verify certificate)" - additionalProperties: false - description: "Verify and use the cert provided by the server." - required: - - "ssl_method" - - "trustStoreName" - - "trustStorePassword" - properties: - ssl_method: - type: "string" - const: "encrypted_verify_certificate" - enum: - - "encrypted_verify_certificate" - default: "encrypted_verify_certificate" - hostNameInCertificate: - title: "Host Name In Certificate" - type: "string" - description: - "Specifies the host name of the server. The value of\ - \ this property must match the subject property of the certificate." - order: 7 + - title: "Unencrypted" + additionalProperties: false + description: "Data transfer will not be encrypted." + required: + - "ssl_method" + properties: + ssl_method: + type: "string" + const: "unencrypted" + enum: + - "unencrypted" + default: "unencrypted" + - title: "Encrypted (trust server certificate)" + additionalProperties: false + description: "Use the cert provided by the server without verification.\ + \ (For testing purposes only!)" + required: + - "ssl_method" + properties: + ssl_method: + type: "string" + const: "encrypted_trust_server_certificate" + enum: + - "encrypted_trust_server_certificate" + default: "encrypted_trust_server_certificate" + - title: "Encrypted (verify certificate)" + additionalProperties: false + description: "Verify and use the cert provided by the server." + required: + - "ssl_method" + - "trustStoreName" + - "trustStorePassword" + properties: + ssl_method: + type: "string" + const: "encrypted_verify_certificate" + enum: + - "encrypted_verify_certificate" + default: "encrypted_verify_certificate" + hostNameInCertificate: + title: "Host Name In Certificate" + type: "string" + description: "Specifies the host name of the server. The value of\ + \ this property must match the subject property of the certificate." + order: 7 replication_method: type: "string" title: "Replication Method" - description: - "Replication method to use for extracting data from the database.\ + description: "Replication method to use for extracting data from the database.\ \ STANDARD replication requires no setup on the DB side but will not be\ \ able to represent deletions incrementally. CDC uses {TBC} to detect\ \ inserts, updates, and deletes. This needs to be configured on the source\ \ database itself." default: "STANDARD" enum: - - "STANDARD" - - "CDC" + - "STANDARD" + - "CDC" tunnel_method: type: "object" title: "SSH Tunnel Method" - description: - "Whether to initiate an SSH tunnel before connecting to the\ + description: "Whether to initiate an SSH tunnel before connecting to the\ \ database, and if so, which kind of authentication to use." oneOf: - - title: "No Tunnel" - required: - - "tunnel_method" - properties: - tunnel_method: - description: "No ssh tunnel needed to connect to database" - type: "string" - const: "NO_TUNNEL" - order: 0 - - title: "SSH Key Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "ssh_key" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and ssh key" - type: "string" - const: "SSH_KEY_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host." - type: "string" - order: 3 - ssh_key: - title: "SSH Private Key" - description: - "OS-level user account ssh key credentials for logging\ - \ into the jump server host." - type: "string" - airbyte_secret: true - multiline: true - order: 4 - - title: "Password Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "tunnel_user_password" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and password authentication" - type: "string" - const: "SSH_PASSWORD_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host" - type: "string" - order: 3 - tunnel_user_password: - title: "Password" - description: "OS-level password for logging into the jump server host" - type: "string" - airbyte_secret: true - order: 4 + - title: "No Tunnel" + required: + - "tunnel_method" + properties: + tunnel_method: + description: "No ssh tunnel needed to connect to database" + type: "string" + const: "NO_TUNNEL" + order: 0 + - title: "SSH Key Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "ssh_key" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and ssh key" + type: "string" + const: "SSH_KEY_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host." + type: "string" + order: 3 + ssh_key: + title: "SSH Private Key" + description: "OS-level user account ssh key credentials for logging\ + \ into the jump server host." + type: "string" + airbyte_secret: true + multiline: true + order: 4 + - title: "Password Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "tunnel_user_password" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and password authentication" + type: "string" + const: "SSH_PASSWORD_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host" + type: "string" + order: 3 + tunnel_user_password: + title: "Password" + description: "OS-level password for logging into the jump server host" + type: "string" + airbyte_secret: true + order: 4 supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -3576,10 +3409,10 @@ title: "Microsoft Teams Spec" type: "object" required: - - "tenant_id" - - "client_id" - - "client_secret" - - "period" + - "tenant_id" + - "client_id" + - "client_secret" + - "period" additionalProperties: false properties: tenant_id: @@ -3597,11 +3430,10 @@ airbyte_secret: true period: type: "string" - description: - "Specifies the length of time over which the Team Device Report\ + description: "Specifies the length of time over which the Team Device Report\ \ stream is aggregated. The supported values are: D7, D30, D90, and D180." examples: - - "D7" + - "D7" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -3613,60 +3445,54 @@ title: "Source Mixpanel Spec" type: "object" required: - - "api_secret" + - "api_secret" additionalProperties: true properties: api_secret: type: "string" - description: - "Mixpanel API Secret. See the docs for more information on how to obtain this key." airbyte_secret: true attribution_window: type: "integer" - description: - "Latency minimum number of days to look-back to account for\ + description: "Latency minimum number of days to look-back to account for\ \ delays in attributing accurate results. Default attribution window is\ \ 5 days." default: 5 date_window_size: type: "integer" - description: - "Number of days for date window looping through transactional\ + description: "Number of days for date window looping through transactional\ \ endpoints with from_date and to_date. Default date_window_size is 30\ \ days. Clients with large volumes of events may want to decrease this\ \ to 14, 7, or even down to 1-2 days." default: 30 project_timezone: type: "string" - description: - "Time zone in which integer date times are stored. The project\ + description: "Time zone in which integer date times are stored. The project\ \ timezone may be found in the project settings in the Mixpanel console." default: "US/Pacific" examples: - - "US/Pacific" - - "UTC" + - "US/Pacific" + - "UTC" select_properties_by_default: type: "boolean" - description: - "Setting this config parameter to true ensures that new properties\ + description: "Setting this config parameter to true ensures that new properties\ \ on events and engage records are captured. Otherwise new properties\ \ will be ignored" default: true start_date: type: "string" - description: - "The default value to use if no bookmark exists for an endpoint.\ + description: "The default value to use if no bookmark exists for an endpoint.\ \ If this option is not set, the connector will replicate data from up\ \ to one year ago by default." examples: - - "2021-11-16" + - "2021-11-16" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}(T[0-9]{2}:[0-9]{2}:[0-9]{2}Z)?$" region: type: "string" enum: - - "US" - - "EU" + - "US" + - "EU" default: "US" supportsNormalization: false supportsDBT: false @@ -3679,13 +3505,12 @@ title: "Monday Spec" type: "object" required: - - "api_token" + - "api_token" additionalProperties: false properties: api_token: type: "string" - description: - "This is the API token to authenticate requests to Monday.\ + description: "This is the API token to authenticate requests to Monday.\ \ Profile picture (bottom left) => Admin => API" airbyte_secret: true supportsNormalization: false @@ -3700,93 +3525,90 @@ title: "MongoDb Source Spec" type: "object" required: - - "database" + - "database" additionalProperties: true properties: instance_type: type: "object" title: "MongoDb instance type" - description: - "MongoDb instance to connect to. For MongoDB Atlas and Replica\ + description: "MongoDb instance to connect to. For MongoDB Atlas and Replica\ \ Set TLS connection is used by default." order: 0 oneOf: - - title: "Standalone MongoDb Instance" - required: - - "instance" - - "host" - - "port" - properties: - instance: - type: "string" - enum: - - "standalone" - default: "standalone" - host: - title: "Host" - type: "string" - description: "Host of a Mongo database to be replicated." - order: 0 - port: - title: "Port" - type: "integer" - description: "Port of a Mongo database to be replicated." - minimum: 0 - maximum: 65536 - default: 27017 - examples: - - "27017" - order: 1 - tls: - title: "TLS connection" - type: "boolean" - description: - "Indicates whether TLS encryption protocol will be used\ - \ to connect to MongoDB. It is recommended to use TLS connection\ - \ if possible. For more information see documentation." - default: false - order: 2 - - title: "Replica Set" - required: - - "instance" - - "server_addresses" - properties: - instance: - type: "string" - enum: - - "replica" - default: "replica" - server_addresses: - title: "Server addresses" - type: "string" - description: - "The members of a replica set. Please specify `host`:`port`\ - \ of each member seperated by comma." - examples: - - "host1:27017,host2:27017,host3:27017" - order: 0 - replica_set: - title: "Replica Set" - type: "string" - description: "A replica set name." - order: 1 - - title: "MongoDB Atlas" - additionalProperties: false - required: - - "instance" - - "cluster_url" - properties: - instance: - type: "string" - enum: - - "atlas" - default: "atlas" - cluster_url: - title: "Cluster URL" - type: "string" - description: "URL of a cluster to connect to." - order: 0 + - title: "Standalone MongoDb Instance" + required: + - "instance" + - "host" + - "port" + properties: + instance: + type: "string" + enum: + - "standalone" + default: "standalone" + host: + title: "Host" + type: "string" + description: "Host of a Mongo database to be replicated." + order: 0 + port: + title: "Port" + type: "integer" + description: "Port of a Mongo database to be replicated." + minimum: 0 + maximum: 65536 + default: 27017 + examples: + - "27017" + order: 1 + tls: + title: "TLS connection" + type: "boolean" + description: "Indicates whether TLS encryption protocol will be used\ + \ to connect to MongoDB. It is recommended to use TLS connection\ + \ if possible. For more information see documentation." + default: false + order: 2 + - title: "Replica Set" + required: + - "instance" + - "server_addresses" + properties: + instance: + type: "string" + enum: + - "replica" + default: "replica" + server_addresses: + title: "Server addresses" + type: "string" + description: "The members of a replica set. Please specify `host`:`port`\ + \ of each member seperated by comma." + examples: + - "host1:27017,host2:27017,host3:27017" + order: 0 + replica_set: + title: "Replica Set" + type: "string" + description: "A replica set name." + order: 1 + - title: "MongoDB Atlas" + additionalProperties: false + required: + - "instance" + - "cluster_url" + properties: + instance: + type: "string" + enum: + - "atlas" + default: "atlas" + cluster_url: + title: "Cluster URL" + type: "string" + description: "URL of a cluster to connect to." + order: 0 database: title: "Database name" type: "string" @@ -3809,7 +3631,7 @@ description: "Authentication source where user information is stored" default: "admin" examples: - - "admin" + - "admin" order: 4 supportsNormalization: false supportsDBT: false @@ -3822,11 +3644,11 @@ title: "MySql Source Spec" type: "object" required: - - "host" - - "port" - - "database" - - "username" - - "replication_method" + - "host" + - "port" + - "database" + - "username" + - "replication_method" additionalProperties: false properties: host: @@ -3840,7 +3662,7 @@ maximum: 65536 default: 3306 examples: - - "3306" + - "3306" order: 1 database: description: "Name of the database." @@ -3856,8 +3678,7 @@ airbyte_secret: true order: 4 jdbc_url_params: - description: - "Additional properties to pass to the jdbc url string when\ + description: "Additional properties to pass to the jdbc url string when\ \ connecting to the database formatted as 'key=value' pairs separated\ \ by the symbol '&'. (example: key1=value1&key2=value2&key3=value3)" type: "string" @@ -3871,8 +3692,7 @@ replication_method: type: "string" title: "Replication Method" - description: - "Replication method to use for extracting data from the database.\ + description: "Replication method to use for extracting data from the database.\ \ STANDARD replication requires no setup on the DB side but will not be\ \ able to represent deletions incrementally. CDC uses the Binlog to detect\ \ inserts, updates, and deletes. This needs to be configured on the source\ @@ -3880,117 +3700,109 @@ order: 7 default: "STANDARD" enum: - - "STANDARD" - - "CDC" + - "STANDARD" + - "CDC" tunnel_method: type: "object" title: "SSH Tunnel Method" - description: - "Whether to initiate an SSH tunnel before connecting to the\ + description: "Whether to initiate an SSH tunnel before connecting to the\ \ database, and if so, which kind of authentication to use." oneOf: - - title: "No Tunnel" - required: - - "tunnel_method" - properties: - tunnel_method: - description: "No ssh tunnel needed to connect to database" - type: "string" - const: "NO_TUNNEL" - order: 0 - - title: "SSH Key Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "ssh_key" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and ssh key" - type: "string" - const: "SSH_KEY_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host." - type: "string" - order: 3 - ssh_key: - title: "SSH Private Key" - description: - "OS-level user account ssh key credentials in RSA PEM\ - \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" - type: "string" - airbyte_secret: true - multiline: true - order: 4 - - title: "Password Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "tunnel_user_password" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and password authentication" - type: "string" - const: "SSH_PASSWORD_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host" - type: "string" - order: 3 - tunnel_user_password: - title: "Password" - description: "OS-level password for logging into the jump server host" - type: "string" - airbyte_secret: true - order: 4 + - title: "No Tunnel" + required: + - "tunnel_method" + properties: + tunnel_method: + description: "No ssh tunnel needed to connect to database" + type: "string" + const: "NO_TUNNEL" + order: 0 + - title: "SSH Key Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "ssh_key" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and ssh key" + type: "string" + const: "SSH_KEY_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host." + type: "string" + order: 3 + ssh_key: + title: "SSH Private Key" + description: "OS-level user account ssh key credentials in RSA PEM\ + \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" + type: "string" + airbyte_secret: true + multiline: true + order: 4 + - title: "Password Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "tunnel_user_password" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and password authentication" + type: "string" + const: "SSH_PASSWORD_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host" + type: "string" + order: 3 + tunnel_user_password: + title: "Password" + description: "OS-level password for logging into the jump server host" + type: "string" + airbyte_secret: true + order: 4 supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -4002,15 +3814,14 @@ title: "Okta Spec" type: "object" required: - - "token" - - "base_url" + - "token" + - "base_url" additionalProperties: false properties: token: type: "string" title: "API Token" - description: - "A Okta token. See the docs for instructions on how to generate it." airbyte_secret: true base_url: @@ -4029,34 +3840,31 @@ title: "OneSignal Source Spec" type: "object" required: - - "user_auth_key" - - "start_date" - - "outcome_names" + - "user_auth_key" + - "start_date" + - "outcome_names" additionalProperties: false properties: user_auth_key: type: "string" - description: - "OneSignal User Auth Key, see the docs for more information on how to obtain this key." airbyte_secret: true start_date: type: "string" - description: - "The date from which you'd like to replicate data for OneSignal\ + description: "The date from which you'd like to replicate data for OneSignal\ \ API, in the format YYYY-MM-DDT00:00:00Z. All data generated after this\ \ date will be replicated." examples: - - "2020-11-16T00:00:00Z" + - "2020-11-16T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" outcome_names: type: "string" - description: - "Comma-separated list of names and the value (sum/count) for\ + description: "Comma-separated list of names and the value (sum/count) for\ \ the returned outcome data. See the docs for more details" examples: - - "os__session_duration.count,os__click.count,CustomOutcomeName.sum" + - "os__session_duration.count,os__click.count,CustomOutcomeName.sum" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -4068,10 +3876,10 @@ title: "Oracle Source Spec" type: "object" required: - - "host" - - "port" - - "sid" - - "username" + - "host" + - "port" + - "sid" + - "username" additionalProperties: false properties: host: @@ -4080,8 +3888,7 @@ type: "string" port: title: "Port" - description: - "Port of the database.\nOracle Corporations recommends the\ + description: "Port of the database.\nOracle Corporations recommends the\ \ following port numbers:\n1521 - Default listening port for client connections\ \ to the listener. \n2484 - Recommended and officially registered listening\ \ port for client connections to the listener using TCP/IP with SSL" @@ -4115,175 +3922,164 @@ description: "Encryption method to use when communicating with the database" order: 6 oneOf: - - title: "Unencrypted" - additionalProperties: false - description: "Data transfer will not be encrypted." - required: - - "encryption_method" - properties: - encryption_method: - type: "string" - const: "unencrypted" - enum: - - "unencrypted" - default: "unencrypted" - - title: "Native Network Ecryption (NNE)" - additionalProperties: false - description: - "Native network encryption gives you the ability to encrypt\ - \ database connections, without the configuration overhead of TCP/IP\ - \ and SSL/TLS and without the need to open and listen on different ports." - required: - - "encryption_method" - properties: - encryption_method: - type: "string" - const: "client_nne" - enum: - - "client_nne" - default: "client_nne" - encryption_algorithm: - type: "string" - description: - "This parameter defines the encryption algorithm to be\ - \ used" - title: "Encryption Algorithm" - default: "AES256" - enum: - - "AES256" - - "RC4_56" - - "3DES168" - - title: "TLS Encrypted (verify certificate)" - additionalProperties: false - description: "Verify and use the cert provided by the server." - required: - - "encryption_method" - - "ssl_certificate" - properties: - encryption_method: - type: "string" - const: "encrypted_verify_certificate" - enum: - - "encrypted_verify_certificate" - default: "encrypted_verify_certificate" - ssl_certificate: - title: "SSL PEM file" - description: - "Privacy Enhanced Mail (PEM) files are concatenated certificate\ - \ containers frequently used in certificate installations" - type: "string" - airbyte_secret: true - multiline: true - order: 4 + - title: "Unencrypted" + additionalProperties: false + description: "Data transfer will not be encrypted." + required: + - "encryption_method" + properties: + encryption_method: + type: "string" + const: "unencrypted" + enum: + - "unencrypted" + default: "unencrypted" + - title: "Native Network Ecryption (NNE)" + additionalProperties: false + description: "Native network encryption gives you the ability to encrypt\ + \ database connections, without the configuration overhead of TCP/IP\ + \ and SSL/TLS and without the need to open and listen on different ports." + required: + - "encryption_method" + properties: + encryption_method: + type: "string" + const: "client_nne" + enum: + - "client_nne" + default: "client_nne" + encryption_algorithm: + type: "string" + description: "This parameter defines the encryption algorithm to be\ + \ used" + title: "Encryption Algorithm" + default: "AES256" + enum: + - "AES256" + - "RC4_56" + - "3DES168" + - title: "TLS Encrypted (verify certificate)" + additionalProperties: false + description: "Verify and use the cert provided by the server." + required: + - "encryption_method" + - "ssl_certificate" + properties: + encryption_method: + type: "string" + const: "encrypted_verify_certificate" + enum: + - "encrypted_verify_certificate" + default: "encrypted_verify_certificate" + ssl_certificate: + title: "SSL PEM file" + description: "Privacy Enhanced Mail (PEM) files are concatenated certificate\ + \ containers frequently used in certificate installations" + type: "string" + airbyte_secret: true + multiline: true + order: 4 tunnel_method: type: "object" title: "SSH Tunnel Method" - description: - "Whether to initiate an SSH tunnel before connecting to the\ + description: "Whether to initiate an SSH tunnel before connecting to the\ \ database, and if so, which kind of authentication to use." oneOf: - - title: "No Tunnel" - required: - - "tunnel_method" - properties: - tunnel_method: - description: "No ssh tunnel needed to connect to database" - type: "string" - const: "NO_TUNNEL" - order: 0 - - title: "SSH Key Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "ssh_key" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and ssh key" - type: "string" - const: "SSH_KEY_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host." - type: "string" - order: 3 - ssh_key: - title: "SSH Private Key" - description: - "OS-level user account ssh key credentials in RSA PEM\ - \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" - type: "string" - airbyte_secret: true - multiline: true - order: 4 - - title: "Password Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "tunnel_user_password" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and password authentication" - type: "string" - const: "SSH_PASSWORD_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host" - type: "string" - order: 3 - tunnel_user_password: - title: "Password" - description: "OS-level password for logging into the jump server host" - type: "string" - airbyte_secret: true - order: 4 + - title: "No Tunnel" + required: + - "tunnel_method" + properties: + tunnel_method: + description: "No ssh tunnel needed to connect to database" + type: "string" + const: "NO_TUNNEL" + order: 0 + - title: "SSH Key Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "ssh_key" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and ssh key" + type: "string" + const: "SSH_KEY_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host." + type: "string" + order: 3 + ssh_key: + title: "SSH Private Key" + description: "OS-level user account ssh key credentials in RSA PEM\ + \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" + type: "string" + airbyte_secret: true + multiline: true + order: 4 + - title: "Password Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "tunnel_user_password" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and password authentication" + type: "string" + const: "SSH_PASSWORD_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host" + type: "string" + order: 3 + tunnel_user_password: + title: "Password" + description: "OS-level password for logging into the jump server host" + type: "string" + airbyte_secret: true + order: 4 supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -4295,10 +4091,10 @@ title: "Paypal Transaction Search" type: "object" required: - - "client_id" - - "secret" - - "start_date" - - "is_sandbox" + - "client_id" + - "secret" + - "start_date" + - "is_sandbox" additionalProperties: true properties: client_id: @@ -4313,17 +4109,15 @@ start_date: type: "string" title: "Start Date" - description: - "Start Date for data extraction in ISO format. Date must be in range from 3 years till 12 hrs before\ \ present time" examples: - - "2021-06-11T23:59:59-00:00" + - "2021-06-11T23:59:59-00:00" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}[+-][0-9]{2}:[0-9]{2}$" is_sandbox: title: "Is Sandbox" - description: - "Whether or not to Sandbox or Production environment to extract\ + description: "Whether or not to Sandbox or Production environment to extract\ \ data from" type: "boolean" default: false @@ -4338,32 +4132,29 @@ title: "Paystack Source Spec" type: "object" required: - - "secret_key" - - "start_date" + - "secret_key" + - "start_date" additionalProperties: false properties: secret_key: type: "string" pattern: "^(s|r)k_(live|test)_[a-zA-Z0-9]+$" - description: - "Paystack API key (usually starts with 'sk_live_'; find yours\ + description: "Paystack API key (usually starts with 'sk_live_'; find yours\ \ here)." airbyte_secret: true start_date: type: "string" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" - description: - "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ \ data before this date will not be replicated." examples: - - "2017-01-25T00:00:00Z" + - "2017-01-25T00:00:00Z" lookback_window_days: type: "integer" title: "Lookback Window (in days)" default: 0 minimum: 0 - description: - "When set, the connector will always reload data from the past\ + description: "When set, the connector will always reload data from the past\ \ N days, where N is the value set here. This is useful if your data is\ \ updated after creation." supportsNormalization: false @@ -4377,47 +4168,42 @@ title: "Pinterest Spec" type: "object" required: - - "client_id" - - "client_secret" - - "refresh_token" + - "client_id" + - "client_secret" + - "refresh_token" additionalProperties: true properties: client_id: type: "string" title: "Client id" - description: - "Your Pinterest client id. See the docs for instructions on how to generate it." airbyte_secret: true client_secret: type: "string" title: "Client secret" - description: - "Your Pinterest client secret. See the docs for instructions on how to generate it." airbyte_secret: true refresh_token: type: "string" title: "Refresh token" - description: - "Your Pinterest refresh token. See the docs for instructions on how to generate it." airbyte_secret: true access_token: type: "string" title: "Access token" - description: - "Your Pinterest access token. See the docs for instructions on how to generate it." airbyte_secret: true start_date: type: "string" title: "Start date" - description: - "A date in the format YYYY-MM-DD. If you have not set a date,\ + description: "A date in the format YYYY-MM-DD. If you have not set a date,\ \ it will be 2020-07-28 by default." examples: - - "2020-07-28" + - "2020-07-28" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -4429,85 +4215,82 @@ title: "Pipedrive Spec" type: "object" required: - - "replication_start_date" + - "replication_start_date" additionalProperties: true properties: authorization: type: "object" title: "Authentication Type" oneOf: - - title: "Sign in via Pipedrive (OAuth)" - type: "object" - required: - - "auth_type" - - "client_id" - - "client_secret" - - "refresh_token" - properties: - auth_type: - type: "string" - const: "Client" - enum: - - "Client" - default: "Client" - order: 0 - client_id: - title: "Client ID" - type: "string" - description: "The Client ID of your developer application" - airbyte_secret: true - client_secret: - title: "Client Secret" - type: "string" - description: "The client secret of your developer application" - airbyte_secret: true - access_token: - title: "Access Token" - type: "string" - description: - "An access token generated using the above client ID\ - \ and secret" - airbyte_secret: true - refresh_token: - title: "Refresh Token" - type: "string" - description: - "A refresh token generated using the above client ID\ - \ and secret" - airbyte_secret: true - - type: "object" - title: "API Key Authentication" - required: - - "auth_type" - - "api_token" - properties: - auth_type: - type: "string" - const: "Token" - enum: - - "Token" - default: "Token" - order: 0 - api_token: - title: "API Token" - type: "string" - description: "Pipedrive API Token" - airbyte_secret: true + - title: "Sign in via Pipedrive (OAuth)" + type: "object" + required: + - "auth_type" + - "client_id" + - "client_secret" + - "refresh_token" + properties: + auth_type: + type: "string" + const: "Client" + enum: + - "Client" + default: "Client" + order: 0 + client_id: + title: "Client ID" + type: "string" + description: "The Client ID of your developer application" + airbyte_secret: true + client_secret: + title: "Client Secret" + type: "string" + description: "The client secret of your developer application" + airbyte_secret: true + access_token: + title: "Access Token" + type: "string" + description: "An access token generated using the above client ID\ + \ and secret" + airbyte_secret: true + refresh_token: + title: "Refresh Token" + type: "string" + description: "A refresh token generated using the above client ID\ + \ and secret" + airbyte_secret: true + - type: "object" + title: "API Key Authentication" + required: + - "auth_type" + - "api_token" + properties: + auth_type: + type: "string" + const: "Token" + enum: + - "Token" + default: "Token" + order: 0 + api_token: + title: "API Token" + type: "string" + description: "Pipedrive API Token" + airbyte_secret: true replication_start_date: title: "Replication Start Date" - description: - "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ \ data before this date will not be replicated. When specified and not\ \ None, then stream will behave as incremental" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2017-01-25T00:00:00Z" + - "2017-01-25T00:00:00Z" type: "string" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "append" + - "append" - dockerImage: "airbyte/source-plaid:0.2.1" spec: documentationUrl: "https://plaid.com/docs/api/" @@ -4515,9 +4298,9 @@ $schema: "http://json-schema.org/draft-07/schema#" type: "object" required: - - "access_token" - - "api_key" - - "client_id" + - "access_token" + - "api_key" + - "client_id" additionalProperties: false properties: access_token: @@ -4537,9 +4320,9 @@ title: "Plaid Environment" type: "string" enum: - - "sandbox" - - "development" - - "production" + - "sandbox" + - "development" + - "production" description: "The Plaid environment" supportsNormalization: false supportsDBT: false @@ -4552,7 +4335,7 @@ title: "Pokeapi Spec" type: "object" required: - - "pokemon_name" + - "pokemon_name" additionalProperties: false properties: pokemon_name: @@ -4560,7 +4343,7 @@ description: "Pokemon requested from the API." pattern: "^[a-z0-9_\\-]+$" examples: - - "ditto, luxray, snorlax" + - "ditto, luxray, snorlax" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -4572,8 +4355,8 @@ title: "PostHog Spec" type: "object" required: - - "api_key" - - "start_date" + - "api_key" + - "start_date" additionalProperties: false properties: start_date: @@ -4582,19 +4365,18 @@ description: "The date from which you'd like to replicate the data" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2021-01-01T00:00:00Z" + - "2021-01-01T00:00:00Z" api_key: type: "string" airbyte_secret: true - description: - "API Key. See the docs for information on how to generate this key." base_url: type: "string" default: "https://app.posthog.com" description: "Base PostHog url. Defaults to PostHog Cloud (https://app.posthog.com)." examples: - - "https://posthog.example.com" + - "https://posthog.example.com" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -4606,10 +4388,10 @@ title: "Postgres Source Spec" type: "object" required: - - "host" - - "port" - - "database" - - "username" + - "host" + - "port" + - "database" + - "username" additionalProperties: false properties: host: @@ -4625,7 +4407,7 @@ maximum: 65536 default: 5432 examples: - - "5432" + - "5432" order: 1 database: title: "DB Name" @@ -4655,172 +4437,161 @@ description: "Replication method to use for extracting data from the database." order: 6 oneOf: - - title: "Standard" - additionalProperties: false - description: - "Standard replication requires no setup on the DB side but\ - \ will not be able to represent deletions incrementally." - required: - - "method" - properties: - method: - type: "string" - const: "Standard" - enum: - - "Standard" - default: "Standard" - order: 0 - - title: "Logical Replication (CDC)" - additionalProperties: false - description: - "Logical replication uses the Postgres write-ahead log (WAL)\ - \ to detect inserts, updates, and deletes. This needs to be configured\ - \ on the source database itself. Only available on Postgres 10 and above.\ - \ Read the Postgres Source docs for more information." - required: - - "method" - - "replication_slot" - - "publication" - properties: - method: - type: "string" - const: "CDC" - enum: - - "CDC" - default: "CDC" - order: 0 - plugin: - type: "string" - description: - "A logical decoding plug-in installed on the PostgreSQL\ - \ server. `pgoutput` plug-in is used by default.\nIf replication\ - \ table contains a lot of big jsonb values it is recommended to\ - \ use `wal2json` plug-in. For more information about `wal2json`\ - \ plug-in read Postgres Source docs." - enum: - - "pgoutput" - - "wal2json" - default: "pgoutput" - order: 1 - replication_slot: - type: "string" - description: "A plug-in logical replication slot." - order: 2 - publication: - type: "string" - description: "A Postgres publication used for consuming changes." - order: 3 + - title: "Standard" + additionalProperties: false + description: "Standard replication requires no setup on the DB side but\ + \ will not be able to represent deletions incrementally." + required: + - "method" + properties: + method: + type: "string" + const: "Standard" + enum: + - "Standard" + default: "Standard" + order: 0 + - title: "Logical Replication (CDC)" + additionalProperties: false + description: "Logical replication uses the Postgres write-ahead log (WAL)\ + \ to detect inserts, updates, and deletes. This needs to be configured\ + \ on the source database itself. Only available on Postgres 10 and above.\ + \ Read the Postgres Source docs for more information." + required: + - "method" + - "replication_slot" + - "publication" + properties: + method: + type: "string" + const: "CDC" + enum: + - "CDC" + default: "CDC" + order: 0 + plugin: + type: "string" + description: "A logical decoding plug-in installed on the PostgreSQL\ + \ server. `pgoutput` plug-in is used by default.\nIf replication\ + \ table contains a lot of big jsonb values it is recommended to\ + \ use `wal2json` plug-in. For more information about `wal2json`\ + \ plug-in read Postgres Source docs." + enum: + - "pgoutput" + - "wal2json" + default: "pgoutput" + order: 1 + replication_slot: + type: "string" + description: "A plug-in logical replication slot." + order: 2 + publication: + type: "string" + description: "A Postgres publication used for consuming changes." + order: 3 tunnel_method: type: "object" title: "SSH Tunnel Method" - description: - "Whether to initiate an SSH tunnel before connecting to the\ + description: "Whether to initiate an SSH tunnel before connecting to the\ \ database, and if so, which kind of authentication to use." oneOf: - - title: "No Tunnel" - required: - - "tunnel_method" - properties: - tunnel_method: - description: "No ssh tunnel needed to connect to database" - type: "string" - const: "NO_TUNNEL" - order: 0 - - title: "SSH Key Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "ssh_key" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and ssh key" - type: "string" - const: "SSH_KEY_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host." - type: "string" - order: 3 - ssh_key: - title: "SSH Private Key" - description: - "OS-level user account ssh key credentials in RSA PEM\ - \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" - type: "string" - airbyte_secret: true - multiline: true - order: 4 - - title: "Password Authentication" - required: - - "tunnel_method" - - "tunnel_host" - - "tunnel_port" - - "tunnel_user" - - "tunnel_user_password" - properties: - tunnel_method: - description: - "Connect through a jump server tunnel host using username\ - \ and password authentication" - type: "string" - const: "SSH_PASSWORD_AUTH" - order: 0 - tunnel_host: - title: "SSH Tunnel Jump Server Host" - description: - "Hostname of the jump server host that allows inbound\ - \ ssh tunnel." - type: "string" - order: 1 - tunnel_port: - title: "SSH Connection Port" - description: - "Port on the proxy/jump server that accepts inbound ssh\ - \ connections." - type: "integer" - minimum: 0 - maximum: 65536 - default: 22 - examples: - - "22" - order: 2 - tunnel_user: - title: "SSH Login Username" - description: "OS-level username for logging into the jump server host" - type: "string" - order: 3 - tunnel_user_password: - title: "Password" - description: "OS-level password for logging into the jump server host" - type: "string" - airbyte_secret: true - order: 4 + - title: "No Tunnel" + required: + - "tunnel_method" + properties: + tunnel_method: + description: "No ssh tunnel needed to connect to database" + type: "string" + const: "NO_TUNNEL" + order: 0 + - title: "SSH Key Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "ssh_key" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and ssh key" + type: "string" + const: "SSH_KEY_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host." + type: "string" + order: 3 + ssh_key: + title: "SSH Private Key" + description: "OS-level user account ssh key credentials in RSA PEM\ + \ format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )" + type: "string" + airbyte_secret: true + multiline: true + order: 4 + - title: "Password Authentication" + required: + - "tunnel_method" + - "tunnel_host" + - "tunnel_port" + - "tunnel_user" + - "tunnel_user_password" + properties: + tunnel_method: + description: "Connect through a jump server tunnel host using username\ + \ and password authentication" + type: "string" + const: "SSH_PASSWORD_AUTH" + order: 0 + tunnel_host: + title: "SSH Tunnel Jump Server Host" + description: "Hostname of the jump server host that allows inbound\ + \ ssh tunnel." + type: "string" + order: 1 + tunnel_port: + title: "SSH Connection Port" + description: "Port on the proxy/jump server that accepts inbound ssh\ + \ connections." + type: "integer" + minimum: 0 + maximum: 65536 + default: 22 + examples: + - "22" + order: 2 + tunnel_user: + title: "SSH Login Username" + description: "OS-level username for logging into the jump server host" + type: "string" + order: 3 + tunnel_user_password: + title: "Password" + description: "OS-level password for logging into the jump server host" + type: "string" + airbyte_secret: true + order: 4 supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -4832,8 +4603,8 @@ title: "PrestaShop Spec" type: "object" required: - - "url" - - "access_key" + - "url" + - "access_key" additionalProperties: false properties: url: @@ -4841,8 +4612,7 @@ description: "Shop URL without trailing slash (domain name or IP address)" access_key: type: "string" - description: - "Your PrestaShop access key. See the docs for info on how to obtain this." airbyte_secret: true supportsNormalization: false @@ -4856,24 +4626,22 @@ title: "Source Quickbooks Singer Spec" type: "object" required: - - "client_id" - - "client_secret" - - "refresh_token" - - "realm_id" - - "user_agent" - - "start_date" - - "sandbox" + - "client_id" + - "client_secret" + - "refresh_token" + - "realm_id" + - "user_agent" + - "start_date" + - "sandbox" additionalProperties: false properties: client_id: type: "string" - description: - "Identifies which app is making the request. Obtain this value\ + description: "Identifies which app is making the request. Obtain this value\ \ from the Keys tab on the app profile via My Apps on the developer site.\ \ There are two versions of this key: development and production" client_secret: - description: - " Obtain this value from the Keys tab on the app profile via\ + description: " Obtain this value from the Keys tab on the app profile via\ \ My Apps on the developer site. There are two versions of this key: development\ \ and production" type: "string" @@ -4883,24 +4651,21 @@ type: "string" airbyte_secret: true realm_id: - description: - "Labeled Company ID. The Make API Calls panel is populated\ + description: "Labeled Company ID. The Make API Calls panel is populated\ \ with the realm id and the current access token" type: "string" airbyte_secret: true user_agent: type: "string" - description: - "Process and email for API logging purposes. Example: tap-quickbooks\ + description: "Process and email for API logging purposes. Example: tap-quickbooks\ \ " start_date: - description: - "The default value to use if no bookmark exists for an endpoint\ + description: "The default value to use if no bookmark exists for an endpoint\ \ (rfc3339 date string) E.g, 2021-03-20T00:00:00Z" type: "string" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2021-03-20T00:00:00Z" + - "2021-03-20T00:00:00Z" sandbox: description: "Development or Production." type: "boolean" @@ -4916,22 +4681,20 @@ title: "Recharge Spec" type: "object" required: - - "start_date" - - "access_token" + - "start_date" + - "access_token" additionalProperties: false properties: start_date: type: "string" - description: - "The date from which you'd like to replicate data for Recharge\ + description: "The date from which you'd like to replicate data for Recharge\ \ API, in the format YYYY-MM-DDT00:00:00Z." examples: - - "2021-05-14T00:00:00Z" + - "2021-05-14T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" access_token: type: "string" - description: - "The value of the Access Token generated. See the docs for more\ \ information" airbyte_secret: true @@ -4946,13 +4709,12 @@ title: "Recurly Source Spec" type: "object" required: - - "api_key" + - "api_key" additionalProperties: false properties: api_key: type: "string" - description: - "Recurly API Key. See the docs for more information on how to generate this key." supportsNormalization: false supportsDBT: false @@ -4965,16 +4727,15 @@ title: "Redshift Source Spec" type: "object" required: - - "host" - - "port" - - "database" - - "username" - - "password" + - "host" + - "port" + - "database" + - "username" + - "password" additionalProperties: false properties: host: - description: - "Host Endpoint of the Redshift Cluster (must include the cluster-id,\ + description: "Host Endpoint of the Redshift Cluster (must include the cluster-id,\ \ region and end with .redshift.amazonaws.com)" type: "string" port: @@ -4984,12 +4745,12 @@ maximum: 65536 default: 5439 examples: - - "5439" + - "5439" database: description: "Name of the database." type: "string" examples: - - "master" + - "master" username: description: "Username to use to access the database." type: "string" @@ -5008,7 +4769,7 @@ title: "Retently Api Spec" type: "object" required: - - "api_key" + - "api_key" additionalProperties: false properties: api_key: @@ -5027,170 +4788,153 @@ properties: dataset: title: "Dataset" - description: - "This source creates one table per connection, this field is\ + description: "This source creates one table per connection, this field is\ \ the name of that table. This should include only letters, numbers, dash\ \ and underscores. Note that this may be altered according to destination." pattern: "^([A-Za-z0-9-_]+)$" type: "string" path_pattern: title: "Path Pattern" - description: - "Add at least 1 pattern here to match filepaths against. Use\ + description: "Add at least 1 pattern here to match filepaths against. Use\ \ | to separate multiple patterns. Airbyte uses these patterns to determine\ \ which files to pick up from the provider storage. See wcmatch.glob to understand pattern syntax (GLOBSTAR\ \ and SPLIT flags are enabled). Use pattern ** to pick\ \ up all files." examples: - - "**" - - "myFolder/myTableFiles/*.csv|myFolder/myOtherTableFiles/*.csv" + - "**" + - "myFolder/myTableFiles/*.csv|myFolder/myOtherTableFiles/*.csv" type: "string" schema: title: "Schema" - description: - "Optionally provide a schema to enforce, as a valid JSON string.\ + description: "Optionally provide a schema to enforce, as a valid JSON string.\ \ Ensure this is a mapping of { \"column\" : \"type\" },\ \ where types are valid JSON Schema datatypes. Leave as {} to auto-infer\ \ the schema." default: "{}" examples: - - "{\"column_1\": \"number\", \"column_2\": \"string\", \"column_3\": \"\ - array\", \"column_4\": \"object\", \"column_5\": \"boolean\"}" + - "{\"column_1\": \"number\", \"column_2\": \"string\", \"column_3\": \"\ + array\", \"column_4\": \"object\", \"column_5\": \"boolean\"}" type: "string" format: title: "Format" default: "csv" type: "object" oneOf: - - title: "csv" - description: - "This connector utilises PyArrow (Apache Arrow) for CSV parsing." - type: "object" - properties: - filetype: - title: "Filetype" - const: "csv" - type: "string" - delimiter: - title: "Delimiter" - description: - "The character delimiting individual cells in the CSV\ - \ data. This may only be a 1-character string." - default: "," - minLength: 1 - type: "string" - quote_char: - title: "Quote Char" - description: - "The character used optionally for quoting CSV values.\ - \ To disallow quoting, make this field blank." - default: '"' - type: "string" - escape_char: - title: "Escape Char" - description: - "The character used optionally for escaping special characters.\ - \ To disallow escaping, leave this field blank." - type: "string" - encoding: - title: "Encoding" - description: - "The character encoding of the CSV data. Leave blank\ - \ to default to UTF-8. See list of python encodings for allowable options." - type: "string" - double_quote: - title: "Double Quote" - description: - "Whether two quotes in a quoted CSV value denote a single\ - \ quote in the data." - default: true - type: "boolean" - newlines_in_values: - title: "Newlines In Values" - description: - "Whether newline characters are allowed in CSV values.\ - \ Turning this on may affect performance. Leave blank to default\ - \ to False." - default: false - type: "boolean" - block_size: - title: "Block Size" - description: - "The chunk size in bytes to process at a time in memory\ - \ from each file. If your data is particularly wide and failing\ - \ during schema detection, increasing this should solve it. Beware\ - \ of raising this too high as you could hit OOM errors." - default: 10000 - type: "integer" - additional_reader_options: - title: "Additional Reader Options" - description: - "Optionally add a valid JSON string here to provide additional\ - \ options to the csv reader. Mappings must correspond to options\ - \ detailed here. 'column_types' is used internally\ - \ to handle schema so overriding that would likely cause problems." - default: "{}" - examples: - - "{\"timestamp_parsers\": [\"%m/%d/%Y %H:%M\", \"%Y/%m/%d %H:%M\"\ - ], \"strings_can_be_null\": true, \"null_values\": [\"NA\", \"NULL\"\ - ]}" - type: "string" - advanced_options: - title: "Advanced Options" - description: - "Optionally add a valid JSON string here to provide additional\ - \ Pyarrow ReadOptions. Specify 'column_names'\ - \ here if your CSV doesn't have header, or if you want to use custom\ - \ column names. 'block_size' and 'encoding' are already used above,\ - \ specify them again here will override the values above." - default: "{}" - examples: - - '{"column_names": ["column1", "column2"]}' - type: "string" - - title: "parquet" - description: - "This connector utilises PyArrow (Apache Arrow) for Parquet parsing." - type: "object" - properties: - filetype: - title: "Filetype" - const: "parquet" + - title: "csv" + description: "This connector utilises PyArrow (Apache Arrow) for CSV parsing." + type: "object" + properties: + filetype: + title: "Filetype" + const: "csv" + type: "string" + delimiter: + title: "Delimiter" + description: "The character delimiting individual cells in the CSV\ + \ data. This may only be a 1-character string." + default: "," + minLength: 1 + type: "string" + quote_char: + title: "Quote Char" + description: "The character used optionally for quoting CSV values.\ + \ To disallow quoting, make this field blank." + default: "\"" + type: "string" + escape_char: + title: "Escape Char" + description: "The character used optionally for escaping special characters.\ + \ To disallow escaping, leave this field blank." + type: "string" + encoding: + title: "Encoding" + description: "The character encoding of the CSV data. Leave blank\ + \ to default to UTF-8. See list of python encodings for allowable options." + type: "string" + double_quote: + title: "Double Quote" + description: "Whether two quotes in a quoted CSV value denote a single\ + \ quote in the data." + default: true + type: "boolean" + newlines_in_values: + title: "Newlines In Values" + description: "Whether newline characters are allowed in CSV values.\ + \ Turning this on may affect performance. Leave blank to default\ + \ to False." + default: false + type: "boolean" + block_size: + title: "Block Size" + description: "The chunk size in bytes to process at a time in memory\ + \ from each file. If your data is particularly wide and failing\ + \ during schema detection, increasing this should solve it. Beware\ + \ of raising this too high as you could hit OOM errors." + default: 10000 + type: "integer" + additional_reader_options: + title: "Additional Reader Options" + description: "Optionally add a valid JSON string here to provide additional\ + \ options to the csv reader. Mappings must correspond to options\ + \ detailed here. 'column_types' is used internally\ + \ to handle schema so overriding that would likely cause problems." + default: "{}" + examples: + - "{\"timestamp_parsers\": [\"%m/%d/%Y %H:%M\", \"%Y/%m/%d %H:%M\"\ + ], \"strings_can_be_null\": true, \"null_values\": [\"NA\", \"NULL\"\ + ]}" + type: "string" + advanced_options: + title: "Advanced Options" + description: "Optionally add a valid JSON string here to provide additional\ + \ Pyarrow ReadOptions. Specify 'column_names'\ + \ here if your CSV doesn't have header, or if you want to use custom\ + \ column names. 'block_size' and 'encoding' are already used above,\ + \ specify them again here will override the values above." + default: "{}" + examples: + - "{\"column_names\": [\"column1\", \"column2\"]}" + type: "string" + - title: "parquet" + description: "This connector utilises PyArrow (Apache Arrow) for Parquet parsing." + type: "object" + properties: + filetype: + title: "Filetype" + const: "parquet" + type: "string" + buffer_size: + title: "Buffer Size" + description: "Perform read buffering when deserializing individual\ + \ column chunks. By default every group column will be loaded fully\ + \ to memory. This option can help to optimize a work with memory\ + \ if your data is particularly wide or failing during detection\ + \ of OOM errors." + default: 0 + type: "integer" + columns: + title: "Columns" + description: "If you only want to sync a subset of the columns from\ + \ the file(s), add the columns you want here. Leave it empty to\ + \ sync all columns." + type: "array" + items: type: "string" - buffer_size: - title: "Buffer Size" - description: - "Perform read buffering when deserializing individual\ - \ column chunks. By default every group column will be loaded fully\ - \ to memory. This option can help to optimize a work with memory\ - \ if your data is particularly wide or failing during detection\ - \ of OOM errors." - default: 0 - type: "integer" - columns: - title: "Columns" - description: - "If you only want to sync a subset of the columns from\ - \ the file(s), add the columns you want here. Leave it empty to\ - \ sync all columns." - type: "array" - items: - type: "string" - batch_size: - title: "Batch Size" - description: - "Maximum number of records per batch. Batches may be\ - \ smaller if there aren’t enough rows in the file. This option can\ - \ help to optimize a work with memory if your data is particularly\ - \ wide or failing during detection of OOM errors." - default: 65536 - type: "integer" + batch_size: + title: "Batch Size" + description: "Maximum number of records per batch. Batches may be\ + \ smaller if there aren’t enough rows in the file. This option can\ + \ help to optimize a work with memory if your data is particularly\ + \ wide or failing during detection of OOM errors." + default: 65536 + type: "integer" provider: title: "S3: Amazon Web Services" type: "object" @@ -5201,24 +4945,21 @@ type: "string" aws_access_key_id: title: "Aws Access Key Id" - description: - "In order to access private Buckets stored on AWS S3, this\ + description: "In order to access private Buckets stored on AWS S3, this\ \ connector requires credentials with the proper permissions. If accessing\ \ publicly available data, this field is not necessary." airbyte_secret: true type: "string" aws_secret_access_key: title: "Aws Secret Access Key" - description: - "In order to access private Buckets stored on AWS S3, this\ + description: "In order to access private Buckets stored on AWS S3, this\ \ connector requires credentials with the proper permissions. If accessing\ \ publicly available data, this field is not necessary." airbyte_secret: true type: "string" path_prefix: title: "Path Prefix" - description: - "By providing a path-like prefix (e.g. myFolder/thisTable/)\ + description: "By providing a path-like prefix (e.g. myFolder/thisTable/)\ \ under which all the relevant files sit, we can optimise finding\ \ these in S3. This is optional but recommended if your bucket contains\ \ many folders/files." @@ -5226,8 +4967,7 @@ type: "string" endpoint: title: "Endpoint" - description: - "Endpoint to an S3 compatible service. Leave empty to use\ + description: "Endpoint to an S3 compatible service. Leave empty to use\ \ AWS." default: "" type: "string" @@ -5240,18 +4980,18 @@ description: "Allow self signed certificates" type: "boolean" required: - - "bucket" + - "bucket" required: - - "dataset" - - "path_pattern" - - "provider" + - "dataset" + - "path_pattern" + - "provider" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" - - "append_dedup" + - "overwrite" + - "append" + - "append_dedup" - dockerImage: "airbyte/source-salesloft:0.1.0" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/salesloft" @@ -5260,10 +5000,10 @@ title: "Source Salesloft Spec" type: "object" required: - - "client_id" - - "client_secret" - - "refresh_token" - - "start_date" + - "client_id" + - "client_secret" + - "refresh_token" + - "start_date" additionalProperties: false properties: client_id: @@ -5279,12 +5019,11 @@ airbyte_secret: true start_date: type: "string" - description: - "The date from which you'd like to replicate data for Salesloft\ + description: "The date from which you'd like to replicate data for Salesloft\ \ API, in the format YYYY-MM-DDT00:00:00Z. All data generated after this\ \ date will be replicated." examples: - - "2020-11-16T00:00:00Z" + - "2020-11-16T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" supportsNormalization: false supportsDBT: false @@ -5297,58 +5036,52 @@ title: "Salesforce Source Spec" type: "object" required: - - "client_id" - - "client_secret" - - "refresh_token" - - "start_date" - - "api_type" + - "client_id" + - "client_secret" + - "refresh_token" + - "start_date" + - "api_type" additionalProperties: false properties: client_id: - description: - "The Consumer Key that can be found when viewing your app in\ + description: "The Consumer Key that can be found when viewing your app in\ \ Salesforce" type: "string" client_secret: - description: - "The Consumer Secret that can be found when viewing your app\ + description: "The Consumer Secret that can be found when viewing your app\ \ in Salesforce" type: "string" airbyte_secret: true refresh_token: - description: - "Salesforce Refresh Token used for Airbyte to access your Salesforce\ + description: "Salesforce Refresh Token used for Airbyte to access your Salesforce\ \ account. If you don't know what this is, follow this guide to retrieve it." type: "string" airbyte_secret: true start_date: - description: - "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ \ data before this date will not be replicated. This field uses the \"\ updated\" field if available, otherwise the \"created\" fields if they\ \ are available for a stream." type: "string" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2021-07-25T00:00:00Z" + - "2021-07-25T00:00:00Z" is_sandbox: - description: - "Whether or not the the app is in a Salesforce sandbox. If\ + description: "Whether or not the the app is in a Salesforce sandbox. If\ \ you do not know what this, assume it is false. We provide more info\ \ on this field in the docs." type: "boolean" default: false api_type: - description: - "Unless you know that you are transferring a very small amount\ + description: "Unless you know that you are transferring a very small amount\ \ of data, prefer using the BULK API. This will help avoid using up all\ \ of your API call quota with Salesforce. Valid values are BULK or REST." type: "string" enum: - - "BULK" - - "REST" + - "BULK" + - "REST" default: "BULK" supportsNormalization: false supportsDBT: false @@ -5358,10 +5091,10 @@ oauth2Specification: rootObject: [] oauthFlowInitParameters: - - - "client_id" - - - "client_secret" + - - "client_id" + - - "client_secret" oauthFlowOutputParameters: - - - "refresh_token" + - - "refresh_token" - dockerImage: "airbyte/source-sendgrid:0.2.6" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/sendgrid" @@ -5370,21 +5103,19 @@ title: "Sendgrid Spec" type: "object" required: - - "apikey" + - "apikey" additionalProperties: false properties: apikey: type: "string" - description: - "API Key, use admin to generate this key." start_time: type: "integer" - description: - "Start time in timestamp integer format. Any data before this\ + description: "Start time in timestamp integer format. Any data before this\ \ timestamp will not be replicated." examples: - - 1558359837 + - 1558359837 supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -5396,70 +5127,68 @@ title: "Shopify Source CDK Specifications" type: "object" required: - - "shop" - - "start_date" - - "auth_method" + - "shop" + - "start_date" + - "auth_method" additionalProperties: false properties: shop: type: "string" - description: - "The name of the shopify store. For https://EXAMPLE.myshopify.com,\ + description: "The name of the shopify store. For https://EXAMPLE.myshopify.com,\ \ the shop name is 'EXAMPLE'." start_date: type: "string" description: "The date you would like to replicate data. Format: YYYY-MM-DD." examples: - - "2021-01-01" + - "2021-01-01" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" auth_method: title: "Shopify Authorization Method" type: "object" oneOf: - - type: "object" - title: "OAuth2.0" - required: - - "client_id" - - "client_secret" + - type: "object" + title: "OAuth2.0" + required: + - "client_id" + - "client_secret" + - "access_token" + properties: + auth_method: + type: "string" + const: "access_token" + enum: - "access_token" - properties: - auth_method: - type: "string" - const: "access_token" - enum: - - "access_token" - default: "access_token" - order: 0 - client_id: - type: "string" - description: "The API Key of the Shopify developer application." - airbyte_secret: true - client_secret: - type: "string" - description: "The API Secret the Shopify developer application." - airbyte_secret: true - access_token: - type: "string" - description: "Access Token for making authenticated requests." - airbyte_secret: true - - title: "API Password" - type: "object" - required: + default: "access_token" + order: 0 + client_id: + type: "string" + description: "The API Key of the Shopify developer application." + airbyte_secret: true + client_secret: + type: "string" + description: "The API Secret the Shopify developer application." + airbyte_secret: true + access_token: + type: "string" + description: "Access Token for making authenticated requests." + airbyte_secret: true + - title: "API Password" + type: "object" + required: + - "api_password" + properties: + auth_method: + type: "string" + const: "api_password" + enum: - "api_password" - properties: - auth_method: - type: "string" - const: "api_password" - enum: - - "api_password" - default: "api_password" - order: 0 - api_password: - type: "string" - description: - "The API PASSWORD for your private application in `Shopify`\ - \ shop." - airbyte_secret: true + default: "api_password" + order: 0 + api_password: + type: "string" + description: "The API PASSWORD for your private application in `Shopify`\ + \ shop." + airbyte_secret: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -5467,13 +5196,13 @@ auth_type: "oauth2.0" oauth2Specification: rootObject: - - "auth_method" - - "0" + - "auth_method" + - "0" oauthFlowInitParameters: - - - "client_id" - - - "client_secret" + - - "client_id" + - - "client_secret" oauthFlowOutputParameters: - - - "access_token" + - - "access_token" - dockerImage: "airbyte/source-shortio:0.1.1" spec: documentationUrl: "https://developers.short.io/reference" @@ -5482,9 +5211,9 @@ title: "Shortio Spec" type: "object" required: - - "domain_id" - - "secret_key" - - "start_date" + - "domain_id" + - "secret_key" + - "start_date" additionalProperties: false properties: domain_id: @@ -5510,33 +5239,31 @@ title: "Slack Spec" type: "object" required: - - "start_date" - - "lookback_window" - - "join_channels" + - "start_date" + - "lookback_window" + - "join_channels" additionalProperties: true properties: start_date: type: "string" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" - description: - "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ \ data before this date will not be replicated." examples: - - "2017-01-25T00:00:00Z" + - "2017-01-25T00:00:00Z" title: "Start Date" lookback_window: type: "integer" title: "Threads Lookback window (Days)" description: "How far into the past to look for messages in threads." examples: - - 7 - - 14 + - 7 + - 14 join_channels: type: "boolean" default: true title: "Join all channels" - description: - "Whether to join all channels or to sync data only from channels\ + description: "Whether to join all channels or to sync data only from channels\ \ the bot is already in. If false, you'll need to manually add the bot\ \ to all the channels from which you'd like to sync messages. " credentials: @@ -5544,70 +5271,65 @@ description: "Choose how to authenticate into Slack" type: "object" oneOf: - - type: "object" - title: "Sign in via Slack (OAuth)" - required: - - "access_token" - - "client_id" - - "client_secret" - - "option_title" - properties: - option_title: - type: "string" - const: "Default OAuth2.0 authorization" - client_id: - title: "Client ID" - description: - "Slack client_id. See our docs if you need help finding this id." - type: "string" - examples: - - "slack-client-id-example" - client_secret: - title: "Client Secret" - description: - "Slack client_secret. See our docs if you need help finding this secret." - type: "string" - examples: - - "slack-client-secret-example" - airbyte_secret: true - access_token: - title: "Access token" - description: - "Slack access_token. See our docs if you need help generating the token." - type: "string" - examples: - - "slack-access-token-example" - airbyte_secret: true - refresh_token: - title: "Refresh token" - description: - "Slack refresh_token. See our docs if you need help generating the token." - type: "string" - examples: - - "slack-refresh-token-example" - airbyte_secret: true - order: 0 - - type: "object" - title: "API Token" - required: - - "api_token" - - "option_title" - properties: - option_title: - type: "string" - const: "API Token Credentials" - api_token: - type: "string" - title: "API Token" - description: - "A Slack bot token. See the docs for instructions on how to generate it." - airbyte_secret: true - order: 1 + - type: "object" + title: "Sign in via Slack (OAuth)" + required: + - "access_token" + - "client_id" + - "client_secret" + - "option_title" + properties: + option_title: + type: "string" + const: "Default OAuth2.0 authorization" + client_id: + title: "Client ID" + description: "Slack client_id. See our docs if you need help finding this id." + type: "string" + examples: + - "slack-client-id-example" + client_secret: + title: "Client Secret" + description: "Slack client_secret. See our docs if you need help finding this secret." + type: "string" + examples: + - "slack-client-secret-example" + airbyte_secret: true + access_token: + title: "Access token" + description: "Slack access_token. See our docs if you need help generating the token." + type: "string" + examples: + - "slack-access-token-example" + airbyte_secret: true + refresh_token: + title: "Refresh token" + description: "Slack refresh_token. See our docs if you need help generating the token." + type: "string" + examples: + - "slack-refresh-token-example" + airbyte_secret: true + order: 0 + - type: "object" + title: "API Token" + required: + - "api_token" + - "option_title" + properties: + option_title: + type: "string" + const: "API Token Credentials" + api_token: + type: "string" + title: "API Token" + description: "A Slack bot token. See the docs for instructions on how to generate it." + airbyte_secret: true + order: 1 supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -5615,14 +5337,14 @@ auth_type: "oauth2.0" oauth2Specification: rootObject: - - "credentials" - - "0" + - "credentials" + - "0" oauthFlowInitParameters: - - - "client_id" - - - "client_secret" + - - "client_id" + - - "client_secret" oauthFlowOutputParameters: - - - "access_token" - - - "refresh_token" + - - "access_token" + - - "refresh_token" - dockerImage: "airbyte/source-smartsheets:0.1.6" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/smartsheets" @@ -5631,14 +5353,13 @@ title: "Smartsheets Source Spec" type: "object" required: - - "access_token" - - "spreadsheet_id" + - "access_token" + - "spreadsheet_id" additionalProperties: false properties: access_token: title: "API Access token" - description: - "Found in Profile > Apps & Integrations > API Access within\ + description: "Found in Profile > Apps & Integrations > API Access within\ \ Smartsheet app" type: "string" airbyte_secret: true @@ -5657,9 +5378,9 @@ title: "Snapchat Marketing Spec" type: "object" required: - - "client_id" - - "client_secret" - - "refresh_token" + - "client_id" + - "client_secret" + - "refresh_token" additionalProperties: false properties: client_id: @@ -5675,18 +5396,16 @@ refresh_token: title: "API Refresh Token" type: "string" - description: - "Refresh Token to get next api key after expiration. Is given\ + description: "Refresh Token to get next api key after expiration. Is given\ \ with API Key" airbyte_secret: true start_date: title: "Start Date" type: "string" - description: - "The start date to sync data. Leave blank for full sync. Format:\ + description: "The start date to sync data. Leave blank for full sync. Format:\ \ YYYY-MM-DD." examples: - - "2021-01-01" + - "2021-01-01" default: "1970-01-01" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" supportsNormalization: false @@ -5700,56 +5419,55 @@ title: "Snowflake Source Spec" type: "object" required: - - "host" - - "role" - - "warehouse" - - "database" - - "schema" - - "username" - - "password" + - "host" + - "role" + - "warehouse" + - "database" + - "schema" + - "username" + - "password" additionalProperties: false properties: host: - description: - "Host domain of the snowflake instance (must include the account,\ + description: "Host domain of the snowflake instance (must include the account,\ \ region, cloud environment, and end with snowflakecomputing.com)." examples: - - "accountname.us-east-2.aws.snowflakecomputing.com" + - "accountname.us-east-2.aws.snowflakecomputing.com" type: "string" title: "Account name" order: 0 role: description: "The role you created for Airbyte to access Snowflake." examples: - - "AIRBYTE_ROLE" + - "AIRBYTE_ROLE" type: "string" title: "Role" order: 1 warehouse: description: "The warehouse you created for Airbyte to access data into." examples: - - "AIRBYTE_WAREHOUSE" + - "AIRBYTE_WAREHOUSE" type: "string" title: "Warehouse" order: 2 database: description: "The database you created for Airbyte to access data into." examples: - - "AIRBYTE_DATABASE" + - "AIRBYTE_DATABASE" type: "string" title: "Database" order: 3 schema: description: "The source Snowflake schema tables." examples: - - "AIRBYTE_SCHEMA" + - "AIRBYTE_SCHEMA" type: "string" title: "Schema" order: 4 username: description: "The username you created to allow Airbyte to access the database." examples: - - "AIRBYTE_USER" + - "AIRBYTE_USER" type: "string" title: "Username" order: 5 @@ -5770,8 +5488,8 @@ title: "Square Source CDK Specifications" type: "object" required: - - "api_key" - - "is_sandbox" + - "api_key" + - "is_sandbox" additionalProperties: false properties: api_key: @@ -5782,26 +5500,24 @@ type: "boolean" description: "Determines the sandbox (true) or production (false) API version" examples: - - true - - false + - true + - false default: true start_date: type: "string" - description: - "The start date to sync data. Leave blank for full sync. Format:\ + description: "The start date to sync data. Leave blank for full sync. Format:\ \ YYYY-MM-DD." examples: - - "2021-01-01" + - "2021-01-01" default: "1970-01-01" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" include_deleted_objects: type: "boolean" - description: - "In some streams there is and option to include deleted objects\ + description: "In some streams there is and option to include deleted objects\ \ (Items, Categories, Discounts, Taxes)" examples: - - true - - false + - true + - false default: false supportsNormalization: false supportsDBT: false @@ -5814,11 +5530,11 @@ title: "Strava Spec" type: "object" required: - - "client_id" - - "client_secret" - - "refresh_token" - - "athlete_id" - - "start_date" + - "client_id" + - "client_secret" + - "refresh_token" + - "athlete_id" + - "start_date" additionalProperties: false properties: client_id: @@ -5826,32 +5542,32 @@ description: "Strava Client ID" pattern: "^[0-9_\\-]+$" examples: - - "12345" + - "12345" client_secret: type: "string" description: "Strava Client Secret" pattern: "^[0-9a-fA-F]+$" examples: - - "fc6243f283e51f6ca989aab298b17da125496f50" + - "fc6243f283e51f6ca989aab298b17da125496f50" airbyte_secret: true refresh_token: type: "string" description: "Strava Refresh Token with activity:read_all permissions" pattern: "^[0-9a-fA-F]+$" examples: - - "fc6243f283e51f6ca989aab298b17da125496f50" + - "fc6243f283e51f6ca989aab298b17da125496f50" airbyte_secret: true athlete_id: type: "integer" description: "Strava Athlete ID" pattern: "^[0-9_\\-]+$" examples: - - "17831421" + - "17831421" start_date: type: "string" description: "Start Query Timestamp in UTC" examples: - - "2016-12-31 23:59:59" + - "2016-12-31 23:59:59" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -5863,38 +5579,34 @@ title: "Stripe Source Spec" type: "object" required: - - "client_secret" - - "account_id" - - "start_date" + - "client_secret" + - "account_id" + - "start_date" additionalProperties: false properties: client_secret: type: "string" pattern: "^(s|r)k_(live|test)_[a-zA-Z0-9]+$" - description: - "Stripe API key (usually starts with 'sk_live_'; find yours\ + description: "Stripe API key (usually starts with 'sk_live_'; find yours\ \ here)." airbyte_secret: true account_id: type: "string" - description: - "Your Stripe account ID (starts with 'acct_', find yours here)." start_date: type: "string" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" - description: - "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ \ data before this date will not be replicated." examples: - - "2017-01-25T00:00:00Z" + - "2017-01-25T00:00:00Z" lookback_window_days: type: "integer" title: "Lookback Window (in days)" default: 0 minimum: 0 - description: - "When set, the connector will always reload data from the past\ + description: "When set, the connector will always reload data from the past\ \ N days, where N is the value set here. This is useful if your data is\ \ updated after creation." supportsNormalization: false @@ -5908,7 +5620,7 @@ title: "SurveyMonkey Spec" type: "object" required: - - "start_date" + - "start_date" additionalProperties: true properties: start_date: @@ -5917,13 +5629,12 @@ description: "The date from which you'd like to replicate the data" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z?$" examples: - - "2021-01-01T00:00:00Z" + - "2021-01-01T00:00:00Z" access_token: title: "Access Token" type: "string" airbyte_secret: true - description: - "API Token. See the docs for information on how to generate this key." supportsNormalization: false supportsDBT: false @@ -5934,7 +5645,7 @@ rootObject: [] oauthFlowInitParameters: [] oauthFlowOutputParameters: - - - "access_token" + - - "access_token" - dockerImage: "airbyte/source-tempo:0.2.4" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/" @@ -5943,7 +5654,7 @@ title: "Tempo Spec" type: "object" required: - - "api_token" + - "api_token" additionalProperties: false properties: api_token: @@ -5965,40 +5676,39 @@ title: "Environment" default: "Production" oneOf: - - title: "Production" - type: "object" - properties: - environment: - title: "Environment" - const: "prod" - type: "string" - app_id: - title: "App Id" - description: "The App id applied by the developer." - type: "string" - secret: - title: "Secret" - description: "The private key of the developer's application." - airbyte_secret: true - type: "string" - required: - - "app_id" - - "secret" - - title: "Sandbox" - type: "object" - properties: - environment: - title: "Environment" - const: "sandbox" - type: "string" - advertiser_id: - title: "Advertiser Id" - description: - "The Advertiser ID which generated for the developer's\ - \ Sandbox application." - type: "string" - required: - - "advertiser_id" + - title: "Production" + type: "object" + properties: + environment: + title: "Environment" + const: "prod" + type: "string" + app_id: + title: "App Id" + description: "The App id applied by the developer." + type: "string" + secret: + title: "Secret" + description: "The private key of the developer's application." + airbyte_secret: true + type: "string" + required: + - "app_id" + - "secret" + - title: "Sandbox" + type: "object" + properties: + environment: + title: "Environment" + const: "sandbox" + type: "string" + advertiser_id: + title: "Advertiser Id" + description: "The Advertiser ID which generated for the developer's\ + \ Sandbox application." + type: "string" + required: + - "advertiser_id" type: "object" access_token: title: "Access Token" @@ -6012,14 +5722,14 @@ pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}$" type: "string" required: - - "access_token" + - "access_token" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "overwrite" - - "append" - - "append_dedup" + - "overwrite" + - "append" + - "append_dedup" - dockerImage: "airbyte/source-trello:0.1.2" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/trello" @@ -6028,34 +5738,31 @@ title: "Trello Spec" type: "object" required: - - "token" - - "key" - - "start_date" + - "token" + - "key" + - "start_date" additionalProperties: true properties: token: type: "string" title: "API token" - description: - "A Trello token. See the docs for instructions on how to generate it." airbyte_secret: true key: type: "string" title: "API key" - description: - "A Trello token. See the docs for instructions on how to generate it." airbyte_secret: true start_date: type: "string" title: "Start date" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}.[0-9]{3}Z$" - description: - "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ + description: "UTC date and time in the format 2017-01-25T00:00:00Z. Any\ \ data before this date will not be replicated." examples: - - "2021-03-01T00:00:00.000Z" + - "2021-03-01T00:00:00.000Z" supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -6065,8 +5772,8 @@ rootObject: [] oauthFlowInitParameters: [] oauthFlowOutputParameters: - - - "token" - - - "key" + - - "token" + - - "key" - dockerImage: "airbyte/source-twilio:0.1.1" spec: documentationUrl: "https://hub.docker.com/r/airbyte/source-twilio" @@ -6075,9 +5782,9 @@ title: "Twilio Spec" type: "object" required: - - "account_sid" - - "auth_token" - - "start_date" + - "account_sid" + - "auth_token" + - "start_date" additionalProperties: false properties: account_sid: @@ -6092,18 +5799,17 @@ type: "string" start_date: title: "Replication Start Date" - description: - "UTC date and time in the format 2020-10-01T00:00:00Z. Any\ + description: "UTC date and time in the format 2020-10-01T00:00:00Z. Any\ \ data before this date will not be replicated." pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" examples: - - "2020-10-01T00:00:00Z" + - "2020-10-01T00:00:00Z" type: "string" supportsIncremental: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: - - "append" + - "append" - dockerImage: "airbyte/source-typeform:0.1.2" spec: documentationUrl: "https://docs.airbyte.io/integrations/sources/typeform" @@ -6112,15 +5818,15 @@ title: "Typeform Spec" type: "object" required: - - "token" - - "start_date" + - "token" + - "start_date" additionalProperties: true properties: start_date: type: "string" description: "The date you would like to replicate data. Format: YYYY-MM-DDTHH:mm:ss[Z]." examples: - - "2020-01-01T00:00:00Z" + - "2020-01-01T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" token: type: "string" @@ -6128,8 +5834,7 @@ airbyte_secret: true form_ids: title: "Form IDs to replicate" - description: - "When this parameter is set, the connector will replicate data\ + description: "When this parameter is set, the connector will replicate data\ \ only from the input forms. Otherwise, all forms in your Typeform account\ \ will be replicated. You can find form IDs in your form URLs. For example,\ \ in the URL \"https://mysite.typeform.com/to/u6nXL7\" the form_id is\ @@ -6149,31 +5854,29 @@ title: "https://api.census.gov/ Source Spec" type: "object" required: - - "api_key" - - "query_path" + - "api_key" + - "query_path" additionalProperties: false properties: query_params: type: "string" - description: - "The query parameters portion of the GET request, without the\ + description: "The query parameters portion of the GET request, without the\ \ api key" pattern: "^\\w+=[\\w,:*]+(&(?!key)\\w+=[\\w,:*]+)*$" examples: - - "get=NAME,NAICS2017_LABEL,LFO_LABEL,EMPSZES_LABEL,ESTAB,PAYANN,PAYQTR1,EMP&for=us:*&NAICS2017=72&LFO=001&EMPSZES=001" - - "get=MOVEDIN,GEOID1,GEOID2,MOVEDOUT,FULL1_NAME,FULL2_NAME,MOVEDNET&for=county:*" + - "get=NAME,NAICS2017_LABEL,LFO_LABEL,EMPSZES_LABEL,ESTAB,PAYANN,PAYQTR1,EMP&for=us:*&NAICS2017=72&LFO=001&EMPSZES=001" + - "get=MOVEDIN,GEOID1,GEOID2,MOVEDOUT,FULL1_NAME,FULL2_NAME,MOVEDNET&for=county:*" query_path: type: "string" description: "The path portion of the GET request" pattern: "^data(\\/[\\w\\d]+)+$" examples: - - "data/2019/cbp" - - "data/2018/acs" - - "data/timeseries/healthins/sahie" + - "data/2019/cbp" + - "data/2018/acs" + - "data/timeseries/healthins/sahie" api_key: type: "string" - description: - "Your API Key. Get your key here." airbyte_secret: true supportsNormalization: false @@ -6187,22 +5890,20 @@ title: "Zendesk Chat Spec" type: "object" required: - - "start_date" - - "access_token" + - "start_date" + - "access_token" additionalProperties: false properties: start_date: type: "string" - description: - "The date from which you'd like to replicate data for Zendesk\ + description: "The date from which you'd like to replicate data for Zendesk\ \ Chat API, in the format YYYY-MM-DDT00:00:00Z." examples: - - "2021-02-01T00:00:00Z" + - "2021-02-01T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" access_token: type: "string" - description: - "The value of the Access Token generated. See the docs for\ \ more information" airbyte_secret: true @@ -6217,17 +5918,16 @@ title: "Zendesk Sunshine Spec" type: "object" required: - - "api_token" - - "email" - - "start_date" - - "subdomain" + - "api_token" + - "email" + - "start_date" + - "subdomain" additionalProperties: false properties: api_token: type: "string" airbyte_secret: true - description: - "API Token. See the docs for information on how to generate this key." email: type: "string" @@ -6252,19 +5952,18 @@ title: "Source Zendesk Support Spec" type: "object" required: - - "start_date" - - "subdomain" - - "auth_method" + - "start_date" + - "subdomain" + - "auth_method" additionalProperties: false properties: start_date: type: "string" - description: - "The date from which you'd like to replicate data for Zendesk\ + description: "The date from which you'd like to replicate data for Zendesk\ \ Support API, in the format YYYY-MM-DDT00:00:00Z. All data generated\ \ after this date will be replicated." examples: - - "2020-10-15T00:00:00Z" + - "2020-10-15T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" subdomain: type: "string" @@ -6273,31 +5972,29 @@ title: "ZenDesk Authorization Method" type: "object" default: "api_token" - description: - "Zendesk service provides 2 auth method: API token and oAuth2.\ + description: "Zendesk service provides 2 auth method: API token and oAuth2.\ \ Now only the first one is available. Another one will be added in the\ \ future" oneOf: - - title: "API Token" - type: "object" - required: - - "email" - - "api_token" - additionalProperties: false - properties: - auth_method: - type: "string" - const: "api_token" - email: - type: "string" - description: "The user email for your Zendesk account" - api_token: - type: "string" - description: - "The value of the API token generated. See the docs\ - \ for more information" - airbyte_secret: true + - title: "API Token" + type: "object" + required: + - "email" + - "api_token" + additionalProperties: false + properties: + auth_method: + type: "string" + const: "api_token" + email: + type: "string" + description: "The user email for your Zendesk account" + api_token: + type: "string" + description: "The value of the API token generated. See the docs\ + \ for more information" + airbyte_secret: true supportsNormalization: false supportsDBT: false supported_destination_sync_modes: [] @@ -6309,27 +6006,25 @@ title: "Zendesk Talk Spec" type: "object" required: - - "start_date" - - "subdomain" - - "access_token" - - "email" + - "start_date" + - "subdomain" + - "access_token" + - "email" additionalProperties: false properties: start_date: type: "string" - description: - "The date from which you'd like to replicate data for Zendesk\ + description: "The date from which you'd like to replicate data for Zendesk\ \ Talk API, in the format YYYY-MM-DDT00:00:00Z." examples: - - "2021-04-01T00:00:00Z" + - "2021-04-01T00:00:00Z" pattern: "^[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z$" subdomain: type: "string" description: "The subdomain for your Zendesk Talk" access_token: type: "string" - description: - "The value of the API token generated. See the docs for more information" airbyte_secret: true email: @@ -6346,24 +6041,22 @@ title: "Sentry Spec" type: "object" required: - - "auth_token" - - "organization" - - "project" + - "auth_token" + - "organization" + - "project" additionalProperties: false properties: auth_token: type: "string" title: "Authentication tokens" - description: - "Log into Sentry and then create authentication tokens.For self-hosted, you can find or create\ \ authentication tokens by visiting \"{instance_url_prefix}/settings/account/api/auth-tokens/\"" airbyte_secret: true hostname: type: "string" title: "Host Name" - description: - "Host name of Sentry API server.For self-hosted, specify your\ + description: "Host name of Sentry API server.For self-hosted, specify your\ \ host name here. Otherwise, leave it empty." default: "sentry.io" organization: @@ -6385,14 +6078,13 @@ title: "Source Zoom Singer Spec" type: "object" required: - - "jwt" + - "jwt" additionalProperties: false properties: jwt: title: "JWT Token" type: "string" - description: - "Zoom JWT Token. See the docs for more information on how to obtain this key." airbyte_secret: true supportsNormalization: false @@ -6406,11 +6098,11 @@ title: "Zuora Connector Configuration" type: "object" required: - - "start_date" - - "tenant_endpoint" - - "data_query" - - "client_id" - - "client_secret" + - "start_date" + - "tenant_endpoint" + - "data_query" + - "client_id" + - "client_secret" properties: start_date: type: "string" @@ -6420,49 +6112,46 @@ window_in_days: type: "string" title: "Query Window (in days)" - description: - "The amount of days for each data-chunk begining from start_date.\ + description: "The amount of days for each data-chunk begining from start_date.\ \ Bigger the value - faster the fetch. (0.1 - as for couple of hours,\ \ 1 - as for a Day; 364 - as for a Year)." examples: - - "0.5" - - "1" - - "30" - - "60" - - "90" - - "120" - - "200" - - "364" + - "0.5" + - "1" + - "30" + - "60" + - "90" + - "120" + - "200" + - "364" pattern: "^(0|[1-9]\\d*)(\\.\\d+)?$" default: "90" tenant_endpoint: title: "Tenant Endpoint Location" type: "string" - description: - "Please choose the right endpoint where your Tenant is located.\ + description: "Please choose the right endpoint where your Tenant is located.\ \ More info by this Link" enum: - - "US Production" - - "US Cloud Production" - - "US API Sandbox" - - "US Cloud API Sandbox" - - "US Central Sandbox" - - "US Performance Test" - - "EU Production" - - "EU API Sandbox" - - "EU Central Sandbox" + - "US Production" + - "US Cloud Production" + - "US API Sandbox" + - "US Cloud API Sandbox" + - "US Central Sandbox" + - "US Performance Test" + - "EU Production" + - "EU API Sandbox" + - "EU Central Sandbox" data_query: title: "Data Query Type" type: "string" - description: - "Choose between `Live`, or `Unlimited` - the optimized, replicated\ + description: "Choose between `Live`, or `Unlimited` - the optimized, replicated\ \ database at 12 hours freshness for high volume extraction Link" enum: - - "Live" - - "Unlimited" + - "Live" + - "Unlimited" default: "Live" client_id: type: "string" diff --git a/airbyte-migration/Dockerfile b/airbyte-migration/Dockerfile index 5505ca5db047b..67ec6da29b9b4 100644 --- a/airbyte-migration/Dockerfile +++ b/airbyte-migration/Dockerfile @@ -6,7 +6,7 @@ ENV APPLICATION airbyte-migration WORKDIR /app # Move and run scheduler -COPY bin/${APPLICATION}-0.30.38-alpha.tar ${APPLICATION}.tar +COPY bin/${APPLICATION}-0.30.39-alpha.tar ${APPLICATION}.tar RUN tar xf ${APPLICATION}.tar --strip-components=1 diff --git a/airbyte-scheduler/app/Dockerfile b/airbyte-scheduler/app/Dockerfile index 64deb051efdb9..cd202fdfacbc9 100644 --- a/airbyte-scheduler/app/Dockerfile +++ b/airbyte-scheduler/app/Dockerfile @@ -5,7 +5,7 @@ ENV APPLICATION airbyte-scheduler WORKDIR /app -ADD bin/${APPLICATION}-0.30.38-alpha.tar /app +ADD bin/${APPLICATION}-0.30.39-alpha.tar /app # wait for upstream dependencies to become available before starting server -ENTRYPOINT ["/bin/bash", "-c", "${APPLICATION}-0.30.38-alpha/bin/${APPLICATION}"] +ENTRYPOINT ["/bin/bash", "-c", "${APPLICATION}-0.30.39-alpha/bin/${APPLICATION}"] diff --git a/airbyte-server/Dockerfile b/airbyte-server/Dockerfile index 3d8b94dcc3af4..ba7490567f359 100644 --- a/airbyte-server/Dockerfile +++ b/airbyte-server/Dockerfile @@ -7,7 +7,7 @@ ENV APPLICATION airbyte-server WORKDIR /app -ADD bin/${APPLICATION}-0.30.38-alpha.tar /app +ADD bin/${APPLICATION}-0.30.39-alpha.tar /app # wait for upstream dependencies to become available before starting server -ENTRYPOINT ["/bin/bash", "-c", "${APPLICATION}-0.30.38-alpha/bin/${APPLICATION}"] +ENTRYPOINT ["/bin/bash", "-c", "${APPLICATION}-0.30.39-alpha/bin/${APPLICATION}"] diff --git a/airbyte-webapp/package-lock.json b/airbyte-webapp/package-lock.json index 05c5a760e22c7..6f9c378b273a1 100644 --- a/airbyte-webapp/package-lock.json +++ b/airbyte-webapp/package-lock.json @@ -1,12 +1,12 @@ { "name": "airbyte-webapp", - "version": "0.30.38-alpha", + "version": "0.30.39-alpha", "lockfileVersion": 2, "requires": true, "packages": { "": { "name": "airbyte-webapp", - "version": "0.30.38-alpha", + "version": "0.30.39-alpha", "dependencies": { "@fortawesome/fontawesome-svg-core": "^1.2.36", "@fortawesome/free-brands-svg-icons": "^5.15.4", diff --git a/airbyte-webapp/package.json b/airbyte-webapp/package.json index 0e8338c315b7a..424a861bcec9b 100644 --- a/airbyte-webapp/package.json +++ b/airbyte-webapp/package.json @@ -1,6 +1,6 @@ { "name": "airbyte-webapp", - "version": "0.30.38-alpha", + "version": "0.30.39-alpha", "private": true, "scripts": { "start": "react-scripts start", diff --git a/airbyte-workers/Dockerfile b/airbyte-workers/Dockerfile index f299d2da00586..7fe23fdafd37d 100644 --- a/airbyte-workers/Dockerfile +++ b/airbyte-workers/Dockerfile @@ -23,7 +23,7 @@ ENV APPLICATION airbyte-workers WORKDIR /app # Move worker app -ADD bin/${APPLICATION}-0.30.38-alpha.tar /app +ADD bin/${APPLICATION}-0.30.39-alpha.tar /app # wait for upstream dependencies to become available before starting server -ENTRYPOINT ["/bin/bash", "-c", "${APPLICATION}-0.30.38-alpha/bin/${APPLICATION}"] +ENTRYPOINT ["/bin/bash", "-c", "${APPLICATION}-0.30.39-alpha/bin/${APPLICATION}"] diff --git a/charts/airbyte/Chart.yaml b/charts/airbyte/Chart.yaml index df06fbdfd93d8..5017199a9beaa 100644 --- a/charts/airbyte/Chart.yaml +++ b/charts/airbyte/Chart.yaml @@ -21,7 +21,7 @@ version: 0.3.0 # incremented each time you make changes to the application. Versions are not expected to # follow Semantic Versioning. They should reflect the version the application is using. # It is recommended to use it with quotes. -appVersion: "0.30.38-alpha" +appVersion: "0.30.39-alpha" dependencies: - name: common diff --git a/charts/airbyte/README.md b/charts/airbyte/README.md index 78e1a4e6d176d..1371fb6d0cfeb 100644 --- a/charts/airbyte/README.md +++ b/charts/airbyte/README.md @@ -29,7 +29,7 @@ | `webapp.replicaCount` | Number of webapp replicas | `1` | | `webapp.image.repository` | The repository to use for the airbyte webapp image. | `airbyte/webapp` | | `webapp.image.pullPolicy` | the pull policy to use for the airbyte webapp image | `IfNotPresent` | -| `webapp.image.tag` | The airbyte webapp image tag. Defaults to the chart's AppVersion | `0.30.38-alpha` | +| `webapp.image.tag` | The airbyte webapp image tag. Defaults to the chart's AppVersion | `0.30.39-alpha` | | `webapp.podAnnotations` | Add extra annotations to the webapp pod(s) | `{}` | | `webapp.service.type` | The service type to use for the webapp service | `ClusterIP` | | `webapp.service.port` | The service port to expose the webapp on | `80` | @@ -56,7 +56,7 @@ | `scheduler.replicaCount` | Number of scheduler replicas | `1` | | `scheduler.image.repository` | The repository to use for the airbyte scheduler image. | `airbyte/scheduler` | | `scheduler.image.pullPolicy` | the pull policy to use for the airbyte scheduler image | `IfNotPresent` | -| `scheduler.image.tag` | The airbyte scheduler image tag. Defaults to the chart's AppVersion | `0.30.38-alpha` | +| `scheduler.image.tag` | The airbyte scheduler image tag. Defaults to the chart's AppVersion | `0.30.39-alpha` | | `scheduler.podAnnotations` | Add extra annotations to the scheduler pod | `{}` | | `scheduler.resources.limits` | The resources limits for the scheduler container | `{}` | | `scheduler.resources.requests` | The requested resources for the scheduler container | `{}` | @@ -87,7 +87,7 @@ | `server.replicaCount` | Number of server replicas | `1` | | `server.image.repository` | The repository to use for the airbyte server image. | `airbyte/server` | | `server.image.pullPolicy` | the pull policy to use for the airbyte server image | `IfNotPresent` | -| `server.image.tag` | The airbyte server image tag. Defaults to the chart's AppVersion | `0.30.38-alpha` | +| `server.image.tag` | The airbyte server image tag. Defaults to the chart's AppVersion | `0.30.39-alpha` | | `server.podAnnotations` | Add extra annotations to the server pod | `{}` | | `server.livenessProbe.enabled` | Enable livenessProbe on the server | `true` | | `server.livenessProbe.initialDelaySeconds` | Initial delay seconds for livenessProbe | `30` | @@ -121,7 +121,7 @@ | `worker.replicaCount` | Number of worker replicas | `1` | | `worker.image.repository` | The repository to use for the airbyte worker image. | `airbyte/worker` | | `worker.image.pullPolicy` | the pull policy to use for the airbyte worker image | `IfNotPresent` | -| `worker.image.tag` | The airbyte worker image tag. Defaults to the chart's AppVersion | `0.30.38-alpha` | +| `worker.image.tag` | The airbyte worker image tag. Defaults to the chart's AppVersion | `0.30.39-alpha` | | `worker.podAnnotations` | Add extra annotations to the worker pod(s) | `{}` | | `worker.livenessProbe.enabled` | Enable livenessProbe on the worker | `true` | | `worker.livenessProbe.initialDelaySeconds` | Initial delay seconds for livenessProbe | `30` | diff --git a/charts/airbyte/values.yaml b/charts/airbyte/values.yaml index 3a6872126f436..5c8ddea873ca1 100644 --- a/charts/airbyte/values.yaml +++ b/charts/airbyte/values.yaml @@ -44,7 +44,7 @@ webapp: image: repository: airbyte/webapp pullPolicy: IfNotPresent - tag: 0.30.38-alpha + tag: 0.30.39-alpha ## @param webapp.podAnnotations [object] Add extra annotations to the webapp pod(s) ## @@ -141,7 +141,7 @@ scheduler: image: repository: airbyte/scheduler pullPolicy: IfNotPresent - tag: 0.30.38-alpha + tag: 0.30.39-alpha ## @param scheduler.podAnnotations [object] Add extra annotations to the scheduler pod ## @@ -248,7 +248,7 @@ server: image: repository: airbyte/server pullPolicy: IfNotPresent - tag: 0.30.38-alpha + tag: 0.30.39-alpha ## @param server.podAnnotations [object] Add extra annotations to the server pod ## @@ -360,7 +360,7 @@ worker: image: repository: airbyte/worker pullPolicy: IfNotPresent - tag: 0.30.38-alpha + tag: 0.30.39-alpha ## @param worker.podAnnotations [object] Add extra annotations to the worker pod(s) ## diff --git a/docs/operator-guides/upgrading-airbyte.md b/docs/operator-guides/upgrading-airbyte.md index 7428a1a12be43..6a866c83d2558 100644 --- a/docs/operator-guides/upgrading-airbyte.md +++ b/docs/operator-guides/upgrading-airbyte.md @@ -82,7 +82,7 @@ If you are upgrading from \(i.e. your current version of Airbyte is\) Airbyte ve Here's an example of what it might look like with the values filled in. It assumes that the downloaded `airbyte_archive.tar.gz` is in `/tmp`. ```bash - docker run --rm -v /tmp:/config airbyte/migration:0.30.38-alpha --\ + docker run --rm -v /tmp:/config airbyte/migration:0.30.39-alpha --\ --input /config/airbyte_archive.tar.gz\ --output /config/airbyte_archive_migrated.tar.gz ``` diff --git a/kube/overlays/stable-with-resource-limits/.env b/kube/overlays/stable-with-resource-limits/.env index 5a9c9b0692fb7..52191167c9a42 100644 --- a/kube/overlays/stable-with-resource-limits/.env +++ b/kube/overlays/stable-with-resource-limits/.env @@ -1,4 +1,4 @@ -AIRBYTE_VERSION=0.30.38-alpha +AIRBYTE_VERSION=0.30.39-alpha # Airbyte Internal Database, see https://docs.airbyte.io/operator-guides/configuring-airbyte-db DATABASE_HOST=airbyte-db-svc diff --git a/kube/overlays/stable-with-resource-limits/kustomization.yaml b/kube/overlays/stable-with-resource-limits/kustomization.yaml index 611dbca8a4f23..2585bbdd88d77 100644 --- a/kube/overlays/stable-with-resource-limits/kustomization.yaml +++ b/kube/overlays/stable-with-resource-limits/kustomization.yaml @@ -8,15 +8,15 @@ bases: images: - name: airbyte/db - newTag: 0.30.38-alpha + newTag: 0.30.39-alpha - name: airbyte/scheduler - newTag: 0.30.38-alpha + newTag: 0.30.39-alpha - name: airbyte/server - newTag: 0.30.38-alpha + newTag: 0.30.39-alpha - name: airbyte/webapp - newTag: 0.30.38-alpha + newTag: 0.30.39-alpha - name: airbyte/worker - newTag: 0.30.38-alpha + newTag: 0.30.39-alpha - name: temporalio/auto-setup newTag: 1.7.0 diff --git a/kube/overlays/stable/.env b/kube/overlays/stable/.env index 5a9c9b0692fb7..52191167c9a42 100644 --- a/kube/overlays/stable/.env +++ b/kube/overlays/stable/.env @@ -1,4 +1,4 @@ -AIRBYTE_VERSION=0.30.38-alpha +AIRBYTE_VERSION=0.30.39-alpha # Airbyte Internal Database, see https://docs.airbyte.io/operator-guides/configuring-airbyte-db DATABASE_HOST=airbyte-db-svc diff --git a/kube/overlays/stable/kustomization.yaml b/kube/overlays/stable/kustomization.yaml index 75a9b70f6bc6f..d4e1946e58f2d 100644 --- a/kube/overlays/stable/kustomization.yaml +++ b/kube/overlays/stable/kustomization.yaml @@ -8,15 +8,15 @@ bases: images: - name: airbyte/db - newTag: 0.30.38-alpha + newTag: 0.30.39-alpha - name: airbyte/scheduler - newTag: 0.30.38-alpha + newTag: 0.30.39-alpha - name: airbyte/server - newTag: 0.30.38-alpha + newTag: 0.30.39-alpha - name: airbyte/webapp - newTag: 0.30.38-alpha + newTag: 0.30.39-alpha - name: airbyte/worker - newTag: 0.30.38-alpha + newTag: 0.30.39-alpha - name: temporalio/auto-setup newTag: 1.7.0