diff --git a/gen.yaml b/gen.yaml
index 93ba212a..468e87d0 100644
--- a/gen.yaml
+++ b/gen.yaml
@@ -1,6 +1,6 @@
configVersion: 2.0.0
generation:
- sdkClassName: airbyte
+ sdkClassName: airbyte_api
usageSnippets:
optionalPropertyRendering: withExample
useClassNamesForArrayFields: true
@@ -11,7 +11,7 @@ generation:
auth:
oAuth2ClientCredentialsEnabled: false
python:
- version: 0.47.3
+ version: 0.48.0
author: Airbyte
clientServerStatusCodesAsErrors: true
description: Python Client SDK for Airbyte API
diff --git a/src/airbyte/__init__.py b/src/airbyte_api/__init__.py
similarity index 100%
rename from src/airbyte/__init__.py
rename to src/airbyte_api/__init__.py
diff --git a/src/airbyte/connections.py b/src/airbyte_api/connections.py
similarity index 96%
rename from src/airbyte/connections.py
rename to src/airbyte_api/connections.py
index 0e86e9fa..4ccecf55 100644
--- a/src/airbyte/connections.py
+++ b/src/airbyte_api/connections.py
@@ -1,8 +1,8 @@
"""Code generated by Speakeasy (https://speakeasyapi.dev). DO NOT EDIT."""
from .sdkconfiguration import SDKConfiguration
-from airbyte import utils
-from airbyte.models import errors, operations, shared
+from airbyte_api import utils
+from airbyte_api.models import errors, operations, shared
from typing import Optional
class Connections:
@@ -10,13 +10,11 @@ class Connections:
def __init__(self, sdk_config: SDKConfiguration) -> None:
self.sdk_configuration = sdk_config
-
-
-
+
def create_connection(self, request: shared.ConnectionCreateRequest) -> operations.CreateConnectionResponse:
r"""Create a connection"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/connections'
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, shared.ConnectionCreateRequest, "request", False, False, 'json')
@@ -26,17 +24,17 @@ def create_connection(self, request: shared.ConnectionCreateRequest) -> operatio
raise Exception('request body is required')
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('POST', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.CreateConnectionResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.ConnectionResponse])
@@ -48,27 +46,25 @@ def create_connection(self, request: shared.ConnectionCreateRequest) -> operatio
return res
-
-
def delete_connection(self, request: operations.DeleteConnectionRequest) -> operations.DeleteConnectionResponse:
r"""Delete a Connection"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.DeleteConnectionRequest, base_url, '/connections/{connectionId}', request)
headers = {}
headers['Accept'] = '*/*'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('DELETE', url, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.DeleteConnectionResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 204:
pass
elif http_res.status_code == 403 or http_res.status_code == 404 or http_res.status_code >= 400 and http_res.status_code < 500 or http_res.status_code >= 500 and http_res.status_code < 600:
@@ -76,27 +72,25 @@ def delete_connection(self, request: operations.DeleteConnectionRequest) -> oper
return res
-
-
def get_connection(self, request: operations.GetConnectionRequest) -> operations.GetConnectionResponse:
r"""Get Connection details"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.GetConnectionRequest, base_url, '/connections/{connectionId}', request)
headers = {}
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.GetConnectionResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.ConnectionResponse])
@@ -108,28 +102,26 @@ def get_connection(self, request: operations.GetConnectionRequest) -> operations
return res
-
-
def list_connections(self, request: operations.ListConnectionsRequest) -> operations.ListConnectionsResponse:
r"""List connections"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/connections'
headers = {}
query_params = utils.get_query_params(operations.ListConnectionsRequest, request)
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, params=query_params, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.ListConnectionsResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.ConnectionsResponse])
@@ -141,12 +133,10 @@ def list_connections(self, request: operations.ListConnectionsRequest) -> operat
return res
-
-
def patch_connection(self, request: operations.PatchConnectionRequest) -> operations.PatchConnectionResponse:
r"""Update Connection details"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.PatchConnectionRequest, base_url, '/connections/{connectionId}', request)
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, operations.PatchConnectionRequest, "connection_patch_request", False, False, 'json')
@@ -156,17 +146,17 @@ def patch_connection(self, request: operations.PatchConnectionRequest) -> operat
raise Exception('request body is required')
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('PATCH', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.PatchConnectionResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.ConnectionResponse])
@@ -177,5 +167,3 @@ def patch_connection(self, request: operations.PatchConnectionRequest) -> operat
raise errors.SDKError('API error occurred', http_res.status_code, http_res.text, http_res)
return res
-
-
\ No newline at end of file
diff --git a/src/airbyte/destinations.py b/src/airbyte_api/destinations.py
similarity index 96%
rename from src/airbyte/destinations.py
rename to src/airbyte_api/destinations.py
index 80d95337..f1dd662a 100644
--- a/src/airbyte/destinations.py
+++ b/src/airbyte_api/destinations.py
@@ -1,8 +1,8 @@
"""Code generated by Speakeasy (https://speakeasyapi.dev). DO NOT EDIT."""
from .sdkconfiguration import SDKConfiguration
-from airbyte import utils
-from airbyte.models import errors, operations, shared
+from airbyte_api import utils
+from airbyte_api.models import errors, operations, shared
from typing import Optional
class Destinations:
@@ -10,15 +10,13 @@ class Destinations:
def __init__(self, sdk_config: SDKConfiguration) -> None:
self.sdk_configuration = sdk_config
-
-
-
+
def create_destination(self, request: Optional[shared.DestinationCreateRequest]) -> operations.CreateDestinationResponse:
r"""Create a destination
Creates a destination given a name, workspace id, and a json blob containing the configuration for the source.
"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/destinations'
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, Optional[shared.DestinationCreateRequest], "request", False, True, 'json')
@@ -26,17 +24,17 @@ def create_destination(self, request: Optional[shared.DestinationCreateRequest])
headers['content-type'] = req_content_type
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('POST', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.CreateDestinationResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.DestinationResponse])
@@ -48,27 +46,25 @@ def create_destination(self, request: Optional[shared.DestinationCreateRequest])
return res
-
-
def delete_destination(self, request: operations.DeleteDestinationRequest) -> operations.DeleteDestinationResponse:
r"""Delete a Destination"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.DeleteDestinationRequest, base_url, '/destinations/{destinationId}', request)
headers = {}
headers['Accept'] = '*/*'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('DELETE', url, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.DeleteDestinationResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 204:
pass
elif http_res.status_code == 403 or http_res.status_code == 404 or http_res.status_code >= 400 and http_res.status_code < 500 or http_res.status_code >= 500 and http_res.status_code < 600:
@@ -76,27 +72,25 @@ def delete_destination(self, request: operations.DeleteDestinationRequest) -> op
return res
-
-
def get_destination(self, request: operations.GetDestinationRequest) -> operations.GetDestinationResponse:
r"""Get Destination details"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.GetDestinationRequest, base_url, '/destinations/{destinationId}', request)
headers = {}
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.GetDestinationResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.DestinationResponse])
@@ -108,28 +102,26 @@ def get_destination(self, request: operations.GetDestinationRequest) -> operatio
return res
-
-
def list_destinations(self, request: operations.ListDestinationsRequest) -> operations.ListDestinationsResponse:
r"""List destinations"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/destinations'
headers = {}
query_params = utils.get_query_params(operations.ListDestinationsRequest, request)
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, params=query_params, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.ListDestinationsResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.DestinationsResponse])
@@ -141,12 +133,10 @@ def list_destinations(self, request: operations.ListDestinationsRequest) -> oper
return res
-
-
def patch_destination(self, request: operations.PatchDestinationRequest) -> operations.PatchDestinationResponse:
r"""Update a Destination"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.PatchDestinationRequest, base_url, '/destinations/{destinationId}', request)
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, operations.PatchDestinationRequest, "destination_patch_request", False, True, 'json')
@@ -154,17 +144,17 @@ def patch_destination(self, request: operations.PatchDestinationRequest) -> oper
headers['content-type'] = req_content_type
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('PATCH', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.PatchDestinationResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.DestinationResponse])
@@ -176,12 +166,10 @@ def patch_destination(self, request: operations.PatchDestinationRequest) -> oper
return res
-
-
def put_destination(self, request: operations.PutDestinationRequest) -> operations.PutDestinationResponse:
r"""Update a Destination and fully overwrite it"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.PutDestinationRequest, base_url, '/destinations/{destinationId}', request)
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, operations.PutDestinationRequest, "destination_put_request", False, True, 'json')
@@ -189,17 +177,17 @@ def put_destination(self, request: operations.PutDestinationRequest) -> operatio
headers['content-type'] = req_content_type
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('PUT', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.PutDestinationResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.DestinationResponse])
@@ -210,5 +198,3 @@ def put_destination(self, request: operations.PutDestinationRequest) -> operatio
raise errors.SDKError('API error occurred', http_res.status_code, http_res.text, http_res)
return res
-
-
\ No newline at end of file
diff --git a/src/airbyte/jobs.py b/src/airbyte_api/jobs.py
similarity index 96%
rename from src/airbyte/jobs.py
rename to src/airbyte_api/jobs.py
index 3fb96f73..48e39ae5 100644
--- a/src/airbyte/jobs.py
+++ b/src/airbyte_api/jobs.py
@@ -1,8 +1,8 @@
"""Code generated by Speakeasy (https://speakeasyapi.dev). DO NOT EDIT."""
from .sdkconfiguration import SDKConfiguration
-from airbyte import utils
-from airbyte.models import errors, operations, shared
+from airbyte_api import utils
+from airbyte_api.models import errors, operations, shared
from typing import Optional
class Jobs:
@@ -10,28 +10,26 @@ class Jobs:
def __init__(self, sdk_config: SDKConfiguration) -> None:
self.sdk_configuration = sdk_config
-
-
-
+
def cancel_job(self, request: operations.CancelJobRequest) -> operations.CancelJobResponse:
r"""Cancel a running Job"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.CancelJobRequest, base_url, '/jobs/{jobId}', request)
headers = {}
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('DELETE', url, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.CancelJobResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.JobResponse])
@@ -43,12 +41,10 @@ def cancel_job(self, request: operations.CancelJobRequest) -> operations.CancelJ
return res
-
-
def create_job(self, request: shared.JobCreateRequest) -> operations.CreateJobResponse:
r"""Trigger a sync or reset job of a connection"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/jobs'
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, shared.JobCreateRequest, "request", False, False, 'json')
@@ -58,17 +54,17 @@ def create_job(self, request: shared.JobCreateRequest) -> operations.CreateJobRe
raise Exception('request body is required')
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('POST', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.CreateJobResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.JobResponse])
@@ -80,27 +76,25 @@ def create_job(self, request: shared.JobCreateRequest) -> operations.CreateJobRe
return res
-
-
def get_job(self, request: operations.GetJobRequest) -> operations.GetJobResponse:
r"""Get Job status and details"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.GetJobRequest, base_url, '/jobs/{jobId}', request)
headers = {}
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.GetJobResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.JobResponse])
@@ -112,28 +106,26 @@ def get_job(self, request: operations.GetJobRequest) -> operations.GetJobRespons
return res
-
-
def list_jobs(self, request: operations.ListJobsRequest) -> operations.ListJobsResponse:
r"""List Jobs by sync type"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/jobs'
headers = {}
query_params = utils.get_query_params(operations.ListJobsRequest, request)
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, params=query_params, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.ListJobsResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.JobsResponse])
@@ -144,5 +136,3 @@ def list_jobs(self, request: operations.ListJobsRequest) -> operations.ListJobsR
raise errors.SDKError('API error occurred', http_res.status_code, http_res.text, http_res)
return res
-
-
\ No newline at end of file
diff --git a/src/airbyte/models/__init__.py b/src/airbyte_api/models/__init__.py
similarity index 100%
rename from src/airbyte/models/__init__.py
rename to src/airbyte_api/models/__init__.py
diff --git a/src/airbyte/models/errors/__init__.py b/src/airbyte_api/models/errors/__init__.py
similarity index 100%
rename from src/airbyte/models/errors/__init__.py
rename to src/airbyte_api/models/errors/__init__.py
diff --git a/src/airbyte/models/errors/sdkerror.py b/src/airbyte_api/models/errors/sdkerror.py
similarity index 100%
rename from src/airbyte/models/errors/sdkerror.py
rename to src/airbyte_api/models/errors/sdkerror.py
diff --git a/src/airbyte/models/operations/__init__.py b/src/airbyte_api/models/operations/__init__.py
similarity index 100%
rename from src/airbyte/models/operations/__init__.py
rename to src/airbyte_api/models/operations/__init__.py
diff --git a/src/airbyte/models/operations/canceljob.py b/src/airbyte_api/models/operations/canceljob.py
similarity index 100%
rename from src/airbyte/models/operations/canceljob.py
rename to src/airbyte_api/models/operations/canceljob.py
diff --git a/src/airbyte/models/operations/createconnection.py b/src/airbyte_api/models/operations/createconnection.py
similarity index 100%
rename from src/airbyte/models/operations/createconnection.py
rename to src/airbyte_api/models/operations/createconnection.py
diff --git a/src/airbyte/models/operations/createdestination.py b/src/airbyte_api/models/operations/createdestination.py
similarity index 100%
rename from src/airbyte/models/operations/createdestination.py
rename to src/airbyte_api/models/operations/createdestination.py
diff --git a/src/airbyte/models/operations/createjob.py b/src/airbyte_api/models/operations/createjob.py
similarity index 100%
rename from src/airbyte/models/operations/createjob.py
rename to src/airbyte_api/models/operations/createjob.py
diff --git a/src/airbyte/models/operations/createorupdateworkspaceoauthcredentials.py b/src/airbyte_api/models/operations/createorupdateworkspaceoauthcredentials.py
similarity index 100%
rename from src/airbyte/models/operations/createorupdateworkspaceoauthcredentials.py
rename to src/airbyte_api/models/operations/createorupdateworkspaceoauthcredentials.py
diff --git a/src/airbyte/models/operations/createsource.py b/src/airbyte_api/models/operations/createsource.py
similarity index 100%
rename from src/airbyte/models/operations/createsource.py
rename to src/airbyte_api/models/operations/createsource.py
diff --git a/src/airbyte/models/operations/createworkspace.py b/src/airbyte_api/models/operations/createworkspace.py
similarity index 100%
rename from src/airbyte/models/operations/createworkspace.py
rename to src/airbyte_api/models/operations/createworkspace.py
diff --git a/src/airbyte/models/operations/deleteconnection.py b/src/airbyte_api/models/operations/deleteconnection.py
similarity index 100%
rename from src/airbyte/models/operations/deleteconnection.py
rename to src/airbyte_api/models/operations/deleteconnection.py
diff --git a/src/airbyte/models/operations/deletedestination.py b/src/airbyte_api/models/operations/deletedestination.py
similarity index 100%
rename from src/airbyte/models/operations/deletedestination.py
rename to src/airbyte_api/models/operations/deletedestination.py
diff --git a/src/airbyte/models/operations/deletesource.py b/src/airbyte_api/models/operations/deletesource.py
similarity index 100%
rename from src/airbyte/models/operations/deletesource.py
rename to src/airbyte_api/models/operations/deletesource.py
diff --git a/src/airbyte/models/operations/deleteworkspace.py b/src/airbyte_api/models/operations/deleteworkspace.py
similarity index 100%
rename from src/airbyte/models/operations/deleteworkspace.py
rename to src/airbyte_api/models/operations/deleteworkspace.py
diff --git a/src/airbyte/models/operations/getconnection.py b/src/airbyte_api/models/operations/getconnection.py
similarity index 100%
rename from src/airbyte/models/operations/getconnection.py
rename to src/airbyte_api/models/operations/getconnection.py
diff --git a/src/airbyte/models/operations/getdestination.py b/src/airbyte_api/models/operations/getdestination.py
similarity index 100%
rename from src/airbyte/models/operations/getdestination.py
rename to src/airbyte_api/models/operations/getdestination.py
diff --git a/src/airbyte/models/operations/getjob.py b/src/airbyte_api/models/operations/getjob.py
similarity index 100%
rename from src/airbyte/models/operations/getjob.py
rename to src/airbyte_api/models/operations/getjob.py
diff --git a/src/airbyte/models/operations/getsource.py b/src/airbyte_api/models/operations/getsource.py
similarity index 100%
rename from src/airbyte/models/operations/getsource.py
rename to src/airbyte_api/models/operations/getsource.py
diff --git a/src/airbyte/models/operations/getstreamproperties.py b/src/airbyte_api/models/operations/getstreamproperties.py
similarity index 100%
rename from src/airbyte/models/operations/getstreamproperties.py
rename to src/airbyte_api/models/operations/getstreamproperties.py
diff --git a/src/airbyte/models/operations/getworkspace.py b/src/airbyte_api/models/operations/getworkspace.py
similarity index 100%
rename from src/airbyte/models/operations/getworkspace.py
rename to src/airbyte_api/models/operations/getworkspace.py
diff --git a/src/airbyte/models/operations/initiateoauth.py b/src/airbyte_api/models/operations/initiateoauth.py
similarity index 100%
rename from src/airbyte/models/operations/initiateoauth.py
rename to src/airbyte_api/models/operations/initiateoauth.py
diff --git a/src/airbyte/models/operations/listconnections.py b/src/airbyte_api/models/operations/listconnections.py
similarity index 100%
rename from src/airbyte/models/operations/listconnections.py
rename to src/airbyte_api/models/operations/listconnections.py
diff --git a/src/airbyte/models/operations/listdestinations.py b/src/airbyte_api/models/operations/listdestinations.py
similarity index 100%
rename from src/airbyte/models/operations/listdestinations.py
rename to src/airbyte_api/models/operations/listdestinations.py
diff --git a/src/airbyte/models/operations/listjobs.py b/src/airbyte_api/models/operations/listjobs.py
similarity index 100%
rename from src/airbyte/models/operations/listjobs.py
rename to src/airbyte_api/models/operations/listjobs.py
diff --git a/src/airbyte/models/operations/listsources.py b/src/airbyte_api/models/operations/listsources.py
similarity index 100%
rename from src/airbyte/models/operations/listsources.py
rename to src/airbyte_api/models/operations/listsources.py
diff --git a/src/airbyte/models/operations/listworkspaces.py b/src/airbyte_api/models/operations/listworkspaces.py
similarity index 100%
rename from src/airbyte/models/operations/listworkspaces.py
rename to src/airbyte_api/models/operations/listworkspaces.py
diff --git a/src/airbyte/models/operations/patchconnection.py b/src/airbyte_api/models/operations/patchconnection.py
similarity index 100%
rename from src/airbyte/models/operations/patchconnection.py
rename to src/airbyte_api/models/operations/patchconnection.py
diff --git a/src/airbyte/models/operations/patchdestination.py b/src/airbyte_api/models/operations/patchdestination.py
similarity index 100%
rename from src/airbyte/models/operations/patchdestination.py
rename to src/airbyte_api/models/operations/patchdestination.py
diff --git a/src/airbyte/models/operations/patchsource.py b/src/airbyte_api/models/operations/patchsource.py
similarity index 100%
rename from src/airbyte/models/operations/patchsource.py
rename to src/airbyte_api/models/operations/patchsource.py
diff --git a/src/airbyte/models/operations/putdestination.py b/src/airbyte_api/models/operations/putdestination.py
similarity index 100%
rename from src/airbyte/models/operations/putdestination.py
rename to src/airbyte_api/models/operations/putdestination.py
diff --git a/src/airbyte/models/operations/putsource.py b/src/airbyte_api/models/operations/putsource.py
similarity index 100%
rename from src/airbyte/models/operations/putsource.py
rename to src/airbyte_api/models/operations/putsource.py
diff --git a/src/airbyte/models/operations/updateworkspace.py b/src/airbyte_api/models/operations/updateworkspace.py
similarity index 100%
rename from src/airbyte/models/operations/updateworkspace.py
rename to src/airbyte_api/models/operations/updateworkspace.py
diff --git a/src/airbyte/models/shared/__init__.py b/src/airbyte_api/models/shared/__init__.py
similarity index 100%
rename from src/airbyte/models/shared/__init__.py
rename to src/airbyte_api/models/shared/__init__.py
diff --git a/src/airbyte/models/shared/actortypeenum.py b/src/airbyte_api/models/shared/actortypeenum.py
similarity index 100%
rename from src/airbyte/models/shared/actortypeenum.py
rename to src/airbyte_api/models/shared/actortypeenum.py
diff --git a/src/airbyte/models/shared/airtable.py b/src/airbyte_api/models/shared/airtable.py
similarity index 96%
rename from src/airbyte/models/shared/airtable.py
rename to src/airbyte_api/models/shared/airtable.py
index 5617e818..e68123e0 100644
--- a/src/airbyte/models/shared/airtable.py
+++ b/src/airbyte_api/models/shared/airtable.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class Credentials:
r"""The client ID of the Airtable developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The client secret the Airtable developer application."""
-
+
@@ -22,5 +22,3 @@ class Credentials:
@dataclasses.dataclass
class Airtable:
credentials: Optional[Credentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/amazon_ads.py b/src/airbyte_api/models/shared/amazon_ads.py
similarity index 97%
rename from src/airbyte/models/shared/amazon_ads.py
rename to src/airbyte_api/models/shared/amazon_ads.py
index b8ff24b8..d31239f7 100644
--- a/src/airbyte/models/shared/amazon_ads.py
+++ b/src/airbyte_api/models/shared/amazon_ads.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,5 +14,3 @@ class AmazonAds:
r"""The client ID of your Amazon Ads developer application. See the docs for more information."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The client secret of your Amazon Ads developer application. See the docs for more information."""
-
-
diff --git a/src/airbyte/models/shared/amazon_seller_partner.py b/src/airbyte_api/models/shared/amazon_seller_partner.py
similarity index 96%
rename from src/airbyte/models/shared/amazon_seller_partner.py
rename to src/airbyte_api/models/shared/amazon_seller_partner.py
index 6397acf4..a68b74c0 100644
--- a/src/airbyte/models/shared/amazon_seller_partner.py
+++ b/src/airbyte_api/models/shared/amazon_seller_partner.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,5 +14,3 @@ class AmazonSellerPartner:
r"""Your Login with Amazon Client ID."""
lwa_client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('lwa_client_secret'), 'exclude': lambda f: f is None }})
r"""Your Login with Amazon Client Secret."""
-
-
diff --git a/src/airbyte/models/shared/asana.py b/src/airbyte_api/models/shared/asana.py
similarity index 96%
rename from src/airbyte/models/shared/asana.py
rename to src/airbyte_api/models/shared/asana.py
index a1a26c67..3ad65abf 100644
--- a/src/airbyte/models/shared/asana.py
+++ b/src/airbyte_api/models/shared/asana.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -12,7 +12,7 @@
class AsanaCredentials:
client_id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_id'), 'exclude': lambda f: f is None }})
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
-
+
@@ -20,5 +20,3 @@ class AsanaCredentials:
@dataclasses.dataclass
class Asana:
credentials: Optional[AsanaCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/bing_ads.py b/src/airbyte_api/models/shared/bing_ads.py
similarity index 96%
rename from src/airbyte/models/shared/bing_ads.py
rename to src/airbyte_api/models/shared/bing_ads.py
index fbfefc9f..cc8b0241 100644
--- a/src/airbyte/models/shared/bing_ads.py
+++ b/src/airbyte_api/models/shared/bing_ads.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,5 +14,3 @@ class BingAds:
r"""The Client ID of your Microsoft Advertising developer application."""
client_secret: Optional[str] = dataclasses.field(default='', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Microsoft Advertising developer application."""
-
-
diff --git a/src/airbyte/models/shared/connectioncreaterequest.py b/src/airbyte_api/models/shared/connectioncreaterequest.py
similarity index 99%
rename from src/airbyte/models/shared/connectioncreaterequest.py
rename to src/airbyte_api/models/shared/connectioncreaterequest.py
index 62adbdb9..a2b83b6d 100644
--- a/src/airbyte/models/shared/connectioncreaterequest.py
+++ b/src/airbyte_api/models/shared/connectioncreaterequest.py
@@ -8,7 +8,7 @@
from .namespacedefinitionenum import NamespaceDefinitionEnum
from .nonbreakingschemaupdatesbehaviorenum import NonBreakingSchemaUpdatesBehaviorEnum
from .streamconfigurations import StreamConfigurations
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -34,5 +34,3 @@ class ConnectionCreateRequest:
schedule: Optional[ConnectionSchedule] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('schedule'), 'exclude': lambda f: f is None }})
r"""schedule for when the the connection should run, per the schedule type"""
status: Optional[ConnectionStatusEnum] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('status'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/connectionpatchrequest.py b/src/airbyte_api/models/shared/connectionpatchrequest.py
similarity index 99%
rename from src/airbyte/models/shared/connectionpatchrequest.py
rename to src/airbyte_api/models/shared/connectionpatchrequest.py
index 0cd4b8d1..69b5121d 100644
--- a/src/airbyte/models/shared/connectionpatchrequest.py
+++ b/src/airbyte_api/models/shared/connectionpatchrequest.py
@@ -8,7 +8,7 @@
from .namespacedefinitionenumnodefault import NamespaceDefinitionEnumNoDefault
from .nonbreakingschemaupdatesbehaviorenumnodefault import NonBreakingSchemaUpdatesBehaviorEnumNoDefault
from .streamconfigurations import StreamConfigurations
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -32,5 +32,3 @@ class ConnectionPatchRequest:
schedule: Optional[ConnectionSchedule] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('schedule'), 'exclude': lambda f: f is None }})
r"""schedule for when the the connection should run, per the schedule type"""
status: Optional[ConnectionStatusEnum] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('status'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/connectionresponse.py b/src/airbyte_api/models/shared/connectionresponse.py
similarity index 99%
rename from src/airbyte/models/shared/connectionresponse.py
rename to src/airbyte_api/models/shared/connectionresponse.py
index a4f27b2b..8ea8bcc6 100644
--- a/src/airbyte/models/shared/connectionresponse.py
+++ b/src/airbyte_api/models/shared/connectionresponse.py
@@ -8,7 +8,7 @@
from .namespacedefinitionenum import NamespaceDefinitionEnum
from .nonbreakingschemaupdatesbehaviorenum import NonBreakingSchemaUpdatesBehaviorEnum
from .streamconfigurations import StreamConfigurations
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -34,5 +34,3 @@ class ConnectionResponse:
non_breaking_schema_updates_behavior: Optional[NonBreakingSchemaUpdatesBehaviorEnum] = dataclasses.field(default=NonBreakingSchemaUpdatesBehaviorEnum.IGNORE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('nonBreakingSchemaUpdatesBehavior'), 'exclude': lambda f: f is None }})
r"""Set how Airbyte handles syncs when it detects a non-breaking schema change in the source"""
prefix: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('prefix'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/connectionschedule.py b/src/airbyte_api/models/shared/connectionschedule.py
similarity index 96%
rename from src/airbyte/models/shared/connectionschedule.py
rename to src/airbyte_api/models/shared/connectionschedule.py
index bbc1a11b..fb630672 100644
--- a/src/airbyte/models/shared/connectionschedule.py
+++ b/src/airbyte_api/models/shared/connectionschedule.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .scheduletypeenum import ScheduleTypeEnum
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,5 +14,3 @@ class ConnectionSchedule:
r"""schedule for when the the connection should run, per the schedule type"""
schedule_type: ScheduleTypeEnum = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('scheduleType') }})
cron_expression: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cronExpression'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/connectionscheduleresponse.py b/src/airbyte_api/models/shared/connectionscheduleresponse.py
similarity index 96%
rename from src/airbyte/models/shared/connectionscheduleresponse.py
rename to src/airbyte_api/models/shared/connectionscheduleresponse.py
index 8948bb44..dad4a824 100644
--- a/src/airbyte/models/shared/connectionscheduleresponse.py
+++ b/src/airbyte_api/models/shared/connectionscheduleresponse.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .scheduletypewithbasicenum import ScheduleTypeWithBasicEnum
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -15,5 +15,3 @@ class ConnectionScheduleResponse:
schedule_type: ScheduleTypeWithBasicEnum = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('scheduleType') }})
basic_timing: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('basicTiming'), 'exclude': lambda f: f is None }})
cron_expression: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cronExpression'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/connectionsresponse.py b/src/airbyte_api/models/shared/connectionsresponse.py
similarity index 96%
rename from src/airbyte/models/shared/connectionsresponse.py
rename to src/airbyte_api/models/shared/connectionsresponse.py
index 5103b11c..40c3899a 100644
--- a/src/airbyte/models/shared/connectionsresponse.py
+++ b/src/airbyte_api/models/shared/connectionsresponse.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .connectionresponse import ConnectionResponse
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import List, Optional
@@ -14,5 +14,3 @@ class ConnectionsResponse:
data: List[ConnectionResponse] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('data') }})
next: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('next'), 'exclude': lambda f: f is None }})
previous: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('previous'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/connectionstatusenum.py b/src/airbyte_api/models/shared/connectionstatusenum.py
similarity index 100%
rename from src/airbyte/models/shared/connectionstatusenum.py
rename to src/airbyte_api/models/shared/connectionstatusenum.py
diff --git a/src/airbyte/models/shared/connectionsyncmodeenum.py b/src/airbyte_api/models/shared/connectionsyncmodeenum.py
similarity index 100%
rename from src/airbyte/models/shared/connectionsyncmodeenum.py
rename to src/airbyte_api/models/shared/connectionsyncmodeenum.py
diff --git a/src/airbyte/models/shared/destination_astra.py b/src/airbyte_api/models/shared/destination_astra.py
similarity index 99%
rename from src/airbyte/models/shared/destination_astra.py
rename to src/airbyte_api/models/shared/destination_astra.py
index 43814913..b21fd93a 100644
--- a/src/airbyte/models/shared/destination_astra.py
+++ b/src/airbyte_api/models/shared/destination_astra.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional, Union
@@ -26,7 +26,7 @@ class OpenAICompatible:
MODE: Final[Optional[DestinationAstraSchemasEmbeddingEmbeddingMode]] = dataclasses.field(default=DestinationAstraSchemasEmbeddingEmbeddingMode.OPENAI_COMPATIBLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
model_name: Optional[str] = dataclasses.field(default='text-embedding-ada-002', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('model_name'), 'exclude': lambda f: f is None }})
r"""The name of the model to use for embedding"""
-
+
class DestinationAstraSchemasEmbeddingMode(str, Enum):
@@ -44,7 +44,7 @@ class AzureOpenAI:
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
r"""The API key for your Azure OpenAI resource. You can find this in the Azure portal under your Azure OpenAI resource"""
MODE: Final[Optional[DestinationAstraSchemasEmbeddingMode]] = dataclasses.field(default=DestinationAstraSchemasEmbeddingMode.AZURE_OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationAstraSchemasMode(str, Enum):
@@ -56,7 +56,7 @@ class DestinationAstraSchemasMode(str, Enum):
class Fake:
r"""Use a fake embedding made out of random vectors with 1536 embedding dimensions. This is useful for testing the data pipeline without incurring any costs."""
MODE: Final[Optional[DestinationAstraSchemasMode]] = dataclasses.field(default=DestinationAstraSchemasMode.FAKE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationAstraMode(str, Enum):
@@ -69,7 +69,7 @@ class Cohere:
r"""Use the Cohere API to embed text."""
cohere_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cohere_key') }})
MODE: Final[Optional[DestinationAstraMode]] = dataclasses.field(default=DestinationAstraMode.COHERE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationAstraSchemasEmbeddingEmbedding1Mode(str, Enum):
@@ -82,7 +82,7 @@ class OpenAI:
r"""Use the OpenAI API to embed text. This option is using the text-embedding-ada-002 model with 1536 embedding dimensions."""
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
MODE: Final[Optional[DestinationAstraSchemasEmbeddingEmbedding1Mode]] = dataclasses.field(default=DestinationAstraSchemasEmbeddingEmbedding1Mode.OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
@@ -98,7 +98,7 @@ class Indexing:
r"""Keyspaces (or Namespaces) serve as containers for organizing data within a database. You can create a new keyspace uisng the Data Explorer tab in the Astra UI. The keyspace default_keyspace is created for you when you create a Vector Database in Astra DB."""
collection: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('collection') }})
r"""Collections hold data. They are analagous to tables in traditional Cassandra terminology. This tool will create the collection with the provided name automatically if it does not already exist. Alternatively, you can create one thorugh the Data Explorer tab in the Astra UI."""
-
+
@@ -109,7 +109,7 @@ class FieldNameMappingConfigModel:
r"""The field name in the source"""
to_field: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('to_field') }})
r"""The field name to use in the destination"""
-
+
class DestinationAstraLanguage(str, Enum):
@@ -142,7 +142,7 @@ class ByProgrammingLanguage:
language: DestinationAstraLanguage = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('language') }})
r"""Split code in suitable places based on the programming language"""
MODE: Final[Optional[DestinationAstraSchemasProcessingTextSplitterTextSplitterMode]] = dataclasses.field(default=DestinationAstraSchemasProcessingTextSplitterTextSplitterMode.CODE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationAstraSchemasProcessingTextSplitterMode(str, Enum):
@@ -156,7 +156,7 @@ class ByMarkdownHeader:
MODE: Final[Optional[DestinationAstraSchemasProcessingTextSplitterMode]] = dataclasses.field(default=DestinationAstraSchemasProcessingTextSplitterMode.MARKDOWN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
split_level: Optional[int] = dataclasses.field(default=1, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('split_level'), 'exclude': lambda f: f is None }})
r"""Level of markdown headers to split text fields by. Headings down to the specified level will be used as split points"""
-
+
class DestinationAstraSchemasProcessingMode(str, Enum):
@@ -172,7 +172,7 @@ class BySeparator:
MODE: Final[Optional[DestinationAstraSchemasProcessingMode]] = dataclasses.field(default=DestinationAstraSchemasProcessingMode.SEPARATOR, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
separators: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('separators'), 'exclude': lambda f: f is None }})
r"""List of separator strings to split text fields by. The separator itself needs to be wrapped in double quotes, e.g. to split by the dot character, use \\".\\". To split by a newline, use \\"\n\\"."""
-
+
@@ -191,7 +191,7 @@ class ProcessingConfigModel:
r"""List of fields in the record that should be used to calculate the embedding. The field list is applied to all streams in the same way and non-existing fields are ignored. If none are defined, all fields are considered text fields. When specifying text fields, you can access nested fields in the record by using dot notation, e.g. `user.name` will access the `name` field in the `user` object. It's also possible to use wildcards to access all fields in an object, e.g. `users.*.name` will access all `names` fields in all entries of the `users` array."""
text_splitter: Optional[Union[BySeparator, ByMarkdownHeader, ByProgrammingLanguage]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('text_splitter'), 'exclude': lambda f: f is None }})
r"""Split text fields into chunks based on the specified method."""
-
+
@@ -217,5 +217,3 @@ class DestinationAstra:
DESTINATION_TYPE: Final[Astra] = dataclasses.field(default=Astra.ASTRA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
omit_raw_text: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('omit_raw_text'), 'exclude': lambda f: f is None }})
r"""Do not store the text that gets embedded along with the vector and the metadata in the destination. If set to true, only the vector and the metadata will be stored - in this case raw text for LLM use cases needs to be retrieved from another source."""
-
-
diff --git a/src/airbyte/models/shared/destination_aws_datalake.py b/src/airbyte_api/models/shared/destination_aws_datalake.py
similarity index 99%
rename from src/airbyte/models/shared/destination_aws_datalake.py
rename to src/airbyte_api/models/shared/destination_aws_datalake.py
index 095771af..058e8437 100644
--- a/src/airbyte/models/shared/destination_aws_datalake.py
+++ b/src/airbyte_api/models/shared/destination_aws_datalake.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -21,7 +21,7 @@ class IAMUser:
r"""Secret Access Key"""
CREDENTIALS_TITLE: Final[Optional[DestinationAwsDatalakeCredentialsTitle]] = dataclasses.field(default=DestinationAwsDatalakeCredentialsTitle.IAM_USER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials_title'), 'exclude': lambda f: f is None }})
r"""Name of the credentials"""
-
+
class CredentialsTitle(str, Enum):
@@ -36,7 +36,7 @@ class IAMRole:
r"""Will assume this role to write data to s3"""
CREDENTIALS_TITLE: Final[Optional[CredentialsTitle]] = dataclasses.field(default=CredentialsTitle.IAM_ROLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials_title'), 'exclude': lambda f: f is None }})
r"""Name of the credentials"""
-
+
class AwsDatalake(str, Enum):
@@ -59,7 +59,7 @@ class ParquetColumnarStorage:
compression_codec: Optional[DestinationAwsDatalakeCompressionCodecOptional] = dataclasses.field(default=DestinationAwsDatalakeCompressionCodecOptional.SNAPPY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_codec'), 'exclude': lambda f: f is None }})
r"""The compression algorithm used to compress data."""
format_type: Optional[DestinationAwsDatalakeFormatTypeWildcard] = dataclasses.field(default=DestinationAwsDatalakeFormatTypeWildcard.PARQUET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type'), 'exclude': lambda f: f is None }})
-
+
class CompressionCodecOptional(str, Enum):
@@ -77,7 +77,7 @@ class JSONLinesNewlineDelimitedJSON:
compression_codec: Optional[CompressionCodecOptional] = dataclasses.field(default=CompressionCodecOptional.UNCOMPRESSED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_codec'), 'exclude': lambda f: f is None }})
r"""The compression algorithm used to compress data."""
format_type: Optional[FormatTypeWildcard] = dataclasses.field(default=FormatTypeWildcard.JSONL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type'), 'exclude': lambda f: f is None }})
-
+
class ChooseHowToPartitionData(str, Enum):
@@ -156,5 +156,3 @@ class DestinationAwsDatalake:
r"""Partition data by cursor fields when a cursor field is a date"""
region: Optional[S3BucketRegion] = dataclasses.field(default=S3BucketRegion.UNKNOWN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('region'), 'exclude': lambda f: f is None }})
r"""The region of the S3 bucket. See here for all region codes."""
-
-
diff --git a/src/airbyte/models/shared/destination_azure_blob_storage.py b/src/airbyte_api/models/shared/destination_azure_blob_storage.py
similarity index 99%
rename from src/airbyte/models/shared/destination_azure_blob_storage.py
rename to src/airbyte_api/models/shared/destination_azure_blob_storage.py
index 7a5b8e5a..46c5df32 100644
--- a/src/airbyte/models/shared/destination_azure_blob_storage.py
+++ b/src/airbyte_api/models/shared/destination_azure_blob_storage.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -18,7 +18,7 @@ class DestinationAzureBlobStorageFormatType(str, Enum):
@dataclasses.dataclass
class DestinationAzureBlobStorageJSONLinesNewlineDelimitedJSON:
FORMAT_TYPE: Final[DestinationAzureBlobStorageFormatType] = dataclasses.field(default=DestinationAzureBlobStorageFormatType.JSONL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type') }})
-
+
class NormalizationFlattening(str, Enum):
@@ -36,7 +36,7 @@ class CSVCommaSeparatedValues:
flattening: Optional[NormalizationFlattening] = dataclasses.field(default=NormalizationFlattening.NO_FLATTENING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('flattening'), 'exclude': lambda f: f is None }})
r"""Whether the input json data should be normalized (flattened) in the output CSV. Please refer to docs for details."""
FORMAT_TYPE: Final[FormatType] = dataclasses.field(default=FormatType.CSV, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type') }})
-
+
@@ -58,5 +58,3 @@ class DestinationAzureBlobStorage:
azure_blob_storage_spill_size: Optional[int] = dataclasses.field(default=500, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('azure_blob_storage_spill_size'), 'exclude': lambda f: f is None }})
r"""The amount of megabytes after which the connector should spill the records in a new blob object. Make sure to configure size greater than individual records. Enter 0 if not applicable"""
DESTINATION_TYPE: Final[AzureBlobStorage] = dataclasses.field(default=AzureBlobStorage.AZURE_BLOB_STORAGE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
-
-
diff --git a/src/airbyte/models/shared/destination_bigquery.py b/src/airbyte_api/models/shared/destination_bigquery.py
similarity index 99%
rename from src/airbyte/models/shared/destination_bigquery.py
rename to src/airbyte_api/models/shared/destination_bigquery.py
index 107a951d..368d46ab 100644
--- a/src/airbyte/models/shared/destination_bigquery.py
+++ b/src/airbyte_api/models/shared/destination_bigquery.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -66,7 +66,7 @@ class DestinationBigqueryMethod(str, Enum):
class StandardInserts:
r"""(not recommended) Direct loading using SQL INSERT statements. This method is extremely inefficient and provided only for quick testing. In all other cases, you should use GCS staging."""
METHOD: Final[DestinationBigqueryMethod] = dataclasses.field(default=DestinationBigqueryMethod.STANDARD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class DestinationBigqueryCredentialType(str, Enum):
@@ -81,7 +81,7 @@ class DestinationBigqueryHMACKey:
hmac_key_secret: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('hmac_key_secret') }})
r"""The corresponding secret for the access ID. It is a 40-character base-64 encoded string."""
CREDENTIAL_TYPE: Final[DestinationBigqueryCredentialType] = dataclasses.field(default=DestinationBigqueryCredentialType.HMAC_KEY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credential_type') }})
-
+
class GCSTmpFilesAfterwardProcessing(str, Enum):
@@ -106,7 +106,7 @@ class GCSStaging:
keep_files_in_gcs_bucket: Optional[GCSTmpFilesAfterwardProcessing] = dataclasses.field(default=GCSTmpFilesAfterwardProcessing.DELETE_ALL_TMP_FILES_FROM_GCS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('keep_files_in_gcs-bucket'), 'exclude': lambda f: f is None }})
r"""This upload method is supposed to temporary store records in GCS bucket. By this select you can chose if these records should be removed from GCS when migration has finished. The default \\"Delete all tmp files from GCS\\" value is used if not set explicitly."""
METHOD: Final[Method] = dataclasses.field(default=Method.GCS_STAGING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class TransformationQueryRunType(str, Enum):
@@ -137,5 +137,3 @@ class DestinationBigquery:
r"""The dataset to write raw tables into (default: airbyte_internal)"""
transformation_priority: Optional[TransformationQueryRunType] = dataclasses.field(default=TransformationQueryRunType.INTERACTIVE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('transformation_priority'), 'exclude': lambda f: f is None }})
r"""Interactive run type means that the query is executed as soon as possible, and these queries count towards concurrent rate limit and daily limit. Read more about interactive run type here. Batch queries are queued and started as soon as idle resources are available in the BigQuery shared resource pool, which usually occurs within a few minutes. Batch queries don’t count towards your concurrent rate limit. Read more about batch queries here. The default \\"interactive\\" value is used if not set explicitly."""
-
-
diff --git a/src/airbyte/models/shared/destination_clickhouse.py b/src/airbyte_api/models/shared/destination_clickhouse.py
similarity index 99%
rename from src/airbyte/models/shared/destination_clickhouse.py
rename to src/airbyte_api/models/shared/destination_clickhouse.py
index 355c0514..6b63095e 100644
--- a/src/airbyte/models/shared/destination_clickhouse.py
+++ b/src/airbyte_api/models/shared/destination_clickhouse.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -28,7 +28,7 @@ class PasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationClickhouseTunnelMethod(str, Enum):
@@ -49,7 +49,7 @@ class SSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class TunnelMethod(str, Enum):
@@ -62,7 +62,7 @@ class TunnelMethod(str, Enum):
class NoTunnel:
TUNNEL_METHOD: Final[TunnelMethod] = dataclasses.field(default=TunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -84,5 +84,3 @@ class DestinationClickhouse:
r"""HTTP port of the database."""
tunnel_method: Optional[Union[NoTunnel, SSHKeyAuthentication, PasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/destination_convex.py b/src/airbyte_api/models/shared/destination_convex.py
similarity index 96%
rename from src/airbyte/models/shared/destination_convex.py
rename to src/airbyte_api/models/shared/destination_convex.py
index d49a5029..4e1960c0 100644
--- a/src/airbyte/models/shared/destination_convex.py
+++ b/src/airbyte_api/models/shared/destination_convex.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class DestinationConvex:
deployment_url: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('deployment_url') }})
r"""URL of the Convex deployment that is the destination"""
DESTINATION_TYPE: Final[Convex] = dataclasses.field(default=Convex.CONVEX, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
-
-
diff --git a/src/airbyte/models/shared/destination_cumulio.py b/src/airbyte_api/models/shared/destination_cumulio.py
similarity index 97%
rename from src/airbyte/models/shared/destination_cumulio.py
rename to src/airbyte_api/models/shared/destination_cumulio.py
index 68c50de4..eae2db2f 100644
--- a/src/airbyte/models/shared/destination_cumulio.py
+++ b/src/airbyte_api/models/shared/destination_cumulio.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class DestinationCumulio:
api_host: Optional[str] = dataclasses.field(default='https://api.cumul.io', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_host'), 'exclude': lambda f: f is None }})
r"""URL of the Cumul.io API (e.g. 'https://api.cumul.io', 'https://api.us.cumul.io', or VPC-specific API url). Defaults to 'https://api.cumul.io'."""
DESTINATION_TYPE: Final[Cumulio] = dataclasses.field(default=Cumulio.CUMULIO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
-
-
diff --git a/src/airbyte/models/shared/destination_databend.py b/src/airbyte_api/models/shared/destination_databend.py
similarity index 98%
rename from src/airbyte/models/shared/destination_databend.py
rename to src/airbyte_api/models/shared/destination_databend.py
index 6c7411cf..26daf1e2 100644
--- a/src/airbyte/models/shared/destination_databend.py
+++ b/src/airbyte_api/models/shared/destination_databend.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -27,5 +27,3 @@ class DestinationDatabend:
r"""Port of the database."""
table: Optional[str] = dataclasses.field(default='default', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('table'), 'exclude': lambda f: f is None }})
r"""The default table was written to."""
-
-
diff --git a/src/airbyte/models/shared/destination_databricks.py b/src/airbyte_api/models/shared/destination_databricks.py
similarity index 99%
rename from src/airbyte/models/shared/destination_databricks.py
rename to src/airbyte_api/models/shared/destination_databricks.py
index 587d44e2..706ce9b4 100644
--- a/src/airbyte/models/shared/destination_databricks.py
+++ b/src/airbyte_api/models/shared/destination_databricks.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -23,7 +23,7 @@ class DestinationDatabricksAzureBlobStorage:
azure_blob_storage_endpoint_domain_name: Optional[str] = dataclasses.field(default='blob.core.windows.net', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('azure_blob_storage_endpoint_domain_name'), 'exclude': lambda f: f is None }})
r"""This is Azure Blob Storage endpoint domain name. Leave default value (or leave it empty if run container from command line) to use Microsoft native from example."""
DATA_SOURCE_TYPE: Final[DestinationDatabricksSchemasDataSourceType] = dataclasses.field(default=DestinationDatabricksSchemasDataSourceType.AZURE_BLOB_STORAGE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('data_source_type') }})
-
+
class DestinationDatabricksDataSourceType(str, Enum):
@@ -75,7 +75,7 @@ class AmazonS3:
r"""The pattern allows you to set the file-name format for the S3 staging file(s)"""
s3_bucket_region: Optional[DestinationDatabricksS3BucketRegion] = dataclasses.field(default=DestinationDatabricksS3BucketRegion.UNKNOWN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('s3_bucket_region'), 'exclude': lambda f: f is None }})
r"""The region of the S3 staging bucket to use if utilising a copy strategy."""
-
+
class DataSourceType(str, Enum):
@@ -86,7 +86,7 @@ class DataSourceType(str, Enum):
@dataclasses.dataclass
class RecommendedManagedTables:
DATA_SOURCE_TYPE: Final[DataSourceType] = dataclasses.field(default=DataSourceType.MANAGED_TABLES_STORAGE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('data_source_type') }})
-
+
class Databricks(str, Enum):
@@ -117,5 +117,3 @@ class DestinationDatabricks:
r"""Default to 'true'. Switch it to 'false' for debugging purpose."""
schema: Optional[str] = dataclasses.field(default='default', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('schema'), 'exclude': lambda f: f is None }})
r"""The default schema tables are written. If not specified otherwise, the \\"default\\" will be used."""
-
-
diff --git a/src/airbyte/models/shared/destination_dev_null.py b/src/airbyte_api/models/shared/destination_dev_null.py
similarity index 96%
rename from src/airbyte/models/shared/destination_dev_null.py
rename to src/airbyte_api/models/shared/destination_dev_null.py
index 0bc1af45..1dd709ff 100644
--- a/src/airbyte/models/shared/destination_dev_null.py
+++ b/src/airbyte_api/models/shared/destination_dev_null.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -18,7 +18,7 @@ class TestDestinationType(str, Enum):
@dataclasses.dataclass
class Silent:
TEST_DESTINATION_TYPE: Final[Optional[TestDestinationType]] = dataclasses.field(default=TestDestinationType.SILENT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('test_destination_type'), 'exclude': lambda f: f is None }})
-
+
@@ -28,5 +28,3 @@ class DestinationDevNull:
test_destination: Union[Silent] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('test_destination') }})
r"""The type of destination to be used"""
DESTINATION_TYPE: Final[DevNull] = dataclasses.field(default=DevNull.DEV_NULL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
-
-
diff --git a/src/airbyte/models/shared/destination_duckdb.py b/src/airbyte_api/models/shared/destination_duckdb.py
similarity index 97%
rename from src/airbyte/models/shared/destination_duckdb.py
rename to src/airbyte_api/models/shared/destination_duckdb.py
index dfe48afb..1235f95f 100644
--- a/src/airbyte/models/shared/destination_duckdb.py
+++ b/src/airbyte_api/models/shared/destination_duckdb.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class DestinationDuckdb:
r"""API key to use for authentication to a MotherDuck database."""
schema: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('schema'), 'exclude': lambda f: f is None }})
r"""Database schema name, default for duckdb is 'main'."""
-
-
diff --git a/src/airbyte/models/shared/destination_dynamodb.py b/src/airbyte_api/models/shared/destination_dynamodb.py
similarity index 98%
rename from src/airbyte/models/shared/destination_dynamodb.py
rename to src/airbyte_api/models/shared/destination_dynamodb.py
index a5985926..0607528d 100644
--- a/src/airbyte/models/shared/destination_dynamodb.py
+++ b/src/airbyte_api/models/shared/destination_dynamodb.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -62,5 +62,3 @@ class DestinationDynamodb:
r"""This is your DynamoDB endpoint url.(if you are working with AWS DynamoDB, just leave empty)."""
dynamodb_region: Optional[DynamoDBRegion] = dataclasses.field(default=DynamoDBRegion.UNKNOWN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('dynamodb_region'), 'exclude': lambda f: f is None }})
r"""The region of the DynamoDB."""
-
-
diff --git a/src/airbyte/models/shared/destination_elasticsearch.py b/src/airbyte_api/models/shared/destination_elasticsearch.py
similarity index 98%
rename from src/airbyte/models/shared/destination_elasticsearch.py
rename to src/airbyte_api/models/shared/destination_elasticsearch.py
index 8603e70d..d2a361cb 100644
--- a/src/airbyte/models/shared/destination_elasticsearch.py
+++ b/src/airbyte_api/models/shared/destination_elasticsearch.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -20,7 +20,7 @@ class UsernamePassword:
username: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('username') }})
r"""Basic auth username to access a secure Elasticsearch server"""
METHOD: Final[DestinationElasticsearchSchemasMethod] = dataclasses.field(default=DestinationElasticsearchSchemasMethod.BASIC, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class DestinationElasticsearchMethod(str, Enum):
@@ -36,7 +36,7 @@ class APIKeySecret:
api_key_secret: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('apiKeySecret') }})
r"""The secret associated with the API Key ID."""
METHOD: Final[DestinationElasticsearchMethod] = dataclasses.field(default=DestinationElasticsearchMethod.SECRET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class Elasticsearch(str, Enum):
@@ -55,5 +55,3 @@ class DestinationElasticsearch:
DESTINATION_TYPE: Final[Elasticsearch] = dataclasses.field(default=Elasticsearch.ELASTICSEARCH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
upsert: Optional[bool] = dataclasses.field(default=True, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('upsert'), 'exclude': lambda f: f is None }})
r"""If a primary key identifier is defined in the source, an upsert will be performed using the primary key value as the elasticsearch doc id. Does not support composite primary keys."""
-
-
diff --git a/src/airbyte/models/shared/destination_firebolt.py b/src/airbyte_api/models/shared/destination_firebolt.py
similarity index 98%
rename from src/airbyte/models/shared/destination_firebolt.py
rename to src/airbyte_api/models/shared/destination_firebolt.py
index e2bfd4d8..0e4d33ac 100644
--- a/src/airbyte/models/shared/destination_firebolt.py
+++ b/src/airbyte_api/models/shared/destination_firebolt.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -26,7 +26,7 @@ class ExternalTableViaS3:
s3_region: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('s3_region') }})
r"""Region name of the S3 bucket."""
METHOD: Final[DestinationFireboltSchemasMethod] = dataclasses.field(default=DestinationFireboltSchemasMethod.S3, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class DestinationFireboltMethod(str, Enum):
@@ -37,7 +37,7 @@ class DestinationFireboltMethod(str, Enum):
@dataclasses.dataclass
class SQLInserts:
METHOD: Final[DestinationFireboltMethod] = dataclasses.field(default=DestinationFireboltMethod.SQL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
@@ -59,5 +59,3 @@ class DestinationFirebolt:
r"""The host name of your Firebolt database."""
loading_method: Optional[Union[SQLInserts, ExternalTableViaS3]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('loading_method'), 'exclude': lambda f: f is None }})
r"""Loading method used to select the way data will be uploaded to Firebolt"""
-
-
diff --git a/src/airbyte/models/shared/destination_firestore.py b/src/airbyte_api/models/shared/destination_firestore.py
similarity index 97%
rename from src/airbyte/models/shared/destination_firestore.py
rename to src/airbyte_api/models/shared/destination_firestore.py
index b45f7dcf..1cfa1669 100644
--- a/src/airbyte/models/shared/destination_firestore.py
+++ b/src/airbyte_api/models/shared/destination_firestore.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -19,5 +19,3 @@ class DestinationFirestore:
credentials_json: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials_json'), 'exclude': lambda f: f is None }})
r"""The contents of the JSON service account key. Check out the docs if you need help generating this key. Default credentials will be used if this field is left empty."""
DESTINATION_TYPE: Final[Firestore] = dataclasses.field(default=Firestore.FIRESTORE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
-
-
diff --git a/src/airbyte/models/shared/destination_gcs.py b/src/airbyte_api/models/shared/destination_gcs.py
similarity index 99%
rename from src/airbyte/models/shared/destination_gcs.py
rename to src/airbyte_api/models/shared/destination_gcs.py
index d7cd25e4..c3dc0560 100644
--- a/src/airbyte/models/shared/destination_gcs.py
+++ b/src/airbyte_api/models/shared/destination_gcs.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -19,7 +19,7 @@ class HMACKey:
hmac_key_secret: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('hmac_key_secret') }})
r"""The corresponding secret for the access ID. It is a 40-character base-64 encoded string. Read more here."""
credential_type: Optional[CredentialType] = dataclasses.field(default=CredentialType.HMAC_KEY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credential_type'), 'exclude': lambda f: f is None }})
-
+
class Gcs(str, Enum):
@@ -55,7 +55,7 @@ class DestinationGcsParquetColumnarStorage:
r"""Maximum size allowed as padding to align row groups. This is also the minimum size of a row group. Default: 8 MB."""
page_size_kb: Optional[int] = dataclasses.field(default=1024, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('page_size_kb'), 'exclude': lambda f: f is None }})
r"""The page size is for compression. A block is composed of pages. A page is the smallest unit that must be read fully to access a single record. If this value is too small, the compression will deteriorate. Default: 1024 KB."""
-
+
class DestinationGcsSchemasFormatCompressionType(str, Enum):
@@ -66,7 +66,7 @@ class DestinationGcsSchemasFormatCompressionType(str, Enum):
@dataclasses.dataclass
class DestinationGcsGZIP:
compression_type: Optional[DestinationGcsSchemasFormatCompressionType] = dataclasses.field(default=DestinationGcsSchemasFormatCompressionType.GZIP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationGcsSchemasCompressionType(str, Enum):
@@ -77,7 +77,7 @@ class DestinationGcsSchemasCompressionType(str, Enum):
@dataclasses.dataclass
class DestinationGcsSchemasNoCompression:
compression_type: Optional[DestinationGcsSchemasCompressionType] = dataclasses.field(default=DestinationGcsSchemasCompressionType.NO_COMPRESSION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationGcsSchemasFormatFormatType(str, Enum):
@@ -90,7 +90,7 @@ class DestinationGcsJSONLinesNewlineDelimitedJSON:
compression: Optional[Union[DestinationGcsSchemasNoCompression, DestinationGcsGZIP]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression'), 'exclude': lambda f: f is None }})
r"""Whether the output files should be compressed. If compression is selected, the output filename will have an extra extension (GZIP: \\".jsonl.gz\\")."""
format_type: Optional[DestinationGcsSchemasFormatFormatType] = dataclasses.field(default=DestinationGcsSchemasFormatFormatType.JSONL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationGcsCompressionType(str, Enum):
@@ -101,7 +101,7 @@ class DestinationGcsCompressionType(str, Enum):
@dataclasses.dataclass
class Gzip:
compression_type: Optional[DestinationGcsCompressionType] = dataclasses.field(default=DestinationGcsCompressionType.GZIP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_type'), 'exclude': lambda f: f is None }})
-
+
class CompressionType(str, Enum):
@@ -112,7 +112,7 @@ class CompressionType(str, Enum):
@dataclasses.dataclass
class DestinationGcsNoCompression:
compression_type: Optional[CompressionType] = dataclasses.field(default=CompressionType.NO_COMPRESSION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_type'), 'exclude': lambda f: f is None }})
-
+
class Normalization(str, Enum):
@@ -132,7 +132,7 @@ class DestinationGcsCSVCommaSeparatedValues:
flattening: Optional[Normalization] = dataclasses.field(default=Normalization.NO_FLATTENING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('flattening'), 'exclude': lambda f: f is None }})
r"""Whether the input JSON data should be normalized (flattened) in the output CSV. Please refer to docs for details."""
format_type: Optional[DestinationGcsSchemasFormatType] = dataclasses.field(default=DestinationGcsSchemasFormatType.CSV, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationGcsSchemasFormatOutputFormat1Codec(str, Enum):
@@ -143,7 +143,7 @@ class DestinationGcsSchemasFormatOutputFormat1Codec(str, Enum):
@dataclasses.dataclass
class Snappy:
codec: Optional[DestinationGcsSchemasFormatOutputFormat1Codec] = dataclasses.field(default=DestinationGcsSchemasFormatOutputFormat1Codec.SNAPPY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('codec'), 'exclude': lambda f: f is None }})
-
+
class DestinationGcsSchemasFormatOutputFormatCodec(str, Enum):
@@ -158,7 +158,7 @@ class Zstandard:
r"""Negative levels are 'fast' modes akin to lz4 or snappy, levels above 9 are generally for archival purposes, and levels above 18 use a lot of memory."""
include_checksum: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('include_checksum'), 'exclude': lambda f: f is None }})
r"""If true, include a checksum with each data block."""
-
+
class DestinationGcsSchemasFormatCodec(str, Enum):
@@ -171,7 +171,7 @@ class Xz:
codec: Optional[DestinationGcsSchemasFormatCodec] = dataclasses.field(default=DestinationGcsSchemasFormatCodec.XZ, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('codec'), 'exclude': lambda f: f is None }})
compression_level: Optional[int] = dataclasses.field(default=6, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_level'), 'exclude': lambda f: f is None }})
r"""The presets 0-3 are fast presets with medium compression. The presets 4-6 are fairly slow presets with high compression. The default preset is 6. The presets 7-9 are like the preset 6 but use bigger dictionaries and have higher compressor and decompressor memory requirements. Unless the uncompressed size of the file exceeds 8 MiB, 16 MiB, or 32 MiB, it is waste of memory to use the presets 7, 8, or 9, respectively. Read more here for details."""
-
+
class DestinationGcsSchemasCodec(str, Enum):
@@ -182,7 +182,7 @@ class DestinationGcsSchemasCodec(str, Enum):
@dataclasses.dataclass
class Bzip2:
codec: Optional[DestinationGcsSchemasCodec] = dataclasses.field(default=DestinationGcsSchemasCodec.BZIP2, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('codec'), 'exclude': lambda f: f is None }})
-
+
class DestinationGcsCodec(str, Enum):
@@ -195,7 +195,7 @@ class Deflate:
codec: Optional[DestinationGcsCodec] = dataclasses.field(default=DestinationGcsCodec.DEFLATE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('codec'), 'exclude': lambda f: f is None }})
compression_level: Optional[int] = dataclasses.field(default=0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_level'), 'exclude': lambda f: f is None }})
r"""0: no compression & fastest, 9: best compression & slowest."""
-
+
class Codec(str, Enum):
@@ -206,7 +206,7 @@ class Codec(str, Enum):
@dataclasses.dataclass
class NoCompression:
codec: Optional[Codec] = dataclasses.field(default=Codec.NO_COMPRESSION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('codec'), 'exclude': lambda f: f is None }})
-
+
class DestinationGcsFormatType(str, Enum):
@@ -219,7 +219,7 @@ class AvroApacheAvro:
compression_codec: Union[NoCompression, Deflate, Bzip2, Xz, Zstandard, Snappy] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_codec') }})
r"""The compression algorithm used to compress data. Default to no compression."""
format_type: Optional[DestinationGcsFormatType] = dataclasses.field(default=DestinationGcsFormatType.AVRO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type'), 'exclude': lambda f: f is None }})
-
+
class GCSBucketRegion(str, Enum):
@@ -275,5 +275,3 @@ class DestinationGcs:
DESTINATION_TYPE: Final[Gcs] = dataclasses.field(default=Gcs.GCS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
gcs_bucket_region: Optional[GCSBucketRegion] = dataclasses.field(default=GCSBucketRegion.US, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('gcs_bucket_region'), 'exclude': lambda f: f is None }})
r"""Select a Region of the GCS Bucket. Read more here."""
-
-
diff --git a/src/airbyte/models/shared/destination_google_sheets.py b/src/airbyte_api/models/shared/destination_google_sheets.py
similarity index 98%
rename from src/airbyte/models/shared/destination_google_sheets.py
rename to src/airbyte_api/models/shared/destination_google_sheets.py
index 3c237b5c..4b1c9a6f 100644
--- a/src/airbyte/models/shared/destination_google_sheets.py
+++ b/src/airbyte_api/models/shared/destination_google_sheets.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -18,7 +18,7 @@ class AuthenticationViaGoogleOAuth:
r"""The Client Secret of your Google Sheets developer application."""
refresh_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token') }})
r"""The token for obtaining new access token."""
-
+
class DestinationGoogleSheetsGoogleSheets(str, Enum):
@@ -33,5 +33,3 @@ class DestinationGoogleSheets:
spreadsheet_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('spreadsheet_id') }})
r"""The link to your spreadsheet. See this guide for more details."""
DESTINATION_TYPE: Final[DestinationGoogleSheetsGoogleSheets] = dataclasses.field(default=DestinationGoogleSheetsGoogleSheets.GOOGLE_SHEETS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
-
-
diff --git a/src/airbyte/models/shared/destination_keen.py b/src/airbyte_api/models/shared/destination_keen.py
similarity index 97%
rename from src/airbyte/models/shared/destination_keen.py
rename to src/airbyte_api/models/shared/destination_keen.py
index 1bccb775..f8347961 100644
--- a/src/airbyte/models/shared/destination_keen.py
+++ b/src/airbyte_api/models/shared/destination_keen.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class DestinationKeen:
DESTINATION_TYPE: Final[Keen] = dataclasses.field(default=Keen.KEEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
infer_timestamp: Optional[bool] = dataclasses.field(default=True, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('infer_timestamp'), 'exclude': lambda f: f is None }})
r"""Allow connector to guess keen.timestamp value based on the streamed data."""
-
-
diff --git a/src/airbyte/models/shared/destination_kinesis.py b/src/airbyte_api/models/shared/destination_kinesis.py
similarity index 98%
rename from src/airbyte/models/shared/destination_kinesis.py
rename to src/airbyte_api/models/shared/destination_kinesis.py
index 4490f52e..60754575 100644
--- a/src/airbyte/models/shared/destination_kinesis.py
+++ b/src/airbyte_api/models/shared/destination_kinesis.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -27,5 +27,3 @@ class DestinationKinesis:
DESTINATION_TYPE: Final[Kinesis] = dataclasses.field(default=Kinesis.KINESIS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
shard_count: Optional[int] = dataclasses.field(default=5, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('shardCount'), 'exclude': lambda f: f is None }})
r"""Number of shards to which the data should be streamed."""
-
-
diff --git a/src/airbyte/models/shared/destination_langchain.py b/src/airbyte_api/models/shared/destination_langchain.py
similarity index 99%
rename from src/airbyte/models/shared/destination_langchain.py
rename to src/airbyte_api/models/shared/destination_langchain.py
index ce9a4580..e594cfed 100644
--- a/src/airbyte/models/shared/destination_langchain.py
+++ b/src/airbyte_api/models/shared/destination_langchain.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional, Union
@@ -19,7 +19,7 @@ class DestinationLangchainSchemasMode(str, Enum):
class DestinationLangchainFake:
r"""Use a fake embedding made out of random vectors with 1536 embedding dimensions. This is useful for testing the data pipeline without incurring any costs."""
MODE: Final[Optional[DestinationLangchainSchemasMode]] = dataclasses.field(default=DestinationLangchainSchemasMode.FAKE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationLangchainMode(str, Enum):
@@ -32,7 +32,7 @@ class DestinationLangchainOpenAI:
r"""Use the OpenAI API to embed text. This option is using the text-embedding-ada-002 model with 1536 embedding dimensions."""
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
MODE: Final[Optional[DestinationLangchainMode]] = dataclasses.field(default=DestinationLangchainMode.OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationLangchainSchemasIndexingIndexing3Mode(str, Enum):
@@ -48,7 +48,7 @@ class ChromaLocalPersistance:
collection_name: Optional[str] = dataclasses.field(default='langchain', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('collection_name'), 'exclude': lambda f: f is None }})
r"""Name of the collection to use."""
MODE: Final[Optional[DestinationLangchainSchemasIndexingIndexing3Mode]] = dataclasses.field(default=DestinationLangchainSchemasIndexingIndexing3Mode.CHROMA_LOCAL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationLangchainSchemasIndexingIndexingMode(str, Enum):
@@ -62,7 +62,7 @@ class DocArrayHnswSearch:
destination_path: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destination_path') }})
r"""Path to the directory where hnswlib and meta data files will be written. The files will be placed inside that local mount. All files in the specified destination directory will be deleted on each run."""
MODE: Final[Optional[DestinationLangchainSchemasIndexingIndexingMode]] = dataclasses.field(default=DestinationLangchainSchemasIndexingIndexingMode.DOC_ARRAY_HNSW_SEARCH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationLangchainSchemasIndexingMode(str, Enum):
@@ -79,7 +79,7 @@ class DestinationLangchainPinecone:
r"""Pinecone environment to use"""
pinecone_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('pinecone_key') }})
MODE: Final[Optional[DestinationLangchainSchemasIndexingMode]] = dataclasses.field(default=DestinationLangchainSchemasIndexingMode.PINECONE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
@@ -92,7 +92,7 @@ class DestinationLangchainProcessingConfigModel:
r"""List of fields in the record that should be used to calculate the embedding. All other fields are passed along as meta fields. The field list is applied to all streams in the same way and non-existing fields are ignored. If none are defined, all fields are considered text fields. When specifying text fields, you can access nested fields in the record by using dot notation, e.g. `user.name` will access the `name` field in the `user` object. It's also possible to use wildcards to access all fields in an object, e.g. `users.*.name` will access all `names` fields in all entries of the `users` array."""
chunk_overlap: Optional[int] = dataclasses.field(default=0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('chunk_overlap'), 'exclude': lambda f: f is None }})
r"""Size of overlap between chunks in tokens to store in vector store to better capture relevant context"""
-
+
@@ -105,5 +105,3 @@ class DestinationLangchain:
r"""Indexing configuration"""
processing: DestinationLangchainProcessingConfigModel = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('processing') }})
DESTINATION_TYPE: Final[Langchain] = dataclasses.field(default=Langchain.LANGCHAIN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
-
-
diff --git a/src/airbyte/models/shared/destination_milvus.py b/src/airbyte_api/models/shared/destination_milvus.py
similarity index 99%
rename from src/airbyte/models/shared/destination_milvus.py
rename to src/airbyte_api/models/shared/destination_milvus.py
index 973bd0d7..c2d409cc 100644
--- a/src/airbyte/models/shared/destination_milvus.py
+++ b/src/airbyte_api/models/shared/destination_milvus.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional, Union
@@ -26,7 +26,7 @@ class DestinationMilvusOpenAICompatible:
MODE: Final[Optional[DestinationMilvusSchemasEmbeddingEmbedding5Mode]] = dataclasses.field(default=DestinationMilvusSchemasEmbeddingEmbedding5Mode.OPENAI_COMPATIBLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
model_name: Optional[str] = dataclasses.field(default='text-embedding-ada-002', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('model_name'), 'exclude': lambda f: f is None }})
r"""The name of the model to use for embedding"""
-
+
class DestinationMilvusSchemasEmbeddingEmbeddingMode(str, Enum):
@@ -44,7 +44,7 @@ class DestinationMilvusAzureOpenAI:
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
r"""The API key for your Azure OpenAI resource. You can find this in the Azure portal under your Azure OpenAI resource"""
MODE: Final[Optional[DestinationMilvusSchemasEmbeddingEmbeddingMode]] = dataclasses.field(default=DestinationMilvusSchemasEmbeddingEmbeddingMode.AZURE_OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationMilvusSchemasEmbeddingMode(str, Enum):
@@ -56,7 +56,7 @@ class DestinationMilvusSchemasEmbeddingMode(str, Enum):
class DestinationMilvusFake:
r"""Use a fake embedding made out of random vectors with 1536 embedding dimensions. This is useful for testing the data pipeline without incurring any costs."""
MODE: Final[Optional[DestinationMilvusSchemasEmbeddingMode]] = dataclasses.field(default=DestinationMilvusSchemasEmbeddingMode.FAKE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationMilvusSchemasMode(str, Enum):
@@ -69,7 +69,7 @@ class DestinationMilvusCohere:
r"""Use the Cohere API to embed text."""
cohere_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cohere_key') }})
MODE: Final[Optional[DestinationMilvusSchemasMode]] = dataclasses.field(default=DestinationMilvusSchemasMode.COHERE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationMilvusMode(str, Enum):
@@ -82,7 +82,7 @@ class DestinationMilvusOpenAI:
r"""Use the OpenAI API to embed text. This option is using the text-embedding-ada-002 model with 1536 embedding dimensions."""
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
MODE: Final[Optional[DestinationMilvusMode]] = dataclasses.field(default=DestinationMilvusMode.OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationMilvusSchemasIndexingAuthAuthenticationMode(str, Enum):
@@ -94,7 +94,7 @@ class DestinationMilvusSchemasIndexingAuthAuthenticationMode(str, Enum):
class NoAuth:
r"""Do not authenticate (suitable for locally running test clusters, do not use for clusters with public IP addresses)"""
MODE: Final[Optional[DestinationMilvusSchemasIndexingAuthAuthenticationMode]] = dataclasses.field(default=DestinationMilvusSchemasIndexingAuthAuthenticationMode.NO_AUTH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationMilvusSchemasIndexingAuthMode(str, Enum):
@@ -110,7 +110,7 @@ class DestinationMilvusUsernamePassword:
username: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('username') }})
r"""Username for the Milvus instance"""
MODE: Final[Optional[DestinationMilvusSchemasIndexingAuthMode]] = dataclasses.field(default=DestinationMilvusSchemasIndexingAuthMode.USERNAME_PASSWORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationMilvusSchemasIndexingMode(str, Enum):
@@ -124,7 +124,7 @@ class DestinationMilvusAPIToken:
token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token') }})
r"""API Token for the Milvus instance"""
MODE: Final[Optional[DestinationMilvusSchemasIndexingMode]] = dataclasses.field(default=DestinationMilvusSchemasIndexingMode.TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
@@ -144,7 +144,7 @@ class DestinationMilvusIndexing:
r"""The field in the entity that contains the embedded text"""
vector_field: Optional[str] = dataclasses.field(default='vector', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('vector_field'), 'exclude': lambda f: f is None }})
r"""The field in the entity that contains the vector"""
-
+
@@ -155,7 +155,7 @@ class DestinationMilvusFieldNameMappingConfigModel:
r"""The field name in the source"""
to_field: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('to_field') }})
r"""The field name to use in the destination"""
-
+
class DestinationMilvusLanguage(str, Enum):
@@ -188,7 +188,7 @@ class DestinationMilvusByProgrammingLanguage:
language: DestinationMilvusLanguage = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('language') }})
r"""Split code in suitable places based on the programming language"""
MODE: Final[Optional[DestinationMilvusSchemasProcessingTextSplitterTextSplitterMode]] = dataclasses.field(default=DestinationMilvusSchemasProcessingTextSplitterTextSplitterMode.CODE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationMilvusSchemasProcessingTextSplitterMode(str, Enum):
@@ -202,7 +202,7 @@ class DestinationMilvusByMarkdownHeader:
MODE: Final[Optional[DestinationMilvusSchemasProcessingTextSplitterMode]] = dataclasses.field(default=DestinationMilvusSchemasProcessingTextSplitterMode.MARKDOWN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
split_level: Optional[int] = dataclasses.field(default=1, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('split_level'), 'exclude': lambda f: f is None }})
r"""Level of markdown headers to split text fields by. Headings down to the specified level will be used as split points"""
-
+
class DestinationMilvusSchemasProcessingMode(str, Enum):
@@ -218,7 +218,7 @@ class DestinationMilvusBySeparator:
MODE: Final[Optional[DestinationMilvusSchemasProcessingMode]] = dataclasses.field(default=DestinationMilvusSchemasProcessingMode.SEPARATOR, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
separators: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('separators'), 'exclude': lambda f: f is None }})
r"""List of separator strings to split text fields by. The separator itself needs to be wrapped in double quotes, e.g. to split by the dot character, use \\".\\". To split by a newline, use \\"\n\\"."""
-
+
@@ -237,7 +237,7 @@ class DestinationMilvusProcessingConfigModel:
r"""List of fields in the record that should be used to calculate the embedding. The field list is applied to all streams in the same way and non-existing fields are ignored. If none are defined, all fields are considered text fields. When specifying text fields, you can access nested fields in the record by using dot notation, e.g. `user.name` will access the `name` field in the `user` object. It's also possible to use wildcards to access all fields in an object, e.g. `users.*.name` will access all `names` fields in all entries of the `users` array."""
text_splitter: Optional[Union[DestinationMilvusBySeparator, DestinationMilvusByMarkdownHeader, DestinationMilvusByProgrammingLanguage]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('text_splitter'), 'exclude': lambda f: f is None }})
r"""Split text fields into chunks based on the specified method."""
-
+
@@ -263,5 +263,3 @@ class DestinationMilvus:
DESTINATION_TYPE: Final[Milvus] = dataclasses.field(default=Milvus.MILVUS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
omit_raw_text: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('omit_raw_text'), 'exclude': lambda f: f is None }})
r"""Do not store the text that gets embedded along with the vector and the metadata in the destination. If set to true, only the vector and the metadata will be stored - in this case raw text for LLM use cases needs to be retrieved from another source."""
-
-
diff --git a/src/airbyte/models/shared/destination_mongodb.py b/src/airbyte_api/models/shared/destination_mongodb.py
similarity index 99%
rename from src/airbyte/models/shared/destination_mongodb.py
rename to src/airbyte_api/models/shared/destination_mongodb.py
index 0bd8695b..b79aee1b 100644
--- a/src/airbyte/models/shared/destination_mongodb.py
+++ b/src/airbyte_api/models/shared/destination_mongodb.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -20,7 +20,7 @@ class LoginPassword:
username: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('username') }})
r"""Username to use to access the database."""
AUTHORIZATION: Final[DestinationMongodbAuthorization] = dataclasses.field(default=DestinationMongodbAuthorization.LOGIN_PASSWORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('authorization') }})
-
+
class DestinationMongodbSchemasAuthorization(str, Enum):
@@ -32,7 +32,7 @@ class DestinationMongodbSchemasAuthorization(str, Enum):
class NoneT:
r"""None."""
AUTHORIZATION: Final[DestinationMongodbSchemasAuthorization] = dataclasses.field(default=DestinationMongodbSchemasAuthorization.NONE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('authorization') }})
-
+
class Mongodb(str, Enum):
@@ -48,7 +48,7 @@ class MongoDBAtlas:
cluster_url: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cluster_url') }})
r"""URL of a cluster to connect to."""
instance: Optional[DestinationMongodbSchemasInstance] = dataclasses.field(default=DestinationMongodbSchemasInstance.ATLAS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('instance'), 'exclude': lambda f: f is None }})
-
+
class DestinationMongodbInstance(str, Enum):
@@ -63,7 +63,7 @@ class ReplicaSet:
instance: Optional[DestinationMongodbInstance] = dataclasses.field(default=DestinationMongodbInstance.REPLICA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('instance'), 'exclude': lambda f: f is None }})
replica_set: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('replica_set'), 'exclude': lambda f: f is None }})
r"""A replica set name."""
-
+
class Instance(str, Enum):
@@ -78,7 +78,7 @@ class StandaloneMongoDbInstance:
instance: Optional[Instance] = dataclasses.field(default=Instance.STANDALONE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('instance'), 'exclude': lambda f: f is None }})
port: Optional[int] = dataclasses.field(default=27017, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('port'), 'exclude': lambda f: f is None }})
r"""The Port of a Mongo database to be replicated."""
-
+
class DestinationMongodbSchemasTunnelMethodTunnelMethod(str, Enum):
@@ -99,7 +99,7 @@ class DestinationMongodbPasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationMongodbSchemasTunnelMethod(str, Enum):
@@ -120,7 +120,7 @@ class DestinationMongodbSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationMongodbTunnelMethod(str, Enum):
@@ -133,7 +133,7 @@ class DestinationMongodbTunnelMethod(str, Enum):
class DestinationMongodbNoTunnel:
TUNNEL_METHOD: Final[DestinationMongodbTunnelMethod] = dataclasses.field(default=DestinationMongodbTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -149,5 +149,3 @@ class DestinationMongodb:
r"""MongoDb instance to connect to. For MongoDB Atlas and Replica Set TLS connection is used by default."""
tunnel_method: Optional[Union[DestinationMongodbNoTunnel, DestinationMongodbSSHKeyAuthentication, DestinationMongodbPasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/destination_mssql.py b/src/airbyte_api/models/shared/destination_mssql.py
similarity index 99%
rename from src/airbyte/models/shared/destination_mssql.py
rename to src/airbyte_api/models/shared/destination_mssql.py
index 63805ae7..70b1f5be 100644
--- a/src/airbyte/models/shared/destination_mssql.py
+++ b/src/airbyte_api/models/shared/destination_mssql.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -21,7 +21,7 @@ class EncryptedVerifyCertificate:
host_name_in_certificate: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('hostNameInCertificate'), 'exclude': lambda f: f is None }})
r"""Specifies the host name of the server. The value of this property must match the subject property of the certificate."""
SSL_METHOD: Final[Optional[DestinationMssqlSchemasSslMethod]] = dataclasses.field(default=DestinationMssqlSchemasSslMethod.ENCRYPTED_VERIFY_CERTIFICATE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('ssl_method'), 'exclude': lambda f: f is None }})
-
+
class DestinationMssqlSslMethod(str, Enum):
@@ -33,7 +33,7 @@ class DestinationMssqlSslMethod(str, Enum):
class EncryptedTrustServerCertificate:
r"""Use the certificate provided by the server without verification. (For testing purposes only!)"""
SSL_METHOD: Final[Optional[DestinationMssqlSslMethod]] = dataclasses.field(default=DestinationMssqlSslMethod.ENCRYPTED_TRUST_SERVER_CERTIFICATE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('ssl_method'), 'exclude': lambda f: f is None }})
-
+
class DestinationMssqlSchemasTunnelMethodTunnelMethod(str, Enum):
@@ -54,7 +54,7 @@ class DestinationMssqlPasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationMssqlSchemasTunnelMethod(str, Enum):
@@ -75,7 +75,7 @@ class DestinationMssqlSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationMssqlTunnelMethod(str, Enum):
@@ -88,7 +88,7 @@ class DestinationMssqlTunnelMethod(str, Enum):
class DestinationMssqlNoTunnel:
TUNNEL_METHOD: Final[DestinationMssqlTunnelMethod] = dataclasses.field(default=DestinationMssqlTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -114,5 +114,3 @@ class DestinationMssql:
r"""The encryption method which is used to communicate with the database."""
tunnel_method: Optional[Union[DestinationMssqlNoTunnel, DestinationMssqlSSHKeyAuthentication, DestinationMssqlPasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/destination_mysql.py b/src/airbyte_api/models/shared/destination_mysql.py
similarity index 99%
rename from src/airbyte/models/shared/destination_mysql.py
rename to src/airbyte_api/models/shared/destination_mysql.py
index 7dc2c393..c58cac39 100644
--- a/src/airbyte/models/shared/destination_mysql.py
+++ b/src/airbyte_api/models/shared/destination_mysql.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -28,7 +28,7 @@ class DestinationMysqlPasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationMysqlSchemasTunnelMethod(str, Enum):
@@ -49,7 +49,7 @@ class DestinationMysqlSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationMysqlTunnelMethod(str, Enum):
@@ -62,7 +62,7 @@ class DestinationMysqlTunnelMethod(str, Enum):
class DestinationMysqlNoTunnel:
TUNNEL_METHOD: Final[DestinationMysqlTunnelMethod] = dataclasses.field(default=DestinationMysqlTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -84,5 +84,3 @@ class DestinationMysql:
r"""Port of the database."""
tunnel_method: Optional[Union[DestinationMysqlNoTunnel, DestinationMysqlSSHKeyAuthentication, DestinationMysqlPasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/destination_oracle.py b/src/airbyte_api/models/shared/destination_oracle.py
similarity index 99%
rename from src/airbyte/models/shared/destination_oracle.py
rename to src/airbyte_api/models/shared/destination_oracle.py
index 5e7e0d8e..d1a329e9 100644
--- a/src/airbyte/models/shared/destination_oracle.py
+++ b/src/airbyte_api/models/shared/destination_oracle.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -28,7 +28,7 @@ class DestinationOraclePasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationOracleSchemasTunnelMethod(str, Enum):
@@ -49,7 +49,7 @@ class DestinationOracleSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationOracleTunnelMethod(str, Enum):
@@ -62,7 +62,7 @@ class DestinationOracleTunnelMethod(str, Enum):
class DestinationOracleNoTunnel:
TUNNEL_METHOD: Final[DestinationOracleTunnelMethod] = dataclasses.field(default=DestinationOracleTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -86,5 +86,3 @@ class DestinationOracle:
r"""The default schema is used as the target schema for all statements issued from the connection that do not explicitly specify a schema name. The usual value for this field is \\"airbyte\\". In Oracle, schemas and users are the same thing, so the \\"user\\" parameter is used as the login credentials and this is used for the default Airbyte message schema."""
tunnel_method: Optional[Union[DestinationOracleNoTunnel, DestinationOracleSSHKeyAuthentication, DestinationOraclePasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/destination_pinecone.py b/src/airbyte_api/models/shared/destination_pinecone.py
similarity index 99%
rename from src/airbyte/models/shared/destination_pinecone.py
rename to src/airbyte_api/models/shared/destination_pinecone.py
index 95d283ee..b4bd29d4 100644
--- a/src/airbyte/models/shared/destination_pinecone.py
+++ b/src/airbyte_api/models/shared/destination_pinecone.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional, Union
@@ -26,7 +26,7 @@ class DestinationPineconeOpenAICompatible:
MODE: Final[Optional[DestinationPineconeSchemasEmbeddingEmbedding5Mode]] = dataclasses.field(default=DestinationPineconeSchemasEmbeddingEmbedding5Mode.OPENAI_COMPATIBLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
model_name: Optional[str] = dataclasses.field(default='text-embedding-ada-002', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('model_name'), 'exclude': lambda f: f is None }})
r"""The name of the model to use for embedding"""
-
+
class DestinationPineconeSchemasEmbeddingEmbeddingMode(str, Enum):
@@ -44,7 +44,7 @@ class DestinationPineconeAzureOpenAI:
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
r"""The API key for your Azure OpenAI resource. You can find this in the Azure portal under your Azure OpenAI resource"""
MODE: Final[Optional[DestinationPineconeSchemasEmbeddingEmbeddingMode]] = dataclasses.field(default=DestinationPineconeSchemasEmbeddingEmbeddingMode.AZURE_OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationPineconeSchemasEmbeddingMode(str, Enum):
@@ -56,7 +56,7 @@ class DestinationPineconeSchemasEmbeddingMode(str, Enum):
class DestinationPineconeFake:
r"""Use a fake embedding made out of random vectors with 1536 embedding dimensions. This is useful for testing the data pipeline without incurring any costs."""
MODE: Final[Optional[DestinationPineconeSchemasEmbeddingMode]] = dataclasses.field(default=DestinationPineconeSchemasEmbeddingMode.FAKE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationPineconeSchemasMode(str, Enum):
@@ -69,7 +69,7 @@ class DestinationPineconeCohere:
r"""Use the Cohere API to embed text."""
cohere_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cohere_key') }})
MODE: Final[Optional[DestinationPineconeSchemasMode]] = dataclasses.field(default=DestinationPineconeSchemasMode.COHERE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationPineconeMode(str, Enum):
@@ -82,7 +82,7 @@ class DestinationPineconeOpenAI:
r"""Use the OpenAI API to embed text. This option is using the text-embedding-ada-002 model with 1536 embedding dimensions."""
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
MODE: Final[Optional[DestinationPineconeMode]] = dataclasses.field(default=DestinationPineconeMode.OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
@@ -96,7 +96,7 @@ class DestinationPineconeIndexing:
r"""Pinecone Cloud environment to use"""
pinecone_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('pinecone_key') }})
r"""The Pinecone API key to use matching the environment (copy from Pinecone console)"""
-
+
@@ -107,7 +107,7 @@ class DestinationPineconeFieldNameMappingConfigModel:
r"""The field name in the source"""
to_field: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('to_field') }})
r"""The field name to use in the destination"""
-
+
class DestinationPineconeLanguage(str, Enum):
@@ -140,7 +140,7 @@ class DestinationPineconeByProgrammingLanguage:
language: DestinationPineconeLanguage = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('language') }})
r"""Split code in suitable places based on the programming language"""
MODE: Final[Optional[DestinationPineconeSchemasProcessingTextSplitterTextSplitterMode]] = dataclasses.field(default=DestinationPineconeSchemasProcessingTextSplitterTextSplitterMode.CODE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationPineconeSchemasProcessingTextSplitterMode(str, Enum):
@@ -154,7 +154,7 @@ class DestinationPineconeByMarkdownHeader:
MODE: Final[Optional[DestinationPineconeSchemasProcessingTextSplitterMode]] = dataclasses.field(default=DestinationPineconeSchemasProcessingTextSplitterMode.MARKDOWN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
split_level: Optional[int] = dataclasses.field(default=1, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('split_level'), 'exclude': lambda f: f is None }})
r"""Level of markdown headers to split text fields by. Headings down to the specified level will be used as split points"""
-
+
class DestinationPineconeSchemasProcessingMode(str, Enum):
@@ -170,7 +170,7 @@ class DestinationPineconeBySeparator:
MODE: Final[Optional[DestinationPineconeSchemasProcessingMode]] = dataclasses.field(default=DestinationPineconeSchemasProcessingMode.SEPARATOR, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
separators: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('separators'), 'exclude': lambda f: f is None }})
r"""List of separator strings to split text fields by. The separator itself needs to be wrapped in double quotes, e.g. to split by the dot character, use \\".\\". To split by a newline, use \\"\n\\"."""
-
+
@@ -189,7 +189,7 @@ class DestinationPineconeProcessingConfigModel:
r"""List of fields in the record that should be used to calculate the embedding. The field list is applied to all streams in the same way and non-existing fields are ignored. If none are defined, all fields are considered text fields. When specifying text fields, you can access nested fields in the record by using dot notation, e.g. `user.name` will access the `name` field in the `user` object. It's also possible to use wildcards to access all fields in an object, e.g. `users.*.name` will access all `names` fields in all entries of the `users` array."""
text_splitter: Optional[Union[DestinationPineconeBySeparator, DestinationPineconeByMarkdownHeader, DestinationPineconeByProgrammingLanguage]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('text_splitter'), 'exclude': lambda f: f is None }})
r"""Split text fields into chunks based on the specified method."""
-
+
@@ -215,5 +215,3 @@ class DestinationPinecone:
DESTINATION_TYPE: Final[Pinecone] = dataclasses.field(default=Pinecone.PINECONE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
omit_raw_text: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('omit_raw_text'), 'exclude': lambda f: f is None }})
r"""Do not store the text that gets embedded along with the vector and the metadata in the destination. If set to true, only the vector and the metadata will be stored - in this case raw text for LLM use cases needs to be retrieved from another source."""
-
-
diff --git a/src/airbyte/models/shared/destination_postgres.py b/src/airbyte_api/models/shared/destination_postgres.py
similarity index 99%
rename from src/airbyte/models/shared/destination_postgres.py
rename to src/airbyte_api/models/shared/destination_postgres.py
index 6d18101a..ddcbc569 100644
--- a/src/airbyte/models/shared/destination_postgres.py
+++ b/src/airbyte_api/models/shared/destination_postgres.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -27,7 +27,7 @@ class VerifyFull:
client_key_password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_key_password'), 'exclude': lambda f: f is None }})
r"""Password for keystorage. This field is optional. If you do not add it - the password will be generated automatically."""
MODE: Final[Optional[DestinationPostgresSchemasSSLModeSSLModes6Mode]] = dataclasses.field(default=DestinationPostgresSchemasSSLModeSSLModes6Mode.VERIFY_FULL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationPostgresSchemasSSLModeSSLModesMode(str, Enum):
@@ -43,7 +43,7 @@ class VerifyCa:
client_key_password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_key_password'), 'exclude': lambda f: f is None }})
r"""Password for keystorage. This field is optional. If you do not add it - the password will be generated automatically."""
MODE: Final[Optional[DestinationPostgresSchemasSSLModeSSLModesMode]] = dataclasses.field(default=DestinationPostgresSchemasSSLModeSSLModesMode.VERIFY_CA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationPostgresSchemasSslModeMode(str, Enum):
@@ -55,7 +55,7 @@ class DestinationPostgresSchemasSslModeMode(str, Enum):
class Require:
r"""Require SSL mode."""
MODE: Final[Optional[DestinationPostgresSchemasSslModeMode]] = dataclasses.field(default=DestinationPostgresSchemasSslModeMode.REQUIRE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationPostgresSchemasMode(str, Enum):
@@ -67,7 +67,7 @@ class DestinationPostgresSchemasMode(str, Enum):
class Prefer:
r"""Prefer SSL mode."""
MODE: Final[Optional[DestinationPostgresSchemasMode]] = dataclasses.field(default=DestinationPostgresSchemasMode.PREFER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationPostgresMode(str, Enum):
@@ -79,7 +79,7 @@ class DestinationPostgresMode(str, Enum):
class Allow:
r"""Allow SSL mode."""
MODE: Final[Optional[DestinationPostgresMode]] = dataclasses.field(default=DestinationPostgresMode.ALLOW, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class Mode(str, Enum):
@@ -91,7 +91,7 @@ class Mode(str, Enum):
class Disable:
r"""Disable SSL."""
MODE: Final[Optional[Mode]] = dataclasses.field(default=Mode.DISABLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationPostgresSchemasTunnelMethodTunnelMethod(str, Enum):
@@ -112,7 +112,7 @@ class DestinationPostgresPasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationPostgresSchemasTunnelMethod(str, Enum):
@@ -133,7 +133,7 @@ class DestinationPostgresSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationPostgresTunnelMethod(str, Enum):
@@ -146,7 +146,7 @@ class DestinationPostgresTunnelMethod(str, Enum):
class DestinationPostgresNoTunnel:
TUNNEL_METHOD: Final[DestinationPostgresTunnelMethod] = dataclasses.field(default=DestinationPostgresTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -184,5 +184,3 @@ class DestinationPostgres:
"""
tunnel_method: Optional[Union[DestinationPostgresNoTunnel, DestinationPostgresSSHKeyAuthentication, DestinationPostgresPasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/destination_pubsub.py b/src/airbyte_api/models/shared/destination_pubsub.py
similarity index 98%
rename from src/airbyte/models/shared/destination_pubsub.py
rename to src/airbyte_api/models/shared/destination_pubsub.py
index 465cbab6..4428169a 100644
--- a/src/airbyte/models/shared/destination_pubsub.py
+++ b/src/airbyte_api/models/shared/destination_pubsub.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -31,5 +31,3 @@ class DestinationPubsub:
DESTINATION_TYPE: Final[Pubsub] = dataclasses.field(default=Pubsub.PUBSUB, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
ordering_enabled: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('ordering_enabled'), 'exclude': lambda f: f is None }})
r"""If TRUE PubSub publisher will have message ordering enabled. Every message will have an ordering key of stream"""
-
-
diff --git a/src/airbyte/models/shared/destination_qdrant.py b/src/airbyte_api/models/shared/destination_qdrant.py
similarity index 99%
rename from src/airbyte/models/shared/destination_qdrant.py
rename to src/airbyte_api/models/shared/destination_qdrant.py
index 4a1338f0..5dd1910e 100644
--- a/src/airbyte/models/shared/destination_qdrant.py
+++ b/src/airbyte_api/models/shared/destination_qdrant.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional, Union
@@ -26,7 +26,7 @@ class DestinationQdrantOpenAICompatible:
MODE: Final[Optional[DestinationQdrantSchemasEmbeddingEmbedding5Mode]] = dataclasses.field(default=DestinationQdrantSchemasEmbeddingEmbedding5Mode.OPENAI_COMPATIBLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
model_name: Optional[str] = dataclasses.field(default='text-embedding-ada-002', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('model_name'), 'exclude': lambda f: f is None }})
r"""The name of the model to use for embedding"""
-
+
class DestinationQdrantSchemasEmbeddingEmbeddingMode(str, Enum):
@@ -44,7 +44,7 @@ class DestinationQdrantAzureOpenAI:
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
r"""The API key for your Azure OpenAI resource. You can find this in the Azure portal under your Azure OpenAI resource"""
MODE: Final[Optional[DestinationQdrantSchemasEmbeddingEmbeddingMode]] = dataclasses.field(default=DestinationQdrantSchemasEmbeddingEmbeddingMode.AZURE_OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationQdrantSchemasEmbeddingMode(str, Enum):
@@ -56,7 +56,7 @@ class DestinationQdrantSchemasEmbeddingMode(str, Enum):
class DestinationQdrantFake:
r"""Use a fake embedding made out of random vectors with 1536 embedding dimensions. This is useful for testing the data pipeline without incurring any costs."""
MODE: Final[Optional[DestinationQdrantSchemasEmbeddingMode]] = dataclasses.field(default=DestinationQdrantSchemasEmbeddingMode.FAKE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationQdrantSchemasMode(str, Enum):
@@ -69,7 +69,7 @@ class DestinationQdrantCohere:
r"""Use the Cohere API to embed text."""
cohere_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cohere_key') }})
MODE: Final[Optional[DestinationQdrantSchemasMode]] = dataclasses.field(default=DestinationQdrantSchemasMode.COHERE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationQdrantMode(str, Enum):
@@ -82,7 +82,7 @@ class DestinationQdrantOpenAI:
r"""Use the OpenAI API to embed text. This option is using the text-embedding-ada-002 model with 1536 embedding dimensions."""
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
MODE: Final[Optional[DestinationQdrantMode]] = dataclasses.field(default=DestinationQdrantMode.OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationQdrantSchemasIndexingAuthMethodMode(str, Enum):
@@ -93,7 +93,7 @@ class DestinationQdrantSchemasIndexingAuthMethodMode(str, Enum):
@dataclasses.dataclass
class DestinationQdrantNoAuth:
MODE: Final[Optional[DestinationQdrantSchemasIndexingAuthMethodMode]] = dataclasses.field(default=DestinationQdrantSchemasIndexingAuthMethodMode.NO_AUTH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationQdrantSchemasIndexingMode(str, Enum):
@@ -106,7 +106,7 @@ class APIKeyAuth:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""API Key for the Qdrant instance"""
MODE: Final[Optional[DestinationQdrantSchemasIndexingMode]] = dataclasses.field(default=DestinationQdrantSchemasIndexingMode.API_KEY_AUTH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DistanceMetric(str, Enum):
@@ -132,7 +132,7 @@ class DestinationQdrantIndexing:
r"""Whether to prefer gRPC over HTTP. Set to true for Qdrant cloud clusters"""
text_field: Optional[str] = dataclasses.field(default='text', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('text_field'), 'exclude': lambda f: f is None }})
r"""The field in the payload that contains the embedded text"""
-
+
@@ -143,7 +143,7 @@ class DestinationQdrantFieldNameMappingConfigModel:
r"""The field name in the source"""
to_field: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('to_field') }})
r"""The field name to use in the destination"""
-
+
class DestinationQdrantLanguage(str, Enum):
@@ -176,7 +176,7 @@ class DestinationQdrantByProgrammingLanguage:
language: DestinationQdrantLanguage = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('language') }})
r"""Split code in suitable places based on the programming language"""
MODE: Final[Optional[DestinationQdrantSchemasProcessingTextSplitterTextSplitterMode]] = dataclasses.field(default=DestinationQdrantSchemasProcessingTextSplitterTextSplitterMode.CODE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationQdrantSchemasProcessingTextSplitterMode(str, Enum):
@@ -190,7 +190,7 @@ class DestinationQdrantByMarkdownHeader:
MODE: Final[Optional[DestinationQdrantSchemasProcessingTextSplitterMode]] = dataclasses.field(default=DestinationQdrantSchemasProcessingTextSplitterMode.MARKDOWN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
split_level: Optional[int] = dataclasses.field(default=1, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('split_level'), 'exclude': lambda f: f is None }})
r"""Level of markdown headers to split text fields by. Headings down to the specified level will be used as split points"""
-
+
class DestinationQdrantSchemasProcessingMode(str, Enum):
@@ -206,7 +206,7 @@ class DestinationQdrantBySeparator:
MODE: Final[Optional[DestinationQdrantSchemasProcessingMode]] = dataclasses.field(default=DestinationQdrantSchemasProcessingMode.SEPARATOR, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
separators: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('separators'), 'exclude': lambda f: f is None }})
r"""List of separator strings to split text fields by. The separator itself needs to be wrapped in double quotes, e.g. to split by the dot character, use \\".\\". To split by a newline, use \\"\n\\"."""
-
+
@@ -225,7 +225,7 @@ class DestinationQdrantProcessingConfigModel:
r"""List of fields in the record that should be used to calculate the embedding. The field list is applied to all streams in the same way and non-existing fields are ignored. If none are defined, all fields are considered text fields. When specifying text fields, you can access nested fields in the record by using dot notation, e.g. `user.name` will access the `name` field in the `user` object. It's also possible to use wildcards to access all fields in an object, e.g. `users.*.name` will access all `names` fields in all entries of the `users` array."""
text_splitter: Optional[Union[DestinationQdrantBySeparator, DestinationQdrantByMarkdownHeader, DestinationQdrantByProgrammingLanguage]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('text_splitter'), 'exclude': lambda f: f is None }})
r"""Split text fields into chunks based on the specified method."""
-
+
@@ -251,5 +251,3 @@ class DestinationQdrant:
DESTINATION_TYPE: Final[Qdrant] = dataclasses.field(default=Qdrant.QDRANT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
omit_raw_text: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('omit_raw_text'), 'exclude': lambda f: f is None }})
r"""Do not store the text that gets embedded along with the vector and the metadata in the destination. If set to true, only the vector and the metadata will be stored - in this case raw text for LLM use cases needs to be retrieved from another source."""
-
-
diff --git a/src/airbyte/models/shared/destination_redis.py b/src/airbyte_api/models/shared/destination_redis.py
similarity index 99%
rename from src/airbyte/models/shared/destination_redis.py
rename to src/airbyte_api/models/shared/destination_redis.py
index 51ddca94..8cdb2cac 100644
--- a/src/airbyte/models/shared/destination_redis.py
+++ b/src/airbyte_api/models/shared/destination_redis.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -31,7 +31,7 @@ class DestinationRedisVerifyFull:
client_key_password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_key_password'), 'exclude': lambda f: f is None }})
r"""Password for keystorage. If you do not add it - the password will be generated automatically."""
MODE: Final[Optional[DestinationRedisSchemasMode]] = dataclasses.field(default=DestinationRedisSchemasMode.VERIFY_FULL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationRedisMode(str, Enum):
@@ -43,7 +43,7 @@ class DestinationRedisMode(str, Enum):
class DestinationRedisDisable:
r"""Disable SSL."""
MODE: Final[Optional[DestinationRedisMode]] = dataclasses.field(default=DestinationRedisMode.DISABLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationRedisSchemasTunnelMethodTunnelMethod(str, Enum):
@@ -64,7 +64,7 @@ class DestinationRedisPasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationRedisSchemasTunnelMethod(str, Enum):
@@ -85,7 +85,7 @@ class DestinationRedisSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationRedisTunnelMethod(str, Enum):
@@ -98,7 +98,7 @@ class DestinationRedisTunnelMethod(str, Enum):
class DestinationRedisNoTunnel:
TUNNEL_METHOD: Final[DestinationRedisTunnelMethod] = dataclasses.field(default=DestinationRedisTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -124,5 +124,3 @@ class DestinationRedis:
"""
tunnel_method: Optional[Union[DestinationRedisNoTunnel, DestinationRedisSSHKeyAuthentication, DestinationRedisPasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/destination_redshift.py b/src/airbyte_api/models/shared/destination_redshift.py
similarity index 99%
rename from src/airbyte/models/shared/destination_redshift.py
rename to src/airbyte_api/models/shared/destination_redshift.py
index ec2ab028..55dd8fd4 100644
--- a/src/airbyte/models/shared/destination_redshift.py
+++ b/src/airbyte_api/models/shared/destination_redshift.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -28,7 +28,7 @@ class DestinationRedshiftPasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationRedshiftSchemasTunnelMethod(str, Enum):
@@ -49,7 +49,7 @@ class DestinationRedshiftSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationRedshiftTunnelMethod(str, Enum):
@@ -62,7 +62,7 @@ class DestinationRedshiftTunnelMethod(str, Enum):
class DestinationRedshiftNoTunnel:
TUNNEL_METHOD: Final[DestinationRedshiftTunnelMethod] = dataclasses.field(default=DestinationRedshiftTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
class DestinationRedshiftSchemasMethod(str, Enum):
@@ -74,7 +74,7 @@ class DestinationRedshiftSchemasMethod(str, Enum):
class Standard:
r"""(not recommended) Direct loading using SQL INSERT statements. This method is extremely inefficient and provided only for quick testing. In all other cases, you should use S3 uploading."""
METHOD: Final[DestinationRedshiftSchemasMethod] = dataclasses.field(default=DestinationRedshiftSchemasMethod.STANDARD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class DestinationRedshiftEncryptionType(str, Enum):
@@ -88,7 +88,7 @@ class AESCBCEnvelopeEncryption:
ENCRYPTION_TYPE: Final[Optional[DestinationRedshiftEncryptionType]] = dataclasses.field(default=DestinationRedshiftEncryptionType.AES_CBC_ENVELOPE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('encryption_type'), 'exclude': lambda f: f is None }})
key_encrypting_key: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('key_encrypting_key'), 'exclude': lambda f: f is None }})
r"""The key, base64-encoded. Must be either 128, 192, or 256 bits. Leave blank to have Airbyte generate an ephemeral key for each sync."""
-
+
class EncryptionType(str, Enum):
@@ -100,7 +100,7 @@ class EncryptionType(str, Enum):
class NoEncryption:
r"""Staging data will be stored in plaintext."""
ENCRYPTION_TYPE: Final[Optional[EncryptionType]] = dataclasses.field(default=EncryptionType.NONE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('encryption_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationRedshiftMethod(str, Enum):
@@ -167,7 +167,7 @@ class AWSS3Staging:
r"""The directory under the S3 bucket where data will be written. If not provided, then defaults to the root directory. See path's name recommendations for more details."""
s3_bucket_region: Optional[DestinationRedshiftS3BucketRegion] = dataclasses.field(default=DestinationRedshiftS3BucketRegion.UNKNOWN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('s3_bucket_region'), 'exclude': lambda f: f is None }})
r"""The region of the S3 staging bucket."""
-
+
@@ -199,5 +199,3 @@ class DestinationRedshift:
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
uploading_method: Optional[Union[AWSS3Staging, Standard]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('uploading_method'), 'exclude': lambda f: f is None }})
r"""The way data will be uploaded to Redshift."""
-
-
diff --git a/src/airbyte/models/shared/destination_s3.py b/src/airbyte_api/models/shared/destination_s3.py
similarity index 99%
rename from src/airbyte/models/shared/destination_s3.py
rename to src/airbyte_api/models/shared/destination_s3.py
index c05ef878..c69a8dcd 100644
--- a/src/airbyte/models/shared/destination_s3.py
+++ b/src/airbyte_api/models/shared/destination_s3.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -40,7 +40,7 @@ class DestinationS3ParquetColumnarStorage:
r"""Maximum size allowed as padding to align row groups. This is also the minimum size of a row group. Default: 8 MB."""
page_size_kb: Optional[int] = dataclasses.field(default=1024, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('page_size_kb'), 'exclude': lambda f: f is None }})
r"""The page size is for compression. A block is composed of pages. A page is the smallest unit that must be read fully to access a single record. If this value is too small, the compression will deteriorate. Default: 1024 KB."""
-
+
class DestinationS3SchemasFormatOutputFormat3CompressionCodecCodec(str, Enum):
@@ -51,7 +51,7 @@ class DestinationS3SchemasFormatOutputFormat3CompressionCodecCodec(str, Enum):
@dataclasses.dataclass
class DestinationS3Snappy:
codec: Optional[DestinationS3SchemasFormatOutputFormat3CompressionCodecCodec] = dataclasses.field(default=DestinationS3SchemasFormatOutputFormat3CompressionCodecCodec.SNAPPY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('codec'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3SchemasFormatOutputFormat3Codec(str, Enum):
@@ -66,7 +66,7 @@ class DestinationS3Zstandard:
r"""Negative levels are 'fast' modes akin to lz4 or snappy, levels above 9 are generally for archival purposes, and levels above 18 use a lot of memory."""
include_checksum: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('include_checksum'), 'exclude': lambda f: f is None }})
r"""If true, include a checksum with each data block."""
-
+
class DestinationS3SchemasFormatOutputFormatCodec(str, Enum):
@@ -79,7 +79,7 @@ class DestinationS3Xz:
codec: Optional[DestinationS3SchemasFormatOutputFormatCodec] = dataclasses.field(default=DestinationS3SchemasFormatOutputFormatCodec.XZ, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('codec'), 'exclude': lambda f: f is None }})
compression_level: Optional[int] = dataclasses.field(default=6, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_level'), 'exclude': lambda f: f is None }})
r"""See here for details."""
-
+
class DestinationS3SchemasFormatCodec(str, Enum):
@@ -90,7 +90,7 @@ class DestinationS3SchemasFormatCodec(str, Enum):
@dataclasses.dataclass
class DestinationS3Bzip2:
codec: Optional[DestinationS3SchemasFormatCodec] = dataclasses.field(default=DestinationS3SchemasFormatCodec.BZIP2, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('codec'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3SchemasCodec(str, Enum):
@@ -103,7 +103,7 @@ class DestinationS3Deflate:
codec: Optional[DestinationS3SchemasCodec] = dataclasses.field(default=DestinationS3SchemasCodec.DEFLATE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('codec'), 'exclude': lambda f: f is None }})
compression_level: Optional[int] = dataclasses.field(default=0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_level'), 'exclude': lambda f: f is None }})
r"""0: no compression & fastest, 9: best compression & slowest."""
-
+
class DestinationS3Codec(str, Enum):
@@ -114,7 +114,7 @@ class DestinationS3Codec(str, Enum):
@dataclasses.dataclass
class DestinationS3SchemasFormatNoCompression:
codec: Optional[DestinationS3Codec] = dataclasses.field(default=DestinationS3Codec.NO_COMPRESSION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('codec'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3SchemasFormatFormatType(str, Enum):
@@ -127,7 +127,7 @@ class DestinationS3AvroApacheAvro:
compression_codec: Union[DestinationS3SchemasFormatNoCompression, DestinationS3Deflate, DestinationS3Bzip2, DestinationS3Xz, DestinationS3Zstandard, DestinationS3Snappy] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_codec') }})
r"""The compression algorithm used to compress data. Default to no compression."""
format_type: Optional[DestinationS3SchemasFormatFormatType] = dataclasses.field(default=DestinationS3SchemasFormatFormatType.AVRO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3SchemasFormatOutputFormatCompressionType(str, Enum):
@@ -138,7 +138,7 @@ class DestinationS3SchemasFormatOutputFormatCompressionType(str, Enum):
@dataclasses.dataclass
class DestinationS3SchemasGZIP:
compression_type: Optional[DestinationS3SchemasFormatOutputFormatCompressionType] = dataclasses.field(default=DestinationS3SchemasFormatOutputFormatCompressionType.GZIP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3SchemasFormatCompressionType(str, Enum):
@@ -149,7 +149,7 @@ class DestinationS3SchemasFormatCompressionType(str, Enum):
@dataclasses.dataclass
class DestinationS3SchemasNoCompression:
compression_type: Optional[DestinationS3SchemasFormatCompressionType] = dataclasses.field(default=DestinationS3SchemasFormatCompressionType.NO_COMPRESSION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3SchemasFlattening(str, Enum):
@@ -169,7 +169,7 @@ class DestinationS3JSONLinesNewlineDelimitedJSON:
flattening: Optional[DestinationS3SchemasFlattening] = dataclasses.field(default=DestinationS3SchemasFlattening.NO_FLATTENING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('flattening'), 'exclude': lambda f: f is None }})
r"""Whether the input json data should be normalized (flattened) in the output JSON Lines. Please refer to docs for details."""
format_type: Optional[DestinationS3SchemasFormatType] = dataclasses.field(default=DestinationS3SchemasFormatType.JSONL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3SchemasCompressionType(str, Enum):
@@ -180,7 +180,7 @@ class DestinationS3SchemasCompressionType(str, Enum):
@dataclasses.dataclass
class DestinationS3GZIP:
compression_type: Optional[DestinationS3SchemasCompressionType] = dataclasses.field(default=DestinationS3SchemasCompressionType.GZIP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3CompressionType(str, Enum):
@@ -191,7 +191,7 @@ class DestinationS3CompressionType(str, Enum):
@dataclasses.dataclass
class DestinationS3NoCompression:
compression_type: Optional[DestinationS3CompressionType] = dataclasses.field(default=DestinationS3CompressionType.NO_COMPRESSION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3Flattening(str, Enum):
@@ -211,7 +211,7 @@ class DestinationS3CSVCommaSeparatedValues:
flattening: Optional[DestinationS3Flattening] = dataclasses.field(default=DestinationS3Flattening.NO_FLATTENING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('flattening'), 'exclude': lambda f: f is None }})
r"""Whether the input json data should be normalized (flattened) in the output CSV. Please refer to docs for details."""
format_type: Optional[DestinationS3FormatType] = dataclasses.field(default=DestinationS3FormatType.CSV, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3S3BucketRegion(str, Enum):
@@ -274,5 +274,3 @@ class DestinationS3:
r"""Format string on how data will be organized inside the S3 bucket directory. Read more here"""
secret_access_key: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('secret_access_key'), 'exclude': lambda f: f is None }})
r"""The corresponding secret to the access key ID. Read more here"""
-
-
diff --git a/src/airbyte/models/shared/destination_s3_glue.py b/src/airbyte_api/models/shared/destination_s3_glue.py
similarity index 99%
rename from src/airbyte/models/shared/destination_s3_glue.py
rename to src/airbyte_api/models/shared/destination_s3_glue.py
index 9a8f8083..627f57fe 100644
--- a/src/airbyte/models/shared/destination_s3_glue.py
+++ b/src/airbyte_api/models/shared/destination_s3_glue.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -18,7 +18,7 @@ class DestinationS3GlueSchemasCompressionType(str, Enum):
@dataclasses.dataclass
class DestinationS3GlueGZIP:
compression_type: Optional[DestinationS3GlueSchemasCompressionType] = dataclasses.field(default=DestinationS3GlueSchemasCompressionType.GZIP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationS3GlueCompressionType(str, Enum):
@@ -29,7 +29,7 @@ class DestinationS3GlueCompressionType(str, Enum):
@dataclasses.dataclass
class DestinationS3GlueNoCompression:
compression_type: Optional[DestinationS3GlueCompressionType] = dataclasses.field(default=DestinationS3GlueCompressionType.NO_COMPRESSION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('compression_type'), 'exclude': lambda f: f is None }})
-
+
class Flattening(str, Enum):
@@ -49,7 +49,7 @@ class DestinationS3GlueJSONLinesNewlineDelimitedJSON:
flattening: Optional[Flattening] = dataclasses.field(default=Flattening.ROOT_LEVEL_FLATTENING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('flattening'), 'exclude': lambda f: f is None }})
r"""Whether the input json data should be normalized (flattened) in the output JSON Lines. Please refer to docs for details."""
format_type: Optional[DestinationS3GlueFormatType] = dataclasses.field(default=DestinationS3GlueFormatType.JSONL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('format_type'), 'exclude': lambda f: f is None }})
-
+
class SerializationLibrary(str, Enum):
@@ -121,5 +121,3 @@ class DestinationS3Glue:
r"""Format string on how data will be organized inside the S3 bucket directory. Read more here"""
secret_access_key: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('secret_access_key'), 'exclude': lambda f: f is None }})
r"""The corresponding secret to the access key ID. Read more here"""
-
-
diff --git a/src/airbyte/models/shared/destination_sftp_json.py b/src/airbyte_api/models/shared/destination_sftp_json.py
similarity index 97%
rename from src/airbyte/models/shared/destination_sftp_json.py
rename to src/airbyte_api/models/shared/destination_sftp_json.py
index ca44639a..d54077a6 100644
--- a/src/airbyte/models/shared/destination_sftp_json.py
+++ b/src/airbyte_api/models/shared/destination_sftp_json.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -25,5 +25,3 @@ class DestinationSftpJSON:
DESTINATION_TYPE: Final[SftpJSON] = dataclasses.field(default=SftpJSON.SFTP_JSON, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('port'), 'exclude': lambda f: f is None }})
r"""Port of the SFTP server."""
-
-
diff --git a/src/airbyte/models/shared/destination_snowflake.py b/src/airbyte_api/models/shared/destination_snowflake.py
similarity index 99%
rename from src/airbyte/models/shared/destination_snowflake.py
rename to src/airbyte_api/models/shared/destination_snowflake.py
index c3fedaf7..10f27c15 100644
--- a/src/airbyte/models/shared/destination_snowflake.py
+++ b/src/airbyte_api/models/shared/destination_snowflake.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -23,7 +23,7 @@ class DestinationSnowflakeOAuth20:
r"""Enter your application's Client ID"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""Enter your application's Client secret"""
-
+
class DestinationSnowflakeAuthType(str, Enum):
@@ -36,7 +36,7 @@ class UsernameAndPassword:
password: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('password') }})
r"""Enter the password associated with the username."""
AUTH_TYPE: Final[Optional[DestinationSnowflakeAuthType]] = dataclasses.field(default=DestinationSnowflakeAuthType.USERNAME_AND_PASSWORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class DestinationSnowflakeSchemasCredentialsAuthType(str, Enum):
@@ -51,7 +51,7 @@ class KeyPairAuthentication:
AUTH_TYPE: Final[Optional[DestinationSnowflakeSchemasCredentialsAuthType]] = dataclasses.field(default=DestinationSnowflakeSchemasCredentialsAuthType.KEY_PAIR_AUTHENTICATION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
private_key_password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('private_key_password'), 'exclude': lambda f: f is None }})
r"""Passphrase for private key"""
-
+
class DestinationSnowflakeSnowflake(str, Enum):
@@ -83,5 +83,3 @@ class DestinationSnowflake:
r"""Enter the additional properties to pass to the JDBC URL string when connecting to the database (formatted as key=value pairs separated by the symbol &). Example: key1=value1&key2=value2&key3=value3"""
raw_data_schema: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('raw_data_schema'), 'exclude': lambda f: f is None }})
r"""The schema to write raw tables into (default: airbyte_internal)"""
-
-
diff --git a/src/airbyte/models/shared/destination_teradata.py b/src/airbyte_api/models/shared/destination_teradata.py
similarity index 99%
rename from src/airbyte/models/shared/destination_teradata.py
rename to src/airbyte_api/models/shared/destination_teradata.py
index 7a98150a..e0833aab 100644
--- a/src/airbyte/models/shared/destination_teradata.py
+++ b/src/airbyte_api/models/shared/destination_teradata.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -23,7 +23,7 @@ class DestinationTeradataVerifyFull:
See more information - in the docs.
"""
MODE: Final[Optional[DestinationTeradataSchemasSSLModeSSLModes6Mode]] = dataclasses.field(default=DestinationTeradataSchemasSSLModeSSLModes6Mode.VERIFY_FULL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationTeradataSchemasSSLModeSSLModes5Mode(str, Enum):
@@ -39,7 +39,7 @@ class DestinationTeradataVerifyCa:
See more information - in the docs.
"""
MODE: Final[Optional[DestinationTeradataSchemasSSLModeSSLModes5Mode]] = dataclasses.field(default=DestinationTeradataSchemasSSLModeSSLModes5Mode.VERIFY_CA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationTeradataSchemasSSLModeSSLModesMode(str, Enum):
@@ -51,7 +51,7 @@ class DestinationTeradataSchemasSSLModeSSLModesMode(str, Enum):
class DestinationTeradataRequire:
r"""Require SSL mode."""
MODE: Final[Optional[DestinationTeradataSchemasSSLModeSSLModesMode]] = dataclasses.field(default=DestinationTeradataSchemasSSLModeSSLModesMode.REQUIRE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationTeradataSchemasSslModeMode(str, Enum):
@@ -63,7 +63,7 @@ class DestinationTeradataSchemasSslModeMode(str, Enum):
class DestinationTeradataPrefer:
r"""Prefer SSL mode."""
MODE: Final[Optional[DestinationTeradataSchemasSslModeMode]] = dataclasses.field(default=DestinationTeradataSchemasSslModeMode.PREFER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationTeradataSchemasMode(str, Enum):
@@ -75,7 +75,7 @@ class DestinationTeradataSchemasMode(str, Enum):
class DestinationTeradataAllow:
r"""Allow SSL mode."""
MODE: Final[Optional[DestinationTeradataSchemasMode]] = dataclasses.field(default=DestinationTeradataSchemasMode.ALLOW, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationTeradataMode(str, Enum):
@@ -87,7 +87,7 @@ class DestinationTeradataMode(str, Enum):
class DestinationTeradataDisable:
r"""Disable SSL."""
MODE: Final[Optional[DestinationTeradataMode]] = dataclasses.field(default=DestinationTeradataMode.DISABLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
@@ -117,5 +117,3 @@ class DestinationTeradata:
verify-full - This is the most secure mode. Chose this mode to always require encryption and to verify the identity of the destination database server
See more information - in the docs.
"""
-
-
diff --git a/src/airbyte/models/shared/destination_timeplus.py b/src/airbyte_api/models/shared/destination_timeplus.py
similarity index 96%
rename from src/airbyte/models/shared/destination_timeplus.py
rename to src/airbyte_api/models/shared/destination_timeplus.py
index 379e5dd8..c5d2b133 100644
--- a/src/airbyte/models/shared/destination_timeplus.py
+++ b/src/airbyte_api/models/shared/destination_timeplus.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -19,5 +19,3 @@ class DestinationTimeplus:
DESTINATION_TYPE: Final[Timeplus] = dataclasses.field(default=Timeplus.TIMEPLUS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
endpoint: Optional[str] = dataclasses.field(default='https://us.timeplus.cloud/', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('endpoint'), 'exclude': lambda f: f is None }})
r"""Timeplus workspace endpoint"""
-
-
diff --git a/src/airbyte/models/shared/destination_typesense.py b/src/airbyte_api/models/shared/destination_typesense.py
similarity index 98%
rename from src/airbyte/models/shared/destination_typesense.py
rename to src/airbyte_api/models/shared/destination_typesense.py
index 9db797a8..2126ec37 100644
--- a/src/airbyte/models/shared/destination_typesense.py
+++ b/src/airbyte_api/models/shared/destination_typesense.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -25,5 +25,3 @@ class DestinationTypesense:
r"""Port of the Typesense instance. Ex: 8108, 80, 443. Default is 443"""
protocol: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('protocol'), 'exclude': lambda f: f is None }})
r"""Protocol of the Typesense instance. Ex: http or https. Default is https"""
-
-
diff --git a/src/airbyte/models/shared/destination_vectara.py b/src/airbyte_api/models/shared/destination_vectara.py
similarity index 99%
rename from src/airbyte/models/shared/destination_vectara.py
rename to src/airbyte_api/models/shared/destination_vectara.py
index 68e71592..5fa17779 100644
--- a/src/airbyte/models/shared/destination_vectara.py
+++ b/src/airbyte_api/models/shared/destination_vectara.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional
@@ -19,7 +19,7 @@ class OAuth20Credentials:
r"""OAuth2.0 client id"""
client_secret: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret') }})
r"""OAuth2.0 client secret"""
-
+
@@ -42,5 +42,3 @@ class DestinationVectara:
r"""List of fields in the record that should be in the section of the document. The field list is applied to all streams in the same way and non-existing fields are ignored. If none are defined, all fields are considered text fields. When specifying text fields, you can access nested fields in the record by using dot notation, e.g. `user.name` will access the `name` field in the `user` object. It's also possible to use wildcards to access all fields in an object, e.g. `users.*.name` will access all `names` fields in all entries of the `users` array."""
title_field: Optional[str] = dataclasses.field(default='', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('title_field'), 'exclude': lambda f: f is None }})
r"""A field that will be used to populate the `title` of each document. The field list is applied to all streams in the same way and non-existing fields are ignored. If none are defined, all fields are considered text fields. When specifying text fields, you can access nested fields in the record by using dot notation, e.g. `user.name` will access the `name` field in the `user` object. It's also possible to use wildcards to access all fields in an object, e.g. `users.*.name` will access all `names` fields in all entries of the `users` array."""
-
-
diff --git a/src/airbyte/models/shared/destination_vertica.py b/src/airbyte_api/models/shared/destination_vertica.py
similarity index 99%
rename from src/airbyte/models/shared/destination_vertica.py
rename to src/airbyte_api/models/shared/destination_vertica.py
index b4fd28a5..62bb633f 100644
--- a/src/airbyte/models/shared/destination_vertica.py
+++ b/src/airbyte_api/models/shared/destination_vertica.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -28,7 +28,7 @@ class DestinationVerticaPasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationVerticaSchemasTunnelMethod(str, Enum):
@@ -49,7 +49,7 @@ class DestinationVerticaSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class DestinationVerticaTunnelMethod(str, Enum):
@@ -62,7 +62,7 @@ class DestinationVerticaTunnelMethod(str, Enum):
class DestinationVerticaNoTunnel:
TUNNEL_METHOD: Final[DestinationVerticaTunnelMethod] = dataclasses.field(default=DestinationVerticaTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -86,5 +86,3 @@ class DestinationVertica:
r"""Port of the database."""
tunnel_method: Optional[Union[DestinationVerticaNoTunnel, DestinationVerticaSSHKeyAuthentication, DestinationVerticaPasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/destination_weaviate.py b/src/airbyte_api/models/shared/destination_weaviate.py
similarity index 99%
rename from src/airbyte/models/shared/destination_weaviate.py
rename to src/airbyte_api/models/shared/destination_weaviate.py
index 8cc4d45b..115ee729 100644
--- a/src/airbyte/models/shared/destination_weaviate.py
+++ b/src/airbyte_api/models/shared/destination_weaviate.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional, Union
@@ -26,7 +26,7 @@ class DestinationWeaviateOpenAICompatible:
MODE: Final[Optional[DestinationWeaviateSchemasEmbeddingEmbedding7Mode]] = dataclasses.field(default=DestinationWeaviateSchemasEmbeddingEmbedding7Mode.OPENAI_COMPATIBLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
model_name: Optional[str] = dataclasses.field(default='text-embedding-ada-002', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('model_name'), 'exclude': lambda f: f is None }})
r"""The name of the model to use for embedding"""
-
+
class DestinationWeaviateSchemasEmbeddingEmbedding6Mode(str, Enum):
@@ -38,7 +38,7 @@ class DestinationWeaviateSchemasEmbeddingEmbedding6Mode(str, Enum):
class DestinationWeaviateFake:
r"""Use a fake embedding made out of random vectors with 1536 embedding dimensions. This is useful for testing the data pipeline without incurring any costs."""
MODE: Final[Optional[DestinationWeaviateSchemasEmbeddingEmbedding6Mode]] = dataclasses.field(default=DestinationWeaviateSchemasEmbeddingEmbedding6Mode.FAKE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationWeaviateSchemasEmbeddingEmbedding5Mode(str, Enum):
@@ -54,7 +54,7 @@ class FromField:
field_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('field_name') }})
r"""Name of the field in the record that contains the embedding"""
MODE: Final[Optional[DestinationWeaviateSchemasEmbeddingEmbedding5Mode]] = dataclasses.field(default=DestinationWeaviateSchemasEmbeddingEmbedding5Mode.FROM_FIELD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationWeaviateSchemasEmbeddingEmbeddingMode(str, Enum):
@@ -67,7 +67,7 @@ class DestinationWeaviateCohere:
r"""Use the Cohere API to embed text."""
cohere_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cohere_key') }})
MODE: Final[Optional[DestinationWeaviateSchemasEmbeddingEmbeddingMode]] = dataclasses.field(default=DestinationWeaviateSchemasEmbeddingEmbeddingMode.COHERE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationWeaviateSchemasEmbeddingMode(str, Enum):
@@ -80,7 +80,7 @@ class DestinationWeaviateOpenAI:
r"""Use the OpenAI API to embed text. This option is using the text-embedding-ada-002 model with 1536 embedding dimensions."""
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
MODE: Final[Optional[DestinationWeaviateSchemasEmbeddingMode]] = dataclasses.field(default=DestinationWeaviateSchemasEmbeddingMode.OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationWeaviateSchemasMode(str, Enum):
@@ -98,7 +98,7 @@ class DestinationWeaviateAzureOpenAI:
openai_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('openai_key') }})
r"""The API key for your Azure OpenAI resource. You can find this in the Azure portal under your Azure OpenAI resource"""
MODE: Final[Optional[DestinationWeaviateSchemasMode]] = dataclasses.field(default=DestinationWeaviateSchemasMode.AZURE_OPENAI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationWeaviateMode(str, Enum):
@@ -110,7 +110,7 @@ class DestinationWeaviateMode(str, Enum):
class NoExternalEmbedding:
r"""Do not calculate and pass embeddings to Weaviate. Suitable for clusters with configured vectorizers to calculate embeddings within Weaviate or for classes that should only support regular text search."""
MODE: Final[Optional[DestinationWeaviateMode]] = dataclasses.field(default=DestinationWeaviateMode.NO_EMBEDDING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
@@ -119,7 +119,7 @@ class NoExternalEmbedding:
class Header:
header_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_key') }})
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
-
+
class DestinationWeaviateSchemasIndexingAuthAuthenticationMode(str, Enum):
@@ -131,7 +131,7 @@ class DestinationWeaviateSchemasIndexingAuthAuthenticationMode(str, Enum):
class NoAuthentication:
r"""Do not authenticate (suitable for locally running test clusters, do not use for clusters with public IP addresses)"""
MODE: Final[Optional[DestinationWeaviateSchemasIndexingAuthAuthenticationMode]] = dataclasses.field(default=DestinationWeaviateSchemasIndexingAuthAuthenticationMode.NO_AUTH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationWeaviateSchemasIndexingAuthMode(str, Enum):
@@ -147,7 +147,7 @@ class DestinationWeaviateUsernamePassword:
username: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('username') }})
r"""Username for the Weaviate cluster"""
MODE: Final[Optional[DestinationWeaviateSchemasIndexingAuthMode]] = dataclasses.field(default=DestinationWeaviateSchemasIndexingAuthMode.USERNAME_PASSWORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationWeaviateSchemasIndexingMode(str, Enum):
@@ -161,7 +161,7 @@ class DestinationWeaviateAPIToken:
token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token') }})
r"""API Token for the Weaviate instance"""
MODE: Final[Optional[DestinationWeaviateSchemasIndexingMode]] = dataclasses.field(default=DestinationWeaviateSchemasIndexingMode.TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DefaultVectorizer(str, Enum):
@@ -194,7 +194,7 @@ class DestinationWeaviateIndexing:
r"""The tenant ID to use for multi tenancy"""
text_field: Optional[str] = dataclasses.field(default='text', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('text_field'), 'exclude': lambda f: f is None }})
r"""The field in the object that contains the embedded text"""
-
+
@@ -205,7 +205,7 @@ class DestinationWeaviateFieldNameMappingConfigModel:
r"""The field name in the source"""
to_field: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('to_field') }})
r"""The field name to use in the destination"""
-
+
class DestinationWeaviateLanguage(str, Enum):
@@ -238,7 +238,7 @@ class DestinationWeaviateByProgrammingLanguage:
language: DestinationWeaviateLanguage = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('language') }})
r"""Split code in suitable places based on the programming language"""
MODE: Final[Optional[DestinationWeaviateSchemasProcessingTextSplitterTextSplitterMode]] = dataclasses.field(default=DestinationWeaviateSchemasProcessingTextSplitterTextSplitterMode.CODE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class DestinationWeaviateSchemasProcessingTextSplitterMode(str, Enum):
@@ -252,7 +252,7 @@ class DestinationWeaviateByMarkdownHeader:
MODE: Final[Optional[DestinationWeaviateSchemasProcessingTextSplitterMode]] = dataclasses.field(default=DestinationWeaviateSchemasProcessingTextSplitterMode.MARKDOWN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
split_level: Optional[int] = dataclasses.field(default=1, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('split_level'), 'exclude': lambda f: f is None }})
r"""Level of markdown headers to split text fields by. Headings down to the specified level will be used as split points"""
-
+
class DestinationWeaviateSchemasProcessingMode(str, Enum):
@@ -268,7 +268,7 @@ class DestinationWeaviateBySeparator:
MODE: Final[Optional[DestinationWeaviateSchemasProcessingMode]] = dataclasses.field(default=DestinationWeaviateSchemasProcessingMode.SEPARATOR, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
separators: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('separators'), 'exclude': lambda f: f is None }})
r"""List of separator strings to split text fields by. The separator itself needs to be wrapped in double quotes, e.g. to split by the dot character, use \\".\\". To split by a newline, use \\"\n\\"."""
-
+
@@ -287,7 +287,7 @@ class DestinationWeaviateProcessingConfigModel:
r"""List of fields in the record that should be used to calculate the embedding. The field list is applied to all streams in the same way and non-existing fields are ignored. If none are defined, all fields are considered text fields. When specifying text fields, you can access nested fields in the record by using dot notation, e.g. `user.name` will access the `name` field in the `user` object. It's also possible to use wildcards to access all fields in an object, e.g. `users.*.name` will access all `names` fields in all entries of the `users` array."""
text_splitter: Optional[Union[DestinationWeaviateBySeparator, DestinationWeaviateByMarkdownHeader, DestinationWeaviateByProgrammingLanguage]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('text_splitter'), 'exclude': lambda f: f is None }})
r"""Split text fields into chunks based on the specified method."""
-
+
@@ -313,5 +313,3 @@ class DestinationWeaviate:
DESTINATION_TYPE: Final[Weaviate] = dataclasses.field(default=Weaviate.WEAVIATE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
omit_raw_text: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('omit_raw_text'), 'exclude': lambda f: f is None }})
r"""Do not store the text that gets embedded along with the vector and the metadata in the destination. If set to true, only the vector and the metadata will be stored - in this case raw text for LLM use cases needs to be retrieved from another source."""
-
-
diff --git a/src/airbyte/models/shared/destination_xata.py b/src/airbyte_api/models/shared/destination_xata.py
similarity index 96%
rename from src/airbyte/models/shared/destination_xata.py
rename to src/airbyte_api/models/shared/destination_xata.py
index e7c6dcdf..c1067858 100644
--- a/src/airbyte/models/shared/destination_xata.py
+++ b/src/airbyte_api/models/shared/destination_xata.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class DestinationXata:
db_url: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('db_url') }})
r"""URL pointing to your workspace."""
DESTINATION_TYPE: Final[Xata] = dataclasses.field(default=Xata.XATA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
-
-
diff --git a/src/airbyte/models/shared/destinationcreaterequest.py b/src/airbyte_api/models/shared/destinationcreaterequest.py
similarity index 99%
rename from src/airbyte/models/shared/destinationcreaterequest.py
rename to src/airbyte_api/models/shared/destinationcreaterequest.py
index 2f862d12..b433f271 100644
--- a/src/airbyte/models/shared/destinationcreaterequest.py
+++ b/src/airbyte_api/models/shared/destinationcreaterequest.py
@@ -44,7 +44,7 @@
from .destination_vertica import DestinationVertica
from .destination_weaviate import DestinationWeaviate
from .destination_xata import DestinationXata
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional, Union
@@ -59,5 +59,3 @@ class DestinationCreateRequest:
workspace_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('workspaceId') }})
definition_id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('definitionId'), 'exclude': lambda f: f is None }})
r"""The UUID of the connector definition. One of configuration.destinationType or definitionId must be provided."""
-
-
diff --git a/src/airbyte/models/shared/destinationpatchrequest.py b/src/airbyte_api/models/shared/destinationpatchrequest.py
similarity index 99%
rename from src/airbyte/models/shared/destinationpatchrequest.py
rename to src/airbyte_api/models/shared/destinationpatchrequest.py
index 8beca8cd..2c4372e2 100644
--- a/src/airbyte/models/shared/destinationpatchrequest.py
+++ b/src/airbyte_api/models/shared/destinationpatchrequest.py
@@ -44,7 +44,7 @@
from .destination_vertica import DestinationVertica
from .destination_weaviate import DestinationWeaviate
from .destination_xata import DestinationXata
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional, Union
@@ -55,5 +55,3 @@ class DestinationPatchRequest:
configuration: Optional[Union[DestinationGoogleSheets, DestinationAstra, DestinationAwsDatalake, DestinationAzureBlobStorage, DestinationBigquery, DestinationClickhouse, DestinationConvex, DestinationCumulio, DestinationDatabend, DestinationDatabricks, DestinationDevNull, DestinationDuckdb, DestinationDynamodb, DestinationElasticsearch, DestinationFirebolt, DestinationFirestore, DestinationGcs, DestinationKeen, DestinationKinesis, DestinationLangchain, DestinationMilvus, DestinationMongodb, DestinationMssql, DestinationMysql, DestinationOracle, DestinationPinecone, DestinationPostgres, DestinationPubsub, DestinationQdrant, DestinationRedis, DestinationRedshift, DestinationS3, DestinationS3Glue, DestinationSftpJSON, DestinationSnowflake, DestinationTeradata, DestinationTimeplus, DestinationTypesense, DestinationVectara, DestinationVertica, DestinationWeaviate, DestinationXata]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('configuration'), 'exclude': lambda f: f is None }})
r"""The values required to configure the destination."""
name: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('name'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/destinationputrequest.py b/src/airbyte_api/models/shared/destinationputrequest.py
similarity index 99%
rename from src/airbyte/models/shared/destinationputrequest.py
rename to src/airbyte_api/models/shared/destinationputrequest.py
index 0e869064..bb4a61cf 100644
--- a/src/airbyte/models/shared/destinationputrequest.py
+++ b/src/airbyte_api/models/shared/destinationputrequest.py
@@ -44,7 +44,7 @@
from .destination_vertica import DestinationVertica
from .destination_weaviate import DestinationWeaviate
from .destination_xata import DestinationXata
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Union
@@ -55,5 +55,3 @@ class DestinationPutRequest:
configuration: Union[DestinationGoogleSheets, DestinationAstra, DestinationAwsDatalake, DestinationAzureBlobStorage, DestinationBigquery, DestinationClickhouse, DestinationConvex, DestinationCumulio, DestinationDatabend, DestinationDatabricks, DestinationDevNull, DestinationDuckdb, DestinationDynamodb, DestinationElasticsearch, DestinationFirebolt, DestinationFirestore, DestinationGcs, DestinationKeen, DestinationKinesis, DestinationLangchain, DestinationMilvus, DestinationMongodb, DestinationMssql, DestinationMysql, DestinationOracle, DestinationPinecone, DestinationPostgres, DestinationPubsub, DestinationQdrant, DestinationRedis, DestinationRedshift, DestinationS3, DestinationS3Glue, DestinationSftpJSON, DestinationSnowflake, DestinationTeradata, DestinationTimeplus, DestinationTypesense, DestinationVectara, DestinationVertica, DestinationWeaviate, DestinationXata] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('configuration') }})
r"""The values required to configure the destination."""
name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('name') }})
-
-
diff --git a/src/airbyte/models/shared/destinationresponse.py b/src/airbyte_api/models/shared/destinationresponse.py
similarity index 99%
rename from src/airbyte/models/shared/destinationresponse.py
rename to src/airbyte_api/models/shared/destinationresponse.py
index 071b354c..00c0abfa 100644
--- a/src/airbyte/models/shared/destinationresponse.py
+++ b/src/airbyte_api/models/shared/destinationresponse.py
@@ -44,7 +44,7 @@
from .destination_vertica import DestinationVertica
from .destination_weaviate import DestinationWeaviate
from .destination_xata import DestinationXata
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Union
@@ -59,5 +59,3 @@ class DestinationResponse:
destination_type: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('destinationType') }})
name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('name') }})
workspace_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('workspaceId') }})
-
-
diff --git a/src/airbyte/models/shared/destinationsresponse.py b/src/airbyte_api/models/shared/destinationsresponse.py
similarity index 96%
rename from src/airbyte/models/shared/destinationsresponse.py
rename to src/airbyte_api/models/shared/destinationsresponse.py
index 10a1e74d..81d73778 100644
--- a/src/airbyte/models/shared/destinationsresponse.py
+++ b/src/airbyte_api/models/shared/destinationsresponse.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .destinationresponse import DestinationResponse
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import List, Optional
@@ -14,5 +14,3 @@ class DestinationsResponse:
data: List[DestinationResponse] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('data') }})
next: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('next'), 'exclude': lambda f: f is None }})
previous: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('previous'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/facebook_marketing.py b/src/airbyte_api/models/shared/facebook_marketing.py
similarity index 96%
rename from src/airbyte/models/shared/facebook_marketing.py
rename to src/airbyte_api/models/shared/facebook_marketing.py
index da6676bd..5bcadab1 100644
--- a/src/airbyte/models/shared/facebook_marketing.py
+++ b/src/airbyte_api/models/shared/facebook_marketing.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,5 +14,3 @@ class FacebookMarketing:
r"""The Client Id for your OAuth app"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret for your OAuth app"""
-
-
diff --git a/src/airbyte/models/shared/geographyenum.py b/src/airbyte_api/models/shared/geographyenum.py
similarity index 100%
rename from src/airbyte/models/shared/geographyenum.py
rename to src/airbyte_api/models/shared/geographyenum.py
diff --git a/src/airbyte/models/shared/geographyenumnodefault.py b/src/airbyte_api/models/shared/geographyenumnodefault.py
similarity index 100%
rename from src/airbyte/models/shared/geographyenumnodefault.py
rename to src/airbyte_api/models/shared/geographyenumnodefault.py
diff --git a/src/airbyte/models/shared/github.py b/src/airbyte_api/models/shared/github.py
similarity index 96%
rename from src/airbyte/models/shared/github.py
rename to src/airbyte_api/models/shared/github.py
index e3a50693..01ba4fc0 100644
--- a/src/airbyte/models/shared/github.py
+++ b/src/airbyte_api/models/shared/github.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class GithubCredentials:
r"""OAuth Client Id"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""OAuth Client secret"""
-
+
@@ -22,5 +22,3 @@ class GithubCredentials:
@dataclasses.dataclass
class Github:
credentials: Optional[GithubCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/gitlab.py b/src/airbyte_api/models/shared/gitlab.py
similarity index 96%
rename from src/airbyte/models/shared/gitlab.py
rename to src/airbyte_api/models/shared/gitlab.py
index 0e39edc8..cf00d0db 100644
--- a/src/airbyte/models/shared/gitlab.py
+++ b/src/airbyte_api/models/shared/gitlab.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class GitlabCredentials:
r"""The API ID of the Gitlab developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The API Secret the Gitlab developer application."""
-
+
@@ -22,5 +22,3 @@ class GitlabCredentials:
@dataclasses.dataclass
class Gitlab:
credentials: Optional[GitlabCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/google_ads.py b/src/airbyte_api/models/shared/google_ads.py
similarity index 98%
rename from src/airbyte/models/shared/google_ads.py
rename to src/airbyte_api/models/shared/google_ads.py
index 1b8960c2..e05c4c00 100644
--- a/src/airbyte/models/shared/google_ads.py
+++ b/src/airbyte_api/models/shared/google_ads.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -16,7 +16,7 @@ class GoogleAdsCredentials:
r"""The Client Secret of your Google Ads developer application. For detailed instructions on finding this value, refer to our documentation."""
developer_token: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('developer_token'), 'exclude': lambda f: f is None }})
r"""The Developer Token granted by Google to use their APIs. For detailed instructions on finding this value, refer to our documentation."""
-
+
@@ -24,5 +24,3 @@ class GoogleAdsCredentials:
@dataclasses.dataclass
class GoogleAds:
credentials: Optional[GoogleAdsCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/google_analytics_data_api.py b/src/airbyte_api/models/shared/google_analytics_data_api.py
similarity index 96%
rename from src/airbyte/models/shared/google_analytics_data_api.py
rename to src/airbyte_api/models/shared/google_analytics_data_api.py
index eb64a27b..ba9932aa 100644
--- a/src/airbyte/models/shared/google_analytics_data_api.py
+++ b/src/airbyte_api/models/shared/google_analytics_data_api.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class GoogleAnalyticsDataAPICredentials:
r"""The Client ID of your Google Analytics developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Google Analytics developer application."""
-
+
@@ -22,5 +22,3 @@ class GoogleAnalyticsDataAPICredentials:
@dataclasses.dataclass
class GoogleAnalyticsDataAPI:
credentials: Optional[GoogleAnalyticsDataAPICredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/google_drive.py b/src/airbyte_api/models/shared/google_drive.py
similarity index 96%
rename from src/airbyte/models/shared/google_drive.py
rename to src/airbyte_api/models/shared/google_drive.py
index 10418621..c83ded71 100644
--- a/src/airbyte/models/shared/google_drive.py
+++ b/src/airbyte_api/models/shared/google_drive.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class GoogleDriveCredentials:
r"""Client ID for the Google Drive API"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""Client Secret for the Google Drive API"""
-
+
@@ -22,5 +22,3 @@ class GoogleDriveCredentials:
@dataclasses.dataclass
class GoogleDrive:
credentials: Optional[GoogleDriveCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/google_search_console.py b/src/airbyte_api/models/shared/google_search_console.py
similarity index 97%
rename from src/airbyte/models/shared/google_search_console.py
rename to src/airbyte_api/models/shared/google_search_console.py
index 2cbc4943..22f63958 100644
--- a/src/airbyte/models/shared/google_search_console.py
+++ b/src/airbyte_api/models/shared/google_search_console.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class Authorization:
r"""The client ID of your Google Search Console developer application. Read more here."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The client secret of your Google Search Console developer application. Read more here."""
-
+
@@ -22,5 +22,3 @@ class Authorization:
@dataclasses.dataclass
class GoogleSearchConsole:
authorization: Optional[Authorization] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('authorization'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/google_sheets.py b/src/airbyte_api/models/shared/google_sheets.py
similarity index 97%
rename from src/airbyte/models/shared/google_sheets.py
rename to src/airbyte_api/models/shared/google_sheets.py
index c4ac1fd2..92f6ca59 100644
--- a/src/airbyte/models/shared/google_sheets.py
+++ b/src/airbyte_api/models/shared/google_sheets.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class GoogleSheetsCredentials:
r"""Enter your Google application's Client ID. See Google's documentation for more information."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""Enter your Google application's Client Secret. See Google's documentation for more information."""
-
+
@@ -22,5 +22,3 @@ class GoogleSheetsCredentials:
@dataclasses.dataclass
class GoogleSheets:
credentials: Optional[GoogleSheetsCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/harvest.py b/src/airbyte_api/models/shared/harvest.py
similarity index 96%
rename from src/airbyte/models/shared/harvest.py
rename to src/airbyte_api/models/shared/harvest.py
index 6d1d8755..29e92f6b 100644
--- a/src/airbyte/models/shared/harvest.py
+++ b/src/airbyte_api/models/shared/harvest.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class HarvestCredentials:
r"""The Client ID of your Harvest developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Harvest developer application."""
-
+
@@ -22,5 +22,3 @@ class HarvestCredentials:
@dataclasses.dataclass
class Harvest:
credentials: Optional[HarvestCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/hubspot.py b/src/airbyte_api/models/shared/hubspot.py
similarity index 97%
rename from src/airbyte/models/shared/hubspot.py
rename to src/airbyte_api/models/shared/hubspot.py
index fe8cc96a..be1f3de2 100644
--- a/src/airbyte/models/shared/hubspot.py
+++ b/src/airbyte_api/models/shared/hubspot.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class HubspotCredentials:
r"""The Client ID of your HubSpot developer application. See the Hubspot docs if you need help finding this ID."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The client secret for your HubSpot developer application. See the Hubspot docs if you need help finding this secret."""
-
+
@@ -22,5 +22,3 @@ class HubspotCredentials:
@dataclasses.dataclass
class Hubspot:
credentials: Optional[HubspotCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/initiateoauthrequest.py b/src/airbyte_api/models/shared/initiateoauthrequest.py
similarity index 97%
rename from src/airbyte/models/shared/initiateoauthrequest.py
rename to src/airbyte_api/models/shared/initiateoauthrequest.py
index 717bf074..63a39303 100644
--- a/src/airbyte/models/shared/initiateoauthrequest.py
+++ b/src/airbyte_api/models/shared/initiateoauthrequest.py
@@ -4,7 +4,7 @@
import dataclasses
from .oauthactornames import OAuthActorNames
from .oauthinputconfiguration import OAuthInputConfiguration
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -20,5 +20,3 @@ class InitiateOauthRequest:
r"""The workspace to create the secret and eventually the full source."""
o_auth_input_configuration: Optional[OAuthInputConfiguration] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('oAuthInputConfiguration'), 'exclude': lambda f: f is None }})
r"""Arbitrary vars to pass for OAuth depending on what the source/destination spec requires."""
-
-
diff --git a/src/airbyte/models/shared/instagram.py b/src/airbyte_api/models/shared/instagram.py
similarity index 96%
rename from src/airbyte/models/shared/instagram.py
rename to src/airbyte_api/models/shared/instagram.py
index 1da88057..4dc06a88 100644
--- a/src/airbyte/models/shared/instagram.py
+++ b/src/airbyte_api/models/shared/instagram.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,5 +14,3 @@ class Instagram:
r"""The Client ID for your Oauth application"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret for your Oauth application"""
-
-
diff --git a/src/airbyte/models/shared/intercom.py b/src/airbyte_api/models/shared/intercom.py
similarity index 96%
rename from src/airbyte/models/shared/intercom.py
rename to src/airbyte_api/models/shared/intercom.py
index c6502e11..8bdce374 100644
--- a/src/airbyte/models/shared/intercom.py
+++ b/src/airbyte_api/models/shared/intercom.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,5 +14,3 @@ class Intercom:
r"""Client Id for your Intercom application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""Client Secret for your Intercom application."""
-
-
diff --git a/src/airbyte/models/shared/jobcreaterequest.py b/src/airbyte_api/models/shared/jobcreaterequest.py
similarity index 95%
rename from src/airbyte/models/shared/jobcreaterequest.py
rename to src/airbyte_api/models/shared/jobcreaterequest.py
index 53410c05..bf2c7e2c 100644
--- a/src/airbyte/models/shared/jobcreaterequest.py
+++ b/src/airbyte_api/models/shared/jobcreaterequest.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .jobtypeenum import JobTypeEnum
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
@@ -14,5 +14,3 @@ class JobCreateRequest:
connection_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('connectionId') }})
job_type: JobTypeEnum = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('jobType') }})
r"""Enum that describes the different types of jobs that the platform runs."""
-
-
diff --git a/src/airbyte/models/shared/jobresponse.py b/src/airbyte_api/models/shared/jobresponse.py
similarity index 98%
rename from src/airbyte/models/shared/jobresponse.py
rename to src/airbyte_api/models/shared/jobresponse.py
index cd971ff1..d96fbe83 100644
--- a/src/airbyte/models/shared/jobresponse.py
+++ b/src/airbyte_api/models/shared/jobresponse.py
@@ -4,7 +4,7 @@
import dataclasses
from .jobstatusenum import JobStatusEnum
from .jobtypeenum import JobTypeEnum
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -24,5 +24,3 @@ class JobResponse:
r"""Duration of a sync in ISO_8601 format"""
last_updated_at: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('lastUpdatedAt'), 'exclude': lambda f: f is None }})
rows_synced: Optional[int] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('rowsSynced'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/jobsresponse.py b/src/airbyte_api/models/shared/jobsresponse.py
similarity index 96%
rename from src/airbyte/models/shared/jobsresponse.py
rename to src/airbyte_api/models/shared/jobsresponse.py
index d7ce4a44..0af16692 100644
--- a/src/airbyte/models/shared/jobsresponse.py
+++ b/src/airbyte_api/models/shared/jobsresponse.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .jobresponse import JobResponse
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import List, Optional
@@ -14,5 +14,3 @@ class JobsResponse:
data: List[JobResponse] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('data') }})
next: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('next'), 'exclude': lambda f: f is None }})
previous: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('previous'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/jobstatusenum.py b/src/airbyte_api/models/shared/jobstatusenum.py
similarity index 100%
rename from src/airbyte/models/shared/jobstatusenum.py
rename to src/airbyte_api/models/shared/jobstatusenum.py
diff --git a/src/airbyte/models/shared/jobtypeenum.py b/src/airbyte_api/models/shared/jobtypeenum.py
similarity index 100%
rename from src/airbyte/models/shared/jobtypeenum.py
rename to src/airbyte_api/models/shared/jobtypeenum.py
diff --git a/src/airbyte/models/shared/lever_hiring.py b/src/airbyte_api/models/shared/lever_hiring.py
similarity index 96%
rename from src/airbyte/models/shared/lever_hiring.py
rename to src/airbyte_api/models/shared/lever_hiring.py
index 85a239e9..749e37e7 100644
--- a/src/airbyte/models/shared/lever_hiring.py
+++ b/src/airbyte_api/models/shared/lever_hiring.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class LeverHiringCredentials:
r"""The Client ID of your Lever Hiring developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Lever Hiring developer application."""
-
+
@@ -22,5 +22,3 @@ class LeverHiringCredentials:
@dataclasses.dataclass
class LeverHiring:
credentials: Optional[LeverHiringCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/linkedin_ads.py b/src/airbyte_api/models/shared/linkedin_ads.py
similarity index 97%
rename from src/airbyte/models/shared/linkedin_ads.py
rename to src/airbyte_api/models/shared/linkedin_ads.py
index 71934959..50eed0b6 100644
--- a/src/airbyte/models/shared/linkedin_ads.py
+++ b/src/airbyte_api/models/shared/linkedin_ads.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class LinkedinAdsCredentials:
r"""The client ID of your developer application. Refer to our documentation for more information."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The client secret of your developer application. Refer to our documentation for more information."""
-
+
@@ -22,5 +22,3 @@ class LinkedinAdsCredentials:
@dataclasses.dataclass
class LinkedinAds:
credentials: Optional[LinkedinAdsCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/mailchimp.py b/src/airbyte_api/models/shared/mailchimp.py
similarity index 96%
rename from src/airbyte/models/shared/mailchimp.py
rename to src/airbyte_api/models/shared/mailchimp.py
index a6a07dff..b7d08bf2 100644
--- a/src/airbyte/models/shared/mailchimp.py
+++ b/src/airbyte_api/models/shared/mailchimp.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class MailchimpCredentials:
r"""The Client ID of your OAuth application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your OAuth application."""
-
+
@@ -22,5 +22,3 @@ class MailchimpCredentials:
@dataclasses.dataclass
class Mailchimp:
credentials: Optional[MailchimpCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/microsoft_sharepoint.py b/src/airbyte_api/models/shared/microsoft_sharepoint.py
similarity index 96%
rename from src/airbyte/models/shared/microsoft_sharepoint.py
rename to src/airbyte_api/models/shared/microsoft_sharepoint.py
index 530930b9..600749ae 100644
--- a/src/airbyte/models/shared/microsoft_sharepoint.py
+++ b/src/airbyte_api/models/shared/microsoft_sharepoint.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class MicrosoftSharepointCredentials:
r"""Client ID of your Microsoft developer application"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""Client Secret of your Microsoft developer application"""
-
+
@@ -22,5 +22,3 @@ class MicrosoftSharepointCredentials:
@dataclasses.dataclass
class MicrosoftSharepoint:
credentials: Optional[MicrosoftSharepointCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/microsoft_teams.py b/src/airbyte_api/models/shared/microsoft_teams.py
similarity index 96%
rename from src/airbyte/models/shared/microsoft_teams.py
rename to src/airbyte_api/models/shared/microsoft_teams.py
index d263ecea..d49e085b 100644
--- a/src/airbyte/models/shared/microsoft_teams.py
+++ b/src/airbyte_api/models/shared/microsoft_teams.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class MicrosoftTeamsCredentials:
r"""The Client ID of your Microsoft Teams developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Microsoft Teams developer application."""
-
+
@@ -22,5 +22,3 @@ class MicrosoftTeamsCredentials:
@dataclasses.dataclass
class MicrosoftTeams:
credentials: Optional[MicrosoftTeamsCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/monday.py b/src/airbyte_api/models/shared/monday.py
similarity index 96%
rename from src/airbyte/models/shared/monday.py
rename to src/airbyte_api/models/shared/monday.py
index 60611ef4..dfc09675 100644
--- a/src/airbyte/models/shared/monday.py
+++ b/src/airbyte_api/models/shared/monday.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class MondayCredentials:
r"""The Client ID of your OAuth application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your OAuth application."""
-
+
@@ -22,5 +22,3 @@ class MondayCredentials:
@dataclasses.dataclass
class Monday:
credentials: Optional[MondayCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/namespacedefinitionenum.py b/src/airbyte_api/models/shared/namespacedefinitionenum.py
similarity index 100%
rename from src/airbyte/models/shared/namespacedefinitionenum.py
rename to src/airbyte_api/models/shared/namespacedefinitionenum.py
diff --git a/src/airbyte/models/shared/namespacedefinitionenumnodefault.py b/src/airbyte_api/models/shared/namespacedefinitionenumnodefault.py
similarity index 100%
rename from src/airbyte/models/shared/namespacedefinitionenumnodefault.py
rename to src/airbyte_api/models/shared/namespacedefinitionenumnodefault.py
diff --git a/src/airbyte/models/shared/nonbreakingschemaupdatesbehaviorenum.py b/src/airbyte_api/models/shared/nonbreakingschemaupdatesbehaviorenum.py
similarity index 100%
rename from src/airbyte/models/shared/nonbreakingschemaupdatesbehaviorenum.py
rename to src/airbyte_api/models/shared/nonbreakingschemaupdatesbehaviorenum.py
diff --git a/src/airbyte/models/shared/nonbreakingschemaupdatesbehaviorenumnodefault.py b/src/airbyte_api/models/shared/nonbreakingschemaupdatesbehaviorenumnodefault.py
similarity index 100%
rename from src/airbyte/models/shared/nonbreakingschemaupdatesbehaviorenumnodefault.py
rename to src/airbyte_api/models/shared/nonbreakingschemaupdatesbehaviorenumnodefault.py
diff --git a/src/airbyte/models/shared/notion.py b/src/airbyte_api/models/shared/notion.py
similarity index 97%
rename from src/airbyte/models/shared/notion.py
rename to src/airbyte_api/models/shared/notion.py
index 40c5e8d2..4dc224e2 100644
--- a/src/airbyte/models/shared/notion.py
+++ b/src/airbyte_api/models/shared/notion.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class NotionCredentials:
r"""The Client ID of your Notion integration. See our docs for more information."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Notion integration. See our docs for more information."""
-
+
@@ -22,5 +22,3 @@ class NotionCredentials:
@dataclasses.dataclass
class Notion:
credentials: Optional[NotionCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/oauthactornames.py b/src/airbyte_api/models/shared/oauthactornames.py
similarity index 100%
rename from src/airbyte/models/shared/oauthactornames.py
rename to src/airbyte_api/models/shared/oauthactornames.py
diff --git a/src/airbyte/models/shared/oauthinputconfiguration.py b/src/airbyte_api/models/shared/oauthinputconfiguration.py
similarity index 100%
rename from src/airbyte/models/shared/oauthinputconfiguration.py
rename to src/airbyte_api/models/shared/oauthinputconfiguration.py
diff --git a/src/airbyte/models/shared/pinterest.py b/src/airbyte_api/models/shared/pinterest.py
similarity index 96%
rename from src/airbyte/models/shared/pinterest.py
rename to src/airbyte_api/models/shared/pinterest.py
index 6a1051f4..68933e47 100644
--- a/src/airbyte/models/shared/pinterest.py
+++ b/src/airbyte_api/models/shared/pinterest.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class PinterestCredentials:
r"""The Client ID of your OAuth application"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your OAuth application."""
-
+
@@ -22,5 +22,3 @@ class PinterestCredentials:
@dataclasses.dataclass
class Pinterest:
credentials: Optional[PinterestCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/retently.py b/src/airbyte_api/models/shared/retently.py
similarity index 96%
rename from src/airbyte/models/shared/retently.py
rename to src/airbyte_api/models/shared/retently.py
index cff74e02..0611c768 100644
--- a/src/airbyte/models/shared/retently.py
+++ b/src/airbyte_api/models/shared/retently.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class RetentlyCredentials:
r"""The Client ID of your Retently developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Retently developer application."""
-
+
@@ -22,5 +22,3 @@ class RetentlyCredentials:
@dataclasses.dataclass
class Retently:
credentials: Optional[RetentlyCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/salesforce.py b/src/airbyte_api/models/shared/salesforce.py
similarity index 96%
rename from src/airbyte/models/shared/salesforce.py
rename to src/airbyte_api/models/shared/salesforce.py
index c02e448a..3a013519 100644
--- a/src/airbyte/models/shared/salesforce.py
+++ b/src/airbyte_api/models/shared/salesforce.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,5 +14,3 @@ class Salesforce:
r"""Enter your Salesforce developer application's Client ID"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""Enter your Salesforce developer application's Client secret"""
-
-
diff --git a/src/airbyte/models/shared/scheduletypeenum.py b/src/airbyte_api/models/shared/scheduletypeenum.py
similarity index 100%
rename from src/airbyte/models/shared/scheduletypeenum.py
rename to src/airbyte_api/models/shared/scheduletypeenum.py
diff --git a/src/airbyte/models/shared/scheduletypewithbasicenum.py b/src/airbyte_api/models/shared/scheduletypewithbasicenum.py
similarity index 100%
rename from src/airbyte/models/shared/scheduletypewithbasicenum.py
rename to src/airbyte_api/models/shared/scheduletypewithbasicenum.py
diff --git a/src/airbyte/models/shared/schemebasicauth.py b/src/airbyte_api/models/shared/schemebasicauth.py
similarity index 100%
rename from src/airbyte/models/shared/schemebasicauth.py
rename to src/airbyte_api/models/shared/schemebasicauth.py
diff --git a/src/airbyte/models/shared/security.py b/src/airbyte_api/models/shared/security.py
similarity index 100%
rename from src/airbyte/models/shared/security.py
rename to src/airbyte_api/models/shared/security.py
diff --git a/src/airbyte/models/shared/shopify.py b/src/airbyte_api/models/shared/shopify.py
similarity index 96%
rename from src/airbyte/models/shared/shopify.py
rename to src/airbyte_api/models/shared/shopify.py
index d069ed61..e15a6c28 100644
--- a/src/airbyte/models/shared/shopify.py
+++ b/src/airbyte_api/models/shared/shopify.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class ShopifyCredentials:
r"""The Client ID of the Shopify developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of the Shopify developer application."""
-
+
@@ -22,5 +22,3 @@ class ShopifyCredentials:
@dataclasses.dataclass
class Shopify:
credentials: Optional[ShopifyCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/slack.py b/src/airbyte_api/models/shared/slack.py
similarity index 97%
rename from src/airbyte/models/shared/slack.py
rename to src/airbyte_api/models/shared/slack.py
index 0a93620f..7774acbd 100644
--- a/src/airbyte/models/shared/slack.py
+++ b/src/airbyte_api/models/shared/slack.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class SlackCredentials:
r"""Slack client_id. See our docs if you need help finding this id."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""Slack client_secret. See our docs if you need help finding this secret."""
-
+
@@ -22,5 +22,3 @@ class SlackCredentials:
@dataclasses.dataclass
class Slack:
credentials: Optional[SlackCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/smartsheets.py b/src/airbyte_api/models/shared/smartsheets.py
similarity index 96%
rename from src/airbyte/models/shared/smartsheets.py
rename to src/airbyte_api/models/shared/smartsheets.py
index 42c78d90..afaf5670 100644
--- a/src/airbyte/models/shared/smartsheets.py
+++ b/src/airbyte_api/models/shared/smartsheets.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class SmartsheetsCredentials:
r"""The API ID of the SmartSheets developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The API Secret the SmartSheets developer application."""
-
+
@@ -22,5 +22,3 @@ class SmartsheetsCredentials:
@dataclasses.dataclass
class Smartsheets:
credentials: Optional[SmartsheetsCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/snapchat_marketing.py b/src/airbyte_api/models/shared/snapchat_marketing.py
similarity index 96%
rename from src/airbyte/models/shared/snapchat_marketing.py
rename to src/airbyte_api/models/shared/snapchat_marketing.py
index ae5359e4..f0aa407e 100644
--- a/src/airbyte/models/shared/snapchat_marketing.py
+++ b/src/airbyte_api/models/shared/snapchat_marketing.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,5 +14,3 @@ class SnapchatMarketing:
r"""The Client ID of your Snapchat developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Snapchat developer application."""
-
-
diff --git a/src/airbyte/models/shared/snowflake.py b/src/airbyte_api/models/shared/snowflake.py
similarity index 96%
rename from src/airbyte/models/shared/snowflake.py
rename to src/airbyte_api/models/shared/snowflake.py
index ba8ac89c..f4c2874a 100644
--- a/src/airbyte/models/shared/snowflake.py
+++ b/src/airbyte_api/models/shared/snowflake.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class SnowflakeCredentials:
r"""The Client ID of your Snowflake developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Snowflake developer application."""
-
+
@@ -22,5 +22,3 @@ class SnowflakeCredentials:
@dataclasses.dataclass
class Snowflake:
credentials: Optional[SnowflakeCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/source_aha.py b/src/airbyte_api/models/shared/source_aha.py
similarity index 96%
rename from src/airbyte/models/shared/source_aha.py
rename to src/airbyte_api/models/shared/source_aha.py
index 7844b66a..8161ad67 100644
--- a/src/airbyte/models/shared/source_aha.py
+++ b/src/airbyte_api/models/shared/source_aha.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class SourceAha:
url: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('url') }})
r"""URL"""
SOURCE_TYPE: Final[Aha] = dataclasses.field(default=Aha.AHA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_aircall.py b/src/airbyte_api/models/shared/source_aircall.py
similarity index 97%
rename from src/airbyte/models/shared/source_aircall.py
rename to src/airbyte_api/models/shared/source_aircall.py
index 7022cbc7..fa0a4386 100644
--- a/src/airbyte/models/shared/source_aircall.py
+++ b/src/airbyte_api/models/shared/source_aircall.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -23,5 +23,3 @@ class SourceAircall:
start_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""Date time filter for incremental filter, Specify which date to extract from."""
SOURCE_TYPE: Final[Aircall] = dataclasses.field(default=Aircall.AIRCALL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_airtable.py b/src/airbyte_api/models/shared/source_airtable.py
similarity index 98%
rename from src/airbyte/models/shared/source_airtable.py
rename to src/airbyte_api/models/shared/source_airtable.py
index c4de0090..f122c2a8 100644
--- a/src/airbyte/models/shared/source_airtable.py
+++ b/src/airbyte_api/models/shared/source_airtable.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class PersonalAccessToken:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""The Personal Access Token for the Airtable account. See the Support Guide for more information on how to obtain this token."""
AUTH_METHOD: Final[Optional[SourceAirtableAuthMethod]] = dataclasses.field(default=SourceAirtableAuthMethod.API_KEY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method'), 'exclude': lambda f: f is None }})
-
+
class SourceAirtableSchemasAuthMethod(str, Enum):
@@ -40,7 +40,7 @@ class SourceAirtableOAuth20:
AUTH_METHOD: Final[Optional[SourceAirtableSchemasAuthMethod]] = dataclasses.field(default=SourceAirtableSchemasAuthMethod.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method'), 'exclude': lambda f: f is None }})
token_expiry_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token_expiry_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""The date-time when the access token should be refreshed."""
-
+
class SourceAirtableAirtable(str, Enum):
@@ -52,5 +52,3 @@ class SourceAirtableAirtable(str, Enum):
class SourceAirtable:
credentials: Optional[Union[SourceAirtableOAuth20, PersonalAccessToken]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
SOURCE_TYPE: Final[Optional[SourceAirtableAirtable]] = dataclasses.field(default=SourceAirtableAirtable.AIRTABLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/source_amazon_ads.py b/src/airbyte_api/models/shared/source_amazon_ads.py
similarity index 99%
rename from src/airbyte/models/shared/source_amazon_ads.py
rename to src/airbyte_api/models/shared/source_amazon_ads.py
index 0f77ca2b..7f72e1c0 100644
--- a/src/airbyte/models/shared/source_amazon_ads.py
+++ b/src/airbyte_api/models/shared/source_amazon_ads.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -61,5 +61,3 @@ class SourceAmazonAds:
r"""The Start date for collecting reports, should not be more than 60 days in the past. In YYYY-MM-DD format"""
state_filter: Optional[List[StateFilter]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('state_filter'), 'exclude': lambda f: f is None }})
r"""Reflects the state of the Display, Product, and Brand Campaign streams as enabled, paused, or archived. If you do not populate this field, it will be ignored completely."""
-
-
diff --git a/src/airbyte/models/shared/source_amazon_seller_partner.py b/src/airbyte_api/models/shared/source_amazon_seller_partner.py
similarity index 99%
rename from src/airbyte/models/shared/source_amazon_seller_partner.py
rename to src/airbyte_api/models/shared/source_amazon_seller_partner.py
index 103e4cd2..1f431f17 100644
--- a/src/airbyte/models/shared/source_amazon_seller_partner.py
+++ b/src/airbyte_api/models/shared/source_amazon_seller_partner.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -53,7 +53,7 @@ class AWSRegion(str, Enum):
class OptionsList:
option_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('option_name') }})
option_value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('option_value') }})
-
+
class StreamName(str, Enum):
@@ -109,7 +109,7 @@ class ReportOptions:
options_list: List[OptionsList] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('options_list') }})
r"""List of options"""
stream_name: StreamName = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('stream_name') }})
-
+
class SourceAmazonSellerPartnerAmazonSellerPartner(str, Enum):
@@ -141,5 +141,3 @@ class SourceAmazonSellerPartner:
report_options_list: Optional[List[ReportOptions]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('report_options_list'), 'exclude': lambda f: f is None }})
r"""Additional information passed to reports. This varies by report type."""
SOURCE_TYPE: Final[SourceAmazonSellerPartnerAmazonSellerPartner] = dataclasses.field(default=SourceAmazonSellerPartnerAmazonSellerPartner.AMAZON_SELLER_PARTNER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_amazon_sqs.py b/src/airbyte_api/models/shared/source_amazon_sqs.py
similarity index 99%
rename from src/airbyte/models/shared/source_amazon_sqs.py
rename to src/airbyte_api/models/shared/source_amazon_sqs.py
index cebbac94..b5dcb088 100644
--- a/src/airbyte/models/shared/source_amazon_sqs.py
+++ b/src/airbyte_api/models/shared/source_amazon_sqs.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -69,5 +69,3 @@ class SourceAmazonSqs:
SOURCE_TYPE: Final[AmazonSqs] = dataclasses.field(default=AmazonSqs.AMAZON_SQS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
visibility_timeout: Optional[int] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('visibility_timeout'), 'exclude': lambda f: f is None }})
r"""Modify the Visibility Timeout of the individual message from the Queue's default (seconds)."""
-
-
diff --git a/src/airbyte/models/shared/source_amplitude.py b/src/airbyte_api/models/shared/source_amplitude.py
similarity index 98%
rename from src/airbyte/models/shared/source_amplitude.py
rename to src/airbyte_api/models/shared/source_amplitude.py
index dbcd376e..4841e0f5 100644
--- a/src/airbyte/models/shared/source_amplitude.py
+++ b/src/airbyte_api/models/shared/source_amplitude.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -32,5 +32,3 @@ class SourceAmplitude:
request_time_range: Optional[int] = dataclasses.field(default=24, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('request_time_range'), 'exclude': lambda f: f is None }})
r"""According to Considerations too big time range in request can cause a timeout error. In this case, set shorter time interval in hours."""
SOURCE_TYPE: Final[Amplitude] = dataclasses.field(default=Amplitude.AMPLITUDE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_apify_dataset.py b/src/airbyte_api/models/shared/source_apify_dataset.py
similarity index 97%
rename from src/airbyte/models/shared/source_apify_dataset.py
rename to src/airbyte_api/models/shared/source_apify_dataset.py
index 1d806371..b68980ea 100644
--- a/src/airbyte/models/shared/source_apify_dataset.py
+++ b/src/airbyte_api/models/shared/source_apify_dataset.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class SourceApifyDataset:
token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token') }})
r"""Personal API token of your Apify account. In Apify Console, you can find your API token in the Settings section under the Integrations tab after you login. See the Apify Docs for more information."""
SOURCE_TYPE: Final[ApifyDataset] = dataclasses.field(default=ApifyDataset.APIFY_DATASET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_appfollow.py b/src/airbyte_api/models/shared/source_appfollow.py
similarity index 96%
rename from src/airbyte/models/shared/source_appfollow.py
rename to src/airbyte_api/models/shared/source_appfollow.py
index 47662061..301734da 100644
--- a/src/airbyte/models/shared/source_appfollow.py
+++ b/src/airbyte_api/models/shared/source_appfollow.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -17,5 +17,3 @@ class SourceAppfollow:
api_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_secret'), 'exclude': lambda f: f is None }})
r"""API Key provided by Appfollow"""
SOURCE_TYPE: Final[Appfollow] = dataclasses.field(default=Appfollow.APPFOLLOW, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_asana.py b/src/airbyte_api/models/shared/source_asana.py
similarity index 98%
rename from src/airbyte/models/shared/source_asana.py
rename to src/airbyte_api/models/shared/source_asana.py
index e6bcf18b..453ee065 100644
--- a/src/airbyte/models/shared/source_asana.py
+++ b/src/airbyte_api/models/shared/source_asana.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Any, Final, List, Optional, Union
@@ -19,7 +19,7 @@ class AuthenticateWithPersonalAccessToken:
r"""Asana Personal Access Token (generate yours here)."""
OPTION_TITLE: Final[Optional[SourceAsanaSchemasCredentialsTitle]] = dataclasses.field(default=SourceAsanaSchemasCredentialsTitle.PAT_CREDENTIALS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('option_title'), 'exclude': lambda f: f is None }})
r"""PAT Credentials"""
-
+
class SourceAsanaCredentialsTitle(str, Enum):
@@ -35,7 +35,7 @@ class AuthenticateViaAsanaOauth:
refresh_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token') }})
OPTION_TITLE: Final[Optional[SourceAsanaCredentialsTitle]] = dataclasses.field(default=SourceAsanaCredentialsTitle.O_AUTH_CREDENTIALS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('option_title'), 'exclude': lambda f: f is None }})
r"""OAuth Credentials"""
-
+
class SourceAsanaAsana(str, Enum):
@@ -52,5 +52,3 @@ class SourceAsana:
SOURCE_TYPE: Final[Optional[SourceAsanaAsana]] = dataclasses.field(default=SourceAsanaAsana.ASANA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType'), 'exclude': lambda f: f is None }})
test_mode: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('test_mode'), 'exclude': lambda f: f is None }})
r"""This flag is used for testing purposes for certain streams that return a lot of data. This flag is not meant to be enabled for prod."""
-
-
diff --git a/src/airbyte/models/shared/source_auth0.py b/src/airbyte_api/models/shared/source_auth0.py
similarity index 98%
rename from src/airbyte/models/shared/source_auth0.py
rename to src/airbyte_api/models/shared/source_auth0.py
index 54f0b7d5..116fa956 100644
--- a/src/airbyte/models/shared/source_auth0.py
+++ b/src/airbyte_api/models/shared/source_auth0.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -17,7 +17,7 @@ class OAuth2AccessToken:
access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token') }})
r"""Also called API Access Token The access token used to call the Auth0 Management API Token. It's a JWT that contains specific grant permissions knowns as scopes."""
AUTH_TYPE: Final[SourceAuth0SchemasCredentialsAuthenticationMethod] = dataclasses.field(default=SourceAuth0SchemasCredentialsAuthenticationMethod.OAUTH2_ACCESS_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceAuth0SchemasAuthenticationMethod(str, Enum):
@@ -34,7 +34,7 @@ class OAuth2ConfidentialApplication:
client_secret: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret') }})
r"""Your application's Client Secret. You can find this value on the application's settings tab after you login the admin portal."""
AUTH_TYPE: Final[SourceAuth0SchemasAuthenticationMethod] = dataclasses.field(default=SourceAuth0SchemasAuthenticationMethod.OAUTH2_CONFIDENTIAL_APPLICATION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class Auth0(str, Enum):
@@ -50,5 +50,3 @@ class SourceAuth0:
SOURCE_TYPE: Final[Auth0] = dataclasses.field(default=Auth0.AUTH0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[str] = dataclasses.field(default='2023-08-05T00:43:59.244Z', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00Z. Any data before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_aws_cloudtrail.py b/src/airbyte_api/models/shared/source_aws_cloudtrail.py
similarity index 98%
rename from src/airbyte/models/shared/source_aws_cloudtrail.py
rename to src/airbyte_api/models/shared/source_aws_cloudtrail.py
index cb153094..e37dacd5 100644
--- a/src/airbyte/models/shared/source_aws_cloudtrail.py
+++ b/src/airbyte_api/models/shared/source_aws_cloudtrail.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -25,5 +25,3 @@ class SourceAwsCloudtrail:
SOURCE_TYPE: Final[AwsCloudtrail] = dataclasses.field(default=AwsCloudtrail.AWS_CLOUDTRAIL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[date] = dataclasses.field(default=dateutil.parser.parse('1970-01-01').date(), metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""The date you would like to replicate data. Data in AWS CloudTrail is available for last 90 days only. Format: YYYY-MM-DD."""
-
-
diff --git a/src/airbyte/models/shared/source_azure_blob_storage.py b/src/airbyte_api/models/shared/source_azure_blob_storage.py
similarity index 99%
rename from src/airbyte/models/shared/source_azure_blob_storage.py
rename to src/airbyte_api/models/shared/source_azure_blob_storage.py
index 5bb2d78a..89ceb0f3 100644
--- a/src/airbyte/models/shared/source_azure_blob_storage.py
+++ b/src/airbyte_api/models/shared/source_azure_blob_storage.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -24,7 +24,7 @@ class SourceAzureBlobStorageMode(str, Enum):
class Local:
r"""Process files locally, supporting `fast` and `ocr` modes. This is the default option."""
MODE: Final[Optional[SourceAzureBlobStorageMode]] = dataclasses.field(default=SourceAzureBlobStorageMode.LOCAL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class ParsingStrategy(str, Enum):
@@ -46,7 +46,7 @@ class DocumentFileTypeFormatExperimental:
r"""If true, skip files that cannot be parsed and pass the error message along as the _ab_source_file_parse_error field. If false, fail the sync."""
strategy: Optional[ParsingStrategy] = dataclasses.field(default=ParsingStrategy.AUTO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('strategy'), 'exclude': lambda f: f is None }})
r"""The strategy used to parse documents. `fast` extracts text directly from the document which doesn't work for all files. `ocr_only` is more reliable, but slower. `hi_res` is the most reliable, but requires an API key and a hosted instance of unstructured and can't be used with local mode. See the unstructured.io documentation for more details: https://unstructured-io.github.io/unstructured/core/partition.html#partition-pdf"""
-
+
class SourceAzureBlobStorageSchemasStreamsFiletype(str, Enum):
@@ -59,7 +59,7 @@ class ParquetFormat:
decimal_as_float: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('decimal_as_float'), 'exclude': lambda f: f is None }})
r"""Whether to convert decimal fields to floats. There is a loss of precision when converting decimals to floats, so this is not recommended."""
FILETYPE: Final[Optional[SourceAzureBlobStorageSchemasStreamsFiletype]] = dataclasses.field(default=SourceAzureBlobStorageSchemasStreamsFiletype.PARQUET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceAzureBlobStorageSchemasFiletype(str, Enum):
@@ -70,7 +70,7 @@ class SourceAzureBlobStorageSchemasFiletype(str, Enum):
@dataclasses.dataclass
class JsonlFormat:
FILETYPE: Final[Optional[SourceAzureBlobStorageSchemasFiletype]] = dataclasses.field(default=SourceAzureBlobStorageSchemasFiletype.JSONL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceAzureBlobStorageFiletype(str, Enum):
@@ -86,7 +86,7 @@ class UserProvided:
column_names: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('column_names') }})
r"""The column names that will be used while emitting the CSV records"""
HEADER_DEFINITION_TYPE: Final[Optional[SourceAzureBlobStorageSchemasHeaderDefinitionType]] = dataclasses.field(default=SourceAzureBlobStorageSchemasHeaderDefinitionType.USER_PROVIDED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceAzureBlobStorageHeaderDefinitionType(str, Enum):
@@ -97,7 +97,7 @@ class SourceAzureBlobStorageHeaderDefinitionType(str, Enum):
@dataclasses.dataclass
class Autogenerated:
HEADER_DEFINITION_TYPE: Final[Optional[SourceAzureBlobStorageHeaderDefinitionType]] = dataclasses.field(default=SourceAzureBlobStorageHeaderDefinitionType.AUTOGENERATED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class HeaderDefinitionType(str, Enum):
@@ -108,7 +108,7 @@ class HeaderDefinitionType(str, Enum):
@dataclasses.dataclass
class FromCSV:
HEADER_DEFINITION_TYPE: Final[Optional[HeaderDefinitionType]] = dataclasses.field(default=HeaderDefinitionType.FROM_CSV, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class InferenceType(str, Enum):
@@ -147,7 +147,7 @@ class CSVFormat:
r"""Whether strings can be interpreted as null values. If true, strings that match the null_values set will be interpreted as null. If false, strings that match the null_values set will be interpreted as the string itself."""
true_values: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('true_values'), 'exclude': lambda f: f is None }})
r"""A set of case-sensitive strings that should be interpreted as true values."""
-
+
class SourceAzureBlobStorageSchemasStreamsFormatFormatFiletype(str, Enum):
@@ -160,7 +160,7 @@ class AvroFormat:
double_as_string: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('double_as_string'), 'exclude': lambda f: f is None }})
r"""Whether to convert double fields to strings. This is recommended if you have decimal numbers with a high degree of precision because there can be a loss precision when handling floating point numbers."""
FILETYPE: Final[Optional[SourceAzureBlobStorageSchemasStreamsFormatFormatFiletype]] = dataclasses.field(default=SourceAzureBlobStorageSchemasStreamsFormatFormatFiletype.AVRO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class ValidationPolicy(str, Enum):
@@ -191,7 +191,7 @@ class FileBasedStreamConfig:
r"""When enabled, syncs will not validate or structure records against the stream's schema."""
validation_policy: Optional[ValidationPolicy] = dataclasses.field(default=ValidationPolicy.EMIT_RECORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('validation_policy'), 'exclude': lambda f: f is None }})
r"""The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema."""
-
+
@@ -214,5 +214,3 @@ class SourceAzureBlobStorage:
SOURCE_TYPE: Final[SourceAzureBlobStorageAzureBlobStorage] = dataclasses.field(default=SourceAzureBlobStorageAzureBlobStorage.AZURE_BLOB_STORAGE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_azure_table.py b/src/airbyte_api/models/shared/source_azure_table.py
similarity index 97%
rename from src/airbyte/models/shared/source_azure_table.py
rename to src/airbyte_api/models/shared/source_azure_table.py
index 692a4dc2..f89a9fb1 100644
--- a/src/airbyte/models/shared/source_azure_table.py
+++ b/src/airbyte_api/models/shared/source_azure_table.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class SourceAzureTable:
SOURCE_TYPE: Final[AzureTable] = dataclasses.field(default=AzureTable.AZURE_TABLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
storage_endpoint_suffix: Optional[str] = dataclasses.field(default='core.windows.net', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('storage_endpoint_suffix'), 'exclude': lambda f: f is None }})
r"""Azure Table Storage service account URL suffix. See the docs for more information on how to obtain endpoint suffix"""
-
-
diff --git a/src/airbyte/models/shared/source_bamboo_hr.py b/src/airbyte_api/models/shared/source_bamboo_hr.py
similarity index 97%
rename from src/airbyte/models/shared/source_bamboo_hr.py
rename to src/airbyte_api/models/shared/source_bamboo_hr.py
index ff99aba0..34b2a5be 100644
--- a/src/airbyte/models/shared/source_bamboo_hr.py
+++ b/src/airbyte_api/models/shared/source_bamboo_hr.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -23,5 +23,3 @@ class SourceBambooHr:
custom_reports_include_default_fields: Optional[bool] = dataclasses.field(default=True, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('custom_reports_include_default_fields'), 'exclude': lambda f: f is None }})
r"""If true, the custom reports endpoint will include the default fields defined here: https://documentation.bamboohr.com/docs/list-of-field-names."""
SOURCE_TYPE: Final[BambooHr] = dataclasses.field(default=BambooHr.BAMBOO_HR, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_bigquery.py b/src/airbyte_api/models/shared/source_bigquery.py
similarity index 97%
rename from src/airbyte/models/shared/source_bigquery.py
rename to src/airbyte_api/models/shared/source_bigquery.py
index 3b9c9a15..db37cfd8 100644
--- a/src/airbyte/models/shared/source_bigquery.py
+++ b/src/airbyte_api/models/shared/source_bigquery.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class SourceBigquery:
dataset_id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('dataset_id'), 'exclude': lambda f: f is None }})
r"""The dataset ID to search for tables and views. If you are only loading data from one dataset, setting this option could result in much faster schema discovery."""
SOURCE_TYPE: Final[SourceBigqueryBigquery] = dataclasses.field(default=SourceBigqueryBigquery.BIGQUERY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_bing_ads.py b/src/airbyte_api/models/shared/source_bing_ads.py
similarity index 99%
rename from src/airbyte/models/shared/source_bing_ads.py
rename to src/airbyte_api/models/shared/source_bing_ads.py
index dcc5294a..3510f0dd 100644
--- a/src/airbyte/models/shared/source_bing_ads.py
+++ b/src/airbyte_api/models/shared/source_bing_ads.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -22,7 +22,7 @@ class AccountNames:
r"""Account Name is a string value for comparing with the specified predicate."""
operator: Operator = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('operator') }})
r"""An Operator that will be used to filter accounts. The Contains predicate has features for matching words, matching inflectional forms of words, searching using wildcard characters, and searching using proximity. The Equals is used to return all rows where account name is equal(=) to the string that you provided"""
-
+
class AuthMethod(str, Enum):
@@ -77,7 +77,7 @@ class CustomReportConfig:
r"""The name of the the object derives from the ReportRequest object. You can find it in Bing Ads Api docs - Reporting API - Reporting Data Objects."""
report_aggregation: Optional[str] = dataclasses.field(default='[Hourly]', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('report_aggregation'), 'exclude': lambda f: f is None }})
r"""A list of available aggregations."""
-
+
class SourceBingAdsBingAds(str, Enum):
@@ -107,5 +107,3 @@ class SourceBingAds:
SOURCE_TYPE: Final[SourceBingAdsBingAds] = dataclasses.field(default=SourceBingAdsBingAds.BING_ADS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
tenant_id: Optional[str] = dataclasses.field(default='common', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tenant_id'), 'exclude': lambda f: f is None }})
r"""The Tenant ID of your Microsoft Advertising developer application. Set this to \\"common\\" unless you know you need a different value."""
-
-
diff --git a/src/airbyte/models/shared/source_braintree.py b/src/airbyte_api/models/shared/source_braintree.py
similarity index 98%
rename from src/airbyte/models/shared/source_braintree.py
rename to src/airbyte_api/models/shared/source_braintree.py
index 3c9d4b0c..1a111d34 100644
--- a/src/airbyte/models/shared/source_braintree.py
+++ b/src/airbyte_api/models/shared/source_braintree.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -34,5 +34,3 @@ class SourceBraintree:
SOURCE_TYPE: Final[Braintree] = dataclasses.field(default=Braintree.BRAINTREE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00Z. Any data before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_braze.py b/src/airbyte_api/models/shared/source_braze.py
similarity index 97%
rename from src/airbyte/models/shared/source_braze.py
rename to src/airbyte_api/models/shared/source_braze.py
index 57095d67..2543aa58 100644
--- a/src/airbyte/models/shared/source_braze.py
+++ b/src/airbyte_api/models/shared/source_braze.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -22,5 +22,3 @@ class SourceBraze:
url: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('url') }})
r"""Braze REST API endpoint"""
SOURCE_TYPE: Final[Braze] = dataclasses.field(default=Braze.BRAZE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_cart.py b/src/airbyte_api/models/shared/source_cart.py
similarity index 98%
rename from src/airbyte/models/shared/source_cart.py
rename to src/airbyte_api/models/shared/source_cart.py
index 84e0c84b..dae3efe9 100644
--- a/src/airbyte/models/shared/source_cart.py
+++ b/src/airbyte_api/models/shared/source_cart.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -19,7 +19,7 @@ class SingleStoreAccessToken:
store_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('store_name') }})
r"""The name of Cart.com Online Store. All API URLs start with https://[mystorename.com]/api/v1/, where [mystorename.com] is the domain name of your store."""
AUTH_TYPE: Final[SourceCartSchemasAuthType] = dataclasses.field(default=SourceCartSchemasAuthType.SINGLE_STORE_ACCESS_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceCartAuthType(str, Enum):
@@ -36,7 +36,7 @@ class CentralAPIRouter:
user_secret: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('user_secret') }})
r"""Enter your application's User Secret"""
AUTH_TYPE: Final[SourceCartAuthType] = dataclasses.field(default=SourceCartAuthType.CENTRAL_API_ROUTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class Cart(str, Enum):
@@ -50,5 +50,3 @@ class SourceCart:
r"""The date from which you'd like to replicate the data"""
credentials: Optional[Union[CentralAPIRouter, SingleStoreAccessToken]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
SOURCE_TYPE: Final[Cart] = dataclasses.field(default=Cart.CART, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_chargebee.py b/src/airbyte_api/models/shared/source_chargebee.py
similarity index 98%
rename from src/airbyte/models/shared/source_chargebee.py
rename to src/airbyte_api/models/shared/source_chargebee.py
index d5e20a03..07b5e152 100644
--- a/src/airbyte/models/shared/source_chargebee.py
+++ b/src/airbyte_api/models/shared/source_chargebee.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -30,5 +30,3 @@ class SourceChargebee:
product_catalog: Optional[ProductCatalog] = dataclasses.field(default=ProductCatalog.TWO_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('product_catalog'), 'exclude': lambda f: f is None }})
r"""Product Catalog version of your Chargebee site. Instructions on how to find your version you may find here under `API Version` section. If left blank, the product catalog version will be set to 2.0."""
SOURCE_TYPE: Final[Chargebee] = dataclasses.field(default=Chargebee.CHARGEBEE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_chartmogul.py b/src/airbyte_api/models/shared/source_chartmogul.py
similarity index 97%
rename from src/airbyte/models/shared/source_chartmogul.py
rename to src/airbyte_api/models/shared/source_chartmogul.py
index 03978c83..0b44dadf 100644
--- a/src/airbyte/models/shared/source_chartmogul.py
+++ b/src/airbyte_api/models/shared/source_chartmogul.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -21,5 +21,3 @@ class SourceChartmogul:
start_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""UTC date and time in the format 2017-01-25T00:00:00Z. When feasible, any data before this date will not be replicated."""
SOURCE_TYPE: Final[Chartmogul] = dataclasses.field(default=Chartmogul.CHARTMOGUL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_clickhouse.py b/src/airbyte_api/models/shared/source_clickhouse.py
similarity index 99%
rename from src/airbyte/models/shared/source_clickhouse.py
rename to src/airbyte_api/models/shared/source_clickhouse.py
index 5c170552..430b2db1 100644
--- a/src/airbyte/models/shared/source_clickhouse.py
+++ b/src/airbyte_api/models/shared/source_clickhouse.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -28,7 +28,7 @@ class SourceClickhousePasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class SourceClickhouseSchemasTunnelMethod(str, Enum):
@@ -49,7 +49,7 @@ class SourceClickhouseSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class SourceClickhouseTunnelMethod(str, Enum):
@@ -62,7 +62,7 @@ class SourceClickhouseTunnelMethod(str, Enum):
class SourceClickhouseNoTunnel:
TUNNEL_METHOD: Final[SourceClickhouseTunnelMethod] = dataclasses.field(default=SourceClickhouseTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -84,5 +84,3 @@ class SourceClickhouse:
SOURCE_TYPE: Final[SourceClickhouseClickhouse] = dataclasses.field(default=SourceClickhouseClickhouse.CLICKHOUSE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
tunnel_method: Optional[Union[SourceClickhouseNoTunnel, SourceClickhouseSSHKeyAuthentication, SourceClickhousePasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/source_clickup_api.py b/src/airbyte_api/models/shared/source_clickup_api.py
similarity index 98%
rename from src/airbyte/models/shared/source_clickup_api.py
rename to src/airbyte_api/models/shared/source_clickup_api.py
index 43da32ab..db7920cb 100644
--- a/src/airbyte/models/shared/source_clickup_api.py
+++ b/src/airbyte_api/models/shared/source_clickup_api.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -27,5 +27,3 @@ class SourceClickupAPI:
r"""The ID of your space in your workspace. Retrieve it from the `/team/{team_id}/space` of the ClickUp API. See here."""
team_id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('team_id'), 'exclude': lambda f: f is None }})
r"""The ID of your team in ClickUp. Retrieve it from the `/team` of the ClickUp API. See here."""
-
-
diff --git a/src/airbyte/models/shared/source_clockify.py b/src/airbyte_api/models/shared/source_clockify.py
similarity index 97%
rename from src/airbyte/models/shared/source_clockify.py
rename to src/airbyte_api/models/shared/source_clockify.py
index 907c72e1..ae4ac3be 100644
--- a/src/airbyte/models/shared/source_clockify.py
+++ b/src/airbyte_api/models/shared/source_clockify.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class SourceClockify:
api_url: Optional[str] = dataclasses.field(default='https://api.clockify.me', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_url'), 'exclude': lambda f: f is None }})
r"""The URL for the Clockify API. This should only need to be modified if connecting to an enterprise version of Clockify."""
SOURCE_TYPE: Final[Clockify] = dataclasses.field(default=Clockify.CLOCKIFY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_close_com.py b/src/airbyte_api/models/shared/source_close_com.py
similarity index 97%
rename from src/airbyte/models/shared/source_close_com.py
rename to src/airbyte_api/models/shared/source_close_com.py
index 2eabbb23..fea5b074 100644
--- a/src/airbyte/models/shared/source_close_com.py
+++ b/src/airbyte_api/models/shared/source_close_com.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -21,5 +21,3 @@ class SourceCloseCom:
SOURCE_TYPE: Final[CloseCom] = dataclasses.field(default=CloseCom.CLOSE_COM, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[date] = dataclasses.field(default=dateutil.parser.parse('2021-01-01').date(), metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""The start date to sync data; all data after this date will be replicated. Leave blank to retrieve all the data available in the account. Format: YYYY-MM-DD."""
-
-
diff --git a/src/airbyte/models/shared/source_coda.py b/src/airbyte_api/models/shared/source_coda.py
similarity index 95%
rename from src/airbyte/models/shared/source_coda.py
rename to src/airbyte_api/models/shared/source_coda.py
index c9d794d9..a3fef588 100644
--- a/src/airbyte/models/shared/source_coda.py
+++ b/src/airbyte_api/models/shared/source_coda.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceCoda:
auth_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_token') }})
r"""Bearer token"""
SOURCE_TYPE: Final[Coda] = dataclasses.field(default=Coda.CODA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_coin_api.py b/src/airbyte_api/models/shared/source_coin_api.py
similarity index 98%
rename from src/airbyte/models/shared/source_coin_api.py
rename to src/airbyte_api/models/shared/source_coin_api.py
index eb642f2d..81930c88 100644
--- a/src/airbyte/models/shared/source_coin_api.py
+++ b/src/airbyte_api/models/shared/source_coin_api.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -42,5 +42,3 @@ class SourceCoinAPI:
request for pricing purposes. Maximum value is 100000.
"""
SOURCE_TYPE: Final[CoinAPI] = dataclasses.field(default=CoinAPI.COIN_API, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_coinmarketcap.py b/src/airbyte_api/models/shared/source_coinmarketcap.py
similarity index 98%
rename from src/airbyte/models/shared/source_coinmarketcap.py
rename to src/airbyte_api/models/shared/source_coinmarketcap.py
index 8285c346..88c7ec0f 100644
--- a/src/airbyte/models/shared/source_coinmarketcap.py
+++ b/src/airbyte_api/models/shared/source_coinmarketcap.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional
@@ -26,5 +26,3 @@ class SourceCoinmarketcap:
SOURCE_TYPE: Final[Coinmarketcap] = dataclasses.field(default=Coinmarketcap.COINMARKETCAP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
symbols: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('symbols'), 'exclude': lambda f: f is None }})
r"""Cryptocurrency symbols. (only used for quotes stream)"""
-
-
diff --git a/src/airbyte/models/shared/source_configcat.py b/src/airbyte_api/models/shared/source_configcat.py
similarity index 96%
rename from src/airbyte/models/shared/source_configcat.py
rename to src/airbyte_api/models/shared/source_configcat.py
index 00222c28..129c87b1 100644
--- a/src/airbyte/models/shared/source_configcat.py
+++ b/src/airbyte_api/models/shared/source_configcat.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class SourceConfigcat:
username: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('username') }})
r"""Basic auth user name. See here."""
SOURCE_TYPE: Final[Configcat] = dataclasses.field(default=Configcat.CONFIGCAT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_confluence.py b/src/airbyte_api/models/shared/source_confluence.py
similarity index 97%
rename from src/airbyte/models/shared/source_confluence.py
rename to src/airbyte_api/models/shared/source_confluence.py
index 7f995e69..0da8fa17 100644
--- a/src/airbyte/models/shared/source_confluence.py
+++ b/src/airbyte_api/models/shared/source_confluence.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -21,5 +21,3 @@ class SourceConfluence:
email: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('email') }})
r"""Your Confluence login email"""
SOURCE_TYPE: Final[Confluence] = dataclasses.field(default=Confluence.CONFLUENCE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_convex.py b/src/airbyte_api/models/shared/source_convex.py
similarity index 96%
rename from src/airbyte/models/shared/source_convex.py
rename to src/airbyte_api/models/shared/source_convex.py
index adcb24c3..5f88ae85 100644
--- a/src/airbyte/models/shared/source_convex.py
+++ b/src/airbyte_api/models/shared/source_convex.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -18,5 +18,3 @@ class SourceConvex:
r"""API access key used to retrieve data from Convex."""
deployment_url: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('deployment_url') }})
SOURCE_TYPE: Final[SourceConvexConvex] = dataclasses.field(default=SourceConvexConvex.CONVEX, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_datascope.py b/src/airbyte_api/models/shared/source_datascope.py
similarity index 96%
rename from src/airbyte/models/shared/source_datascope.py
rename to src/airbyte_api/models/shared/source_datascope.py
index 976b3448..e7118535 100644
--- a/src/airbyte/models/shared/source_datascope.py
+++ b/src/airbyte_api/models/shared/source_datascope.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class SourceDatascope:
start_date: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date') }})
r"""Start date for the data to be replicated"""
SOURCE_TYPE: Final[Datascope] = dataclasses.field(default=Datascope.DATASCOPE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_delighted.py b/src/airbyte_api/models/shared/source_delighted.py
similarity index 96%
rename from src/airbyte/models/shared/source_delighted.py
rename to src/airbyte_api/models/shared/source_delighted.py
index e152cc7b..9eb90d94 100644
--- a/src/airbyte/models/shared/source_delighted.py
+++ b/src/airbyte_api/models/shared/source_delighted.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -21,5 +21,3 @@ class SourceDelighted:
since: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('since'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""The date from which you'd like to replicate the data"""
SOURCE_TYPE: Final[Delighted] = dataclasses.field(default=Delighted.DELIGHTED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_dixa.py b/src/airbyte_api/models/shared/source_dixa.py
similarity index 97%
rename from src/airbyte/models/shared/source_dixa.py
rename to src/airbyte_api/models/shared/source_dixa.py
index 5e648411..6b623219 100644
--- a/src/airbyte/models/shared/source_dixa.py
+++ b/src/airbyte_api/models/shared/source_dixa.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -23,5 +23,3 @@ class SourceDixa:
batch_size: Optional[int] = dataclasses.field(default=31, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('batch_size'), 'exclude': lambda f: f is None }})
r"""Number of days to batch into one request. Max 31."""
SOURCE_TYPE: Final[Dixa] = dataclasses.field(default=Dixa.DIXA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_dockerhub.py b/src/airbyte_api/models/shared/source_dockerhub.py
similarity index 96%
rename from src/airbyte/models/shared/source_dockerhub.py
rename to src/airbyte_api/models/shared/source_dockerhub.py
index e050b514..b488526b 100644
--- a/src/airbyte/models/shared/source_dockerhub.py
+++ b/src/airbyte_api/models/shared/source_dockerhub.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceDockerhub:
docker_username: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('docker_username') }})
r"""Username of DockerHub person or organization (for https://hub.docker.com/v2/repositories/USERNAME/ API call)"""
SOURCE_TYPE: Final[Dockerhub] = dataclasses.field(default=Dockerhub.DOCKERHUB, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_dremio.py b/src/airbyte_api/models/shared/source_dremio.py
similarity index 96%
rename from src/airbyte/models/shared/source_dremio.py
rename to src/airbyte_api/models/shared/source_dremio.py
index 74bb110b..5fc55284 100644
--- a/src/airbyte/models/shared/source_dremio.py
+++ b/src/airbyte_api/models/shared/source_dremio.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -19,5 +19,3 @@ class SourceDremio:
base_url: Optional[str] = dataclasses.field(default='https://app.dremio.cloud', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('base_url'), 'exclude': lambda f: f is None }})
r"""URL of your Dremio instance"""
SOURCE_TYPE: Final[Dremio] = dataclasses.field(default=Dremio.DREMIO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_dynamodb.py b/src/airbyte_api/models/shared/source_dynamodb.py
similarity index 98%
rename from src/airbyte/models/shared/source_dynamodb.py
rename to src/airbyte_api/models/shared/source_dynamodb.py
index c7da4b76..bb6ade98 100644
--- a/src/airbyte/models/shared/source_dynamodb.py
+++ b/src/airbyte_api/models/shared/source_dynamodb.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -62,5 +62,3 @@ class SourceDynamodb:
reserved_attribute_names: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('reserved_attribute_names'), 'exclude': lambda f: f is None }})
r"""Comma separated reserved attribute names present in your tables"""
SOURCE_TYPE: Final[SourceDynamodbDynamodb] = dataclasses.field(default=SourceDynamodbDynamodb.DYNAMODB, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_e2e_test_cloud.py b/src/airbyte_api/models/shared/source_e2e_test_cloud.py
similarity index 99%
rename from src/airbyte/models/shared/source_e2e_test_cloud.py
rename to src/airbyte_api/models/shared/source_e2e_test_cloud.py
index 25aa79ba..efa9cf41 100644
--- a/src/airbyte/models/shared/source_e2e_test_cloud.py
+++ b/src/airbyte_api/models/shared/source_e2e_test_cloud.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Any, Dict, Final, Optional, Union
@@ -18,7 +18,7 @@ class MultiSchema:
stream_schemas: Optional[str] = dataclasses.field(default='{ "stream1": { "type": "object", "properties": { "field1": { "type": "string" } } }, "stream2": { "type": "object", "properties": { "field1": { "type": "boolean" } } } }', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('stream_schemas'), 'exclude': lambda f: f is None }})
r"""A Json object specifying multiple data streams and their schemas. Each key in this object is one stream name. Each value is the schema for that stream. The schema should be compatible with draft-07. See this doc for examples."""
TYPE: Final[Optional[SourceE2eTestCloudType]] = dataclasses.field(default=SourceE2eTestCloudType.MULTI_STREAM, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('type'), 'exclude': lambda f: f is None }})
-
+
class SourceE2eTestCloudSchemasType(str, Enum):
@@ -36,7 +36,7 @@ class SingleSchema:
stream_schema: Optional[str] = dataclasses.field(default='{ "type": "object", "properties": { "column1": { "type": "string" } } }', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('stream_schema'), 'exclude': lambda f: f is None }})
r"""A Json schema for the stream. The schema should be compatible with draft-07. See this doc for examples."""
TYPE: Final[Optional[SourceE2eTestCloudSchemasType]] = dataclasses.field(default=SourceE2eTestCloudSchemasType.SINGLE_STREAM, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('type'), 'exclude': lambda f: f is None }})
-
+
class E2eTestCloud(str, Enum):
@@ -60,5 +60,3 @@ class ContinuousFeed:
r"""When the seed is unspecified, the current time millis will be used as the seed. Range: [0, 1000000]."""
SOURCE_TYPE: Final[Optional[E2eTestCloud]] = dataclasses.field(default=E2eTestCloud.E2E_TEST_CLOUD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType'), 'exclude': lambda f: f is None }})
TYPE: Final[Optional[Type]] = dataclasses.field(default=Type.CONTINUOUS_FEED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('type'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/source_emailoctopus.py b/src/airbyte_api/models/shared/source_emailoctopus.py
similarity index 96%
rename from src/airbyte/models/shared/source_emailoctopus.py
rename to src/airbyte_api/models/shared/source_emailoctopus.py
index 017ffeaa..9c210277 100644
--- a/src/airbyte/models/shared/source_emailoctopus.py
+++ b/src/airbyte_api/models/shared/source_emailoctopus.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceEmailoctopus:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""EmailOctopus API Key. See the docs for information on how to generate this key."""
SOURCE_TYPE: Final[Emailoctopus] = dataclasses.field(default=Emailoctopus.EMAILOCTOPUS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_exchange_rates.py b/src/airbyte_api/models/shared/source_exchange_rates.py
similarity index 98%
rename from src/airbyte/models/shared/source_exchange_rates.py
rename to src/airbyte_api/models/shared/source_exchange_rates.py
index 821beb0e..fdb8b9f3 100644
--- a/src/airbyte/models/shared/source_exchange_rates.py
+++ b/src/airbyte_api/models/shared/source_exchange_rates.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -24,5 +24,3 @@ class SourceExchangeRates:
ignore_weekends: Optional[bool] = dataclasses.field(default=True, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('ignore_weekends'), 'exclude': lambda f: f is None }})
r"""Ignore weekends? (Exchanges don't run on weekends)"""
SOURCE_TYPE: Final[ExchangeRates] = dataclasses.field(default=ExchangeRates.EXCHANGE_RATES, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_facebook_marketing.py b/src/airbyte_api/models/shared/source_facebook_marketing.py
similarity index 99%
rename from src/airbyte/models/shared/source_facebook_marketing.py
rename to src/airbyte_api/models/shared/source_facebook_marketing.py
index 3f1eace9..69d77cf6 100644
--- a/src/airbyte/models/shared/source_facebook_marketing.py
+++ b/src/airbyte_api/models/shared/source_facebook_marketing.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -232,7 +232,7 @@ class InsightConfig:
r"""The date from which you'd like to replicate data for this stream, in the format YYYY-MM-DDT00:00:00Z."""
time_increment: Optional[int] = dataclasses.field(default=1, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('time_increment'), 'exclude': lambda f: f is None }})
r"""Time window in days by which to aggregate statistics. The sync will be chunked into N day intervals, where N is the number of days you specified. For example, if you set this value to 7, then all statistics will be reported as 7-day aggregates by starting from the start_date. If the start and end dates are October 1st and October 30th, then the connector will output 5 records: 01 - 06, 07 - 13, 14 - 20, 21 - 27, and 28 - 30 (3 days only)."""
-
+
class SourceFacebookMarketingFacebookMarketing(str, Enum):
@@ -269,5 +269,3 @@ class SourceFacebookMarketing:
SOURCE_TYPE: Final[SourceFacebookMarketingFacebookMarketing] = dataclasses.field(default=SourceFacebookMarketingFacebookMarketing.FACEBOOK_MARKETING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""The date from which you'd like to replicate data for all incremental streams, in the format YYYY-MM-DDT00:00:00Z. If not set then all data will be replicated for usual streams and only last 2 years for insight streams."""
-
-
diff --git a/src/airbyte/models/shared/source_faker.py b/src/airbyte_api/models/shared/source_faker.py
similarity index 98%
rename from src/airbyte/models/shared/source_faker.py
rename to src/airbyte_api/models/shared/source_faker.py
index 2d3c6824..75929660 100644
--- a/src/airbyte/models/shared/source_faker.py
+++ b/src/airbyte_api/models/shared/source_faker.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -25,5 +25,3 @@ class SourceFaker:
seed: Optional[int] = dataclasses.field(default=-1, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('seed'), 'exclude': lambda f: f is None }})
r"""Manually control the faker random seed to return the same values on subsequent runs (leave -1 for random)"""
SOURCE_TYPE: Final[Faker] = dataclasses.field(default=Faker.FAKER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_fauna.py b/src/airbyte_api/models/shared/source_fauna.py
similarity index 98%
rename from src/airbyte/models/shared/source_fauna.py
rename to src/airbyte_api/models/shared/source_fauna.py
index bc745522..bf8c07f3 100644
--- a/src/airbyte/models/shared/source_fauna.py
+++ b/src/airbyte_api/models/shared/source_fauna.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -17,7 +17,7 @@ class Enabled:
column: Optional[str] = dataclasses.field(default='deleted_at', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('column'), 'exclude': lambda f: f is None }})
r"""Name of the \\"deleted at\\" column."""
DELETION_MODE: Final[SourceFaunaSchemasDeletionMode] = dataclasses.field(default=SourceFaunaSchemasDeletionMode.DELETED_FIELD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('deletion_mode') }})
-
+
class SourceFaunaDeletionMode(str, Enum):
@@ -28,7 +28,7 @@ class SourceFaunaDeletionMode(str, Enum):
@dataclasses.dataclass
class Disabled:
DELETION_MODE: Final[SourceFaunaDeletionMode] = dataclasses.field(default=SourceFaunaDeletionMode.IGNORE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('deletion_mode') }})
-
+
@@ -47,7 +47,7 @@ class Collection:
Choose your page size based on how large the documents are.
See the docs.
"""
-
+
class Fauna(str, Enum):
@@ -68,5 +68,3 @@ class SourceFauna:
scheme: Optional[str] = dataclasses.field(default='https', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('scheme'), 'exclude': lambda f: f is None }})
r"""URL scheme."""
SOURCE_TYPE: Final[Fauna] = dataclasses.field(default=Fauna.FAUNA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_file.py b/src/airbyte_api/models/shared/source_file.py
similarity index 99%
rename from src/airbyte/models/shared/source_file.py
rename to src/airbyte_api/models/shared/source_file.py
index 2adeddc7..902ce681 100644
--- a/src/airbyte/models/shared/source_file.py
+++ b/src/airbyte_api/models/shared/source_file.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -31,7 +31,7 @@ class SFTPSecureFileTransferProtocol:
password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('password'), 'exclude': lambda f: f is None }})
port: Optional[str] = dataclasses.field(default='22', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('port'), 'exclude': lambda f: f is None }})
STORAGE: Final[SourceFileSchemasProviderStorageProvider7Storage] = dataclasses.field(default=SourceFileSchemasProviderStorageProvider7Storage.SFTP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('storage') }})
-
+
class SourceFileSchemasProviderStorageProvider6Storage(str, Enum):
@@ -46,7 +46,7 @@ class SCPSecureCopyProtocol:
password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('password'), 'exclude': lambda f: f is None }})
port: Optional[str] = dataclasses.field(default='22', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('port'), 'exclude': lambda f: f is None }})
STORAGE: Final[SourceFileSchemasProviderStorageProvider6Storage] = dataclasses.field(default=SourceFileSchemasProviderStorageProvider6Storage.SCP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('storage') }})
-
+
class SourceFileSchemasProviderStorageProviderStorage(str, Enum):
@@ -61,7 +61,7 @@ class SSHSecureShell:
password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('password'), 'exclude': lambda f: f is None }})
port: Optional[str] = dataclasses.field(default='22', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('port'), 'exclude': lambda f: f is None }})
STORAGE: Final[SourceFileSchemasProviderStorageProviderStorage] = dataclasses.field(default=SourceFileSchemasProviderStorageProviderStorage.SSH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('storage') }})
-
+
class SourceFileSchemasProviderStorage(str, Enum):
@@ -78,7 +78,7 @@ class AzBlobAzureBlobStorage:
shared_key: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('shared_key'), 'exclude': lambda f: f is None }})
r"""To access Azure Blob Storage, this connector would need credentials with the proper permissions. One option is a storage account shared key (aka account key or access key). If accessing publicly available data, this field is not necessary."""
STORAGE: Final[SourceFileSchemasProviderStorage] = dataclasses.field(default=SourceFileSchemasProviderStorage.AZ_BLOB, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('storage') }})
-
+
class SourceFileSchemasStorage(str, Enum):
@@ -93,7 +93,7 @@ class SourceFileS3AmazonWebServices:
aws_secret_access_key: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('aws_secret_access_key'), 'exclude': lambda f: f is None }})
r"""In order to access private Buckets stored on AWS S3, this connector would need credentials with the proper permissions. If accessing publicly available data, this field is not necessary."""
STORAGE: Final[SourceFileSchemasStorage] = dataclasses.field(default=SourceFileSchemasStorage.S3, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('storage') }})
-
+
class SourceFileStorage(str, Enum):
@@ -106,7 +106,7 @@ class GCSGoogleCloudStorage:
service_account_json: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('service_account_json'), 'exclude': lambda f: f is None }})
r"""In order to access private Buckets stored on Google Cloud, this connector would need a service account json credentials with the proper permissions as described here. Please generate the credentials.json file and copy/paste its content to this field (expecting JSON formats). If accessing publicly available data, this field is not necessary."""
STORAGE: Final[SourceFileStorage] = dataclasses.field(default=SourceFileStorage.GCS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('storage') }})
-
+
class Storage(str, Enum):
@@ -119,7 +119,7 @@ class HTTPSPublicWeb:
STORAGE: Final[Storage] = dataclasses.field(default=Storage.HTTPS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('storage') }})
user_agent: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('user_agent'), 'exclude': lambda f: f is None }})
r"""Add User-Agent to request"""
-
+
class File(str, Enum):
@@ -140,5 +140,3 @@ class SourceFile:
reader_options: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('reader_options'), 'exclude': lambda f: f is None }})
r"""This should be a string in JSON format. It depends on the chosen file format to provide additional options and tune its behavior."""
SOURCE_TYPE: Final[File] = dataclasses.field(default=File.FILE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_firebolt.py b/src/airbyte_api/models/shared/source_firebolt.py
similarity index 98%
rename from src/airbyte/models/shared/source_firebolt.py
rename to src/airbyte_api/models/shared/source_firebolt.py
index e849f3ea..052d80d2 100644
--- a/src/airbyte/models/shared/source_firebolt.py
+++ b/src/airbyte_api/models/shared/source_firebolt.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -27,5 +27,3 @@ class SourceFirebolt:
host: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('host'), 'exclude': lambda f: f is None }})
r"""The host name of your Firebolt database."""
SOURCE_TYPE: Final[SourceFireboltFirebolt] = dataclasses.field(default=SourceFireboltFirebolt.FIREBOLT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_freshcaller.py b/src/airbyte_api/models/shared/source_freshcaller.py
similarity index 98%
rename from src/airbyte/models/shared/source_freshcaller.py
rename to src/airbyte_api/models/shared/source_freshcaller.py
index a2b47f72..9ce95f10 100644
--- a/src/airbyte/models/shared/source_freshcaller.py
+++ b/src/airbyte_api/models/shared/source_freshcaller.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -27,5 +27,3 @@ class SourceFreshcaller:
r"""UTC date and time. Any data created after this date will be replicated."""
sync_lag_minutes: Optional[int] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sync_lag_minutes'), 'exclude': lambda f: f is None }})
r"""Lag in minutes for each sync, i.e., at time T, data for the time range [prev_sync_time, T-30] will be fetched"""
-
-
diff --git a/src/airbyte/models/shared/source_freshdesk.py b/src/airbyte_api/models/shared/source_freshdesk.py
similarity index 98%
rename from src/airbyte/models/shared/source_freshdesk.py
rename to src/airbyte_api/models/shared/source_freshdesk.py
index d516adfb..6622bd58 100644
--- a/src/airbyte/models/shared/source_freshdesk.py
+++ b/src/airbyte_api/models/shared/source_freshdesk.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,5 +25,3 @@ class SourceFreshdesk:
SOURCE_TYPE: Final[Freshdesk] = dataclasses.field(default=Freshdesk.FRESHDESK, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time. Any data created after this date will be replicated. If this parameter is not set, all data will be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_freshsales.py b/src/airbyte_api/models/shared/source_freshsales.py
similarity index 97%
rename from src/airbyte/models/shared/source_freshsales.py
rename to src/airbyte_api/models/shared/source_freshsales.py
index a29f87ad..28b03fb9 100644
--- a/src/airbyte/models/shared/source_freshsales.py
+++ b/src/airbyte_api/models/shared/source_freshsales.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class SourceFreshsales:
domain_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('domain_name') }})
r"""The Name of your Freshsales domain"""
SOURCE_TYPE: Final[Freshsales] = dataclasses.field(default=Freshsales.FRESHSALES, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_gainsight_px.py b/src/airbyte_api/models/shared/source_gainsight_px.py
similarity index 96%
rename from src/airbyte/models/shared/source_gainsight_px.py
rename to src/airbyte_api/models/shared/source_gainsight_px.py
index 35c0ae43..f014006d 100644
--- a/src/airbyte/models/shared/source_gainsight_px.py
+++ b/src/airbyte_api/models/shared/source_gainsight_px.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceGainsightPx:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""The Aptrinsic API Key which is recieved from the dashboard settings (ref - https://app.aptrinsic.com/settings/api-keys)"""
SOURCE_TYPE: Final[GainsightPx] = dataclasses.field(default=GainsightPx.GAINSIGHT_PX, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_gcs.py b/src/airbyte_api/models/shared/source_gcs.py
similarity index 99%
rename from src/airbyte/models/shared/source_gcs.py
rename to src/airbyte_api/models/shared/source_gcs.py
index ea4100b5..53aac974 100644
--- a/src/airbyte/models/shared/source_gcs.py
+++ b/src/airbyte_api/models/shared/source_gcs.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,7 +25,7 @@ class SourceGcsUserProvided:
column_names: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('column_names') }})
r"""The column names that will be used while emitting the CSV records"""
HEADER_DEFINITION_TYPE: Final[Optional[SourceGcsSchemasStreamsHeaderDefinitionType]] = dataclasses.field(default=SourceGcsSchemasStreamsHeaderDefinitionType.USER_PROVIDED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGcsSchemasHeaderDefinitionType(str, Enum):
@@ -36,7 +36,7 @@ class SourceGcsSchemasHeaderDefinitionType(str, Enum):
@dataclasses.dataclass
class SourceGcsAutogenerated:
HEADER_DEFINITION_TYPE: Final[Optional[SourceGcsSchemasHeaderDefinitionType]] = dataclasses.field(default=SourceGcsSchemasHeaderDefinitionType.AUTOGENERATED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGcsHeaderDefinitionType(str, Enum):
@@ -47,7 +47,7 @@ class SourceGcsHeaderDefinitionType(str, Enum):
@dataclasses.dataclass
class SourceGcsFromCSV:
HEADER_DEFINITION_TYPE: Final[Optional[SourceGcsHeaderDefinitionType]] = dataclasses.field(default=SourceGcsHeaderDefinitionType.FROM_CSV, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGcsInferenceType(str, Enum):
@@ -86,7 +86,7 @@ class SourceGcsCSVFormat:
r"""Whether strings can be interpreted as null values. If true, strings that match the null_values set will be interpreted as null. If false, strings that match the null_values set will be interpreted as the string itself."""
true_values: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('true_values'), 'exclude': lambda f: f is None }})
r"""A set of case-sensitive strings that should be interpreted as true values."""
-
+
class SourceGcsValidationPolicy(str, Enum):
@@ -117,7 +117,7 @@ class SourceGCSStreamConfig:
r"""When enabled, syncs will not validate or structure records against the stream's schema."""
validation_policy: Optional[SourceGcsValidationPolicy] = dataclasses.field(default=SourceGcsValidationPolicy.EMIT_RECORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('validation_policy'), 'exclude': lambda f: f is None }})
r"""The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema."""
-
+
@@ -137,5 +137,3 @@ class SourceGcs:
SOURCE_TYPE: Final[SourceGcsGcs] = dataclasses.field(default=SourceGcsGcs.GCS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_getlago.py b/src/airbyte_api/models/shared/source_getlago.py
similarity index 96%
rename from src/airbyte/models/shared/source_getlago.py
rename to src/airbyte_api/models/shared/source_getlago.py
index 74339911..9da24810 100644
--- a/src/airbyte/models/shared/source_getlago.py
+++ b/src/airbyte_api/models/shared/source_getlago.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -19,5 +19,3 @@ class SourceGetlago:
api_url: Optional[str] = dataclasses.field(default='https://api.getlago.com/api/v1', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_url'), 'exclude': lambda f: f is None }})
r"""Your Lago API URL"""
SOURCE_TYPE: Final[Getlago] = dataclasses.field(default=Getlago.GETLAGO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_github.py b/src/airbyte_api/models/shared/source_github.py
similarity index 99%
rename from src/airbyte/models/shared/source_github.py
rename to src/airbyte_api/models/shared/source_github.py
index 2b8c66ed..c13f3ff1 100644
--- a/src/airbyte/models/shared/source_github.py
+++ b/src/airbyte_api/models/shared/source_github.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class SourceGithubPersonalAccessToken:
personal_access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('personal_access_token') }})
r"""Log into GitHub and then generate a personal access token. To load balance your API quota consumption across multiple API tokens, input multiple tokens separated with \\",\\" """
OPTION_TITLE: Final[Optional[SourceGithubOptionTitle]] = dataclasses.field(default=SourceGithubOptionTitle.PAT_CREDENTIALS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('option_title'), 'exclude': lambda f: f is None }})
-
+
class OptionTitle(str, Enum):
@@ -36,7 +36,7 @@ class OAuth:
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""OAuth Client secret"""
OPTION_TITLE: Final[Optional[OptionTitle]] = dataclasses.field(default=OptionTitle.O_AUTH_CREDENTIALS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('option_title'), 'exclude': lambda f: f is None }})
-
+
class SourceGithubGithub(str, Enum):
@@ -61,5 +61,3 @@ class SourceGithub:
SOURCE_TYPE: Final[SourceGithubGithub] = dataclasses.field(default=SourceGithubGithub.GITHUB, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""The date from which you'd like to replicate data from GitHub in the format YYYY-MM-DDT00:00:00Z. If the date is not set, all data will be replicated. For the streams which support this configuration, only data generated on or after the start date will be replicated. This field doesn't apply to all streams, see the docs for more info"""
-
-
diff --git a/src/airbyte/models/shared/source_gitlab.py b/src/airbyte_api/models/shared/source_gitlab.py
similarity index 99%
rename from src/airbyte/models/shared/source_gitlab.py
rename to src/airbyte_api/models/shared/source_gitlab.py
index 4d4f5d74..dfbeb28c 100644
--- a/src/airbyte/models/shared/source_gitlab.py
+++ b/src/airbyte_api/models/shared/source_gitlab.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class PrivateToken:
access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token') }})
r"""Log into your Gitlab account and then generate a personal Access Token."""
AUTH_TYPE: Final[Optional[SourceGitlabSchemasAuthType]] = dataclasses.field(default=SourceGitlabSchemasAuthType.ACCESS_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGitlabAuthType(str, Enum):
@@ -40,7 +40,7 @@ class SourceGitlabOAuth20:
token_expiry_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token_expiry_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""The date-time when the access token should be refreshed."""
AUTH_TYPE: Final[Optional[SourceGitlabAuthType]] = dataclasses.field(default=SourceGitlabAuthType.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGitlabGitlab(str, Enum):
@@ -64,5 +64,3 @@ class SourceGitlab:
SOURCE_TYPE: Final[SourceGitlabGitlab] = dataclasses.field(default=SourceGitlabGitlab.GITLAB, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""The date from which you'd like to replicate data for GitLab API, in the format YYYY-MM-DDT00:00:00Z. Optional. If not set, all data will be replicated. All data generated after this date will be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_glassfrog.py b/src/airbyte_api/models/shared/source_glassfrog.py
similarity index 95%
rename from src/airbyte/models/shared/source_glassfrog.py
rename to src/airbyte_api/models/shared/source_glassfrog.py
index 6c1a4782..41f7a032 100644
--- a/src/airbyte/models/shared/source_glassfrog.py
+++ b/src/airbyte_api/models/shared/source_glassfrog.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceGlassfrog:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""API key provided by Glassfrog"""
SOURCE_TYPE: Final[Glassfrog] = dataclasses.field(default=Glassfrog.GLASSFROG, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_gnews.py b/src/airbyte_api/models/shared/source_gnews.py
similarity index 95%
rename from src/airbyte/models/shared/source_gnews.py
rename to src/airbyte_api/models/shared/source_gnews.py
index 907d254f..3fa3796e 100644
--- a/src/airbyte/models/shared/source_gnews.py
+++ b/src/airbyte_api/models/shared/source_gnews.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional
@@ -105,14 +105,14 @@ class SourceGnews:
r"""API Key"""
query: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('query') }})
r"""This parameter allows you to specify your search keywords to find the news articles you are looking for. The keywords will be used to return the most relevant articles. It is possible to use logical operators with keywords. - Phrase Search Operator: This operator allows you to make an exact search. Keywords surrounded by
- quotation marks are used to search for articles with the exact same keyword sequence.
+ quotation marks are used to search for articles with the exact same keyword sequence.
For example the query: \"Apple iPhone\" will return articles matching at least once this sequence of keywords.
- Logical AND Operator: This operator allows you to make sure that several keywords are all used in the article
- search. By default the space character acts as an AND operator, it is possible to replace the space character
+ search. By default the space character acts as an AND operator, it is possible to replace the space character
by AND to obtain the same result. For example the query: Apple Microsoft is equivalent to Apple AND Microsoft
- Logical OR Operator: This operator allows you to retrieve articles matching the keyword a or the keyword b.
- It is important to note that this operator has a higher precedence than the AND operator. For example the
- query: Apple OR Microsoft will return all articles matching the keyword Apple as well as all articles matching
+ It is important to note that this operator has a higher precedence than the AND operator. For example the
+ query: Apple OR Microsoft will return all articles matching the keyword Apple as well as all articles matching
the keyword Microsoft
- Logical NOT Operator: This operator allows you to remove from the results the articles corresponding to the
specified keywords. To use it, you need to add NOT in front of each word or phrase surrounded by quotes.
@@ -138,14 +138,14 @@ class SourceGnews:
r"""This parameter allows you to filter the articles that have a publication date greater than or equal to the specified value. The date must respect the following format: YYYY-MM-DD hh:mm:ss (in UTC)"""
top_headlines_query: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('top_headlines_query'), 'exclude': lambda f: f is None }})
r"""This parameter allows you to specify your search keywords to find the news articles you are looking for. The keywords will be used to return the most relevant articles. It is possible to use logical operators with keywords. - Phrase Search Operator: This operator allows you to make an exact search. Keywords surrounded by
- quotation marks are used to search for articles with the exact same keyword sequence.
+ quotation marks are used to search for articles with the exact same keyword sequence.
For example the query: \"Apple iPhone\" will return articles matching at least once this sequence of keywords.
- Logical AND Operator: This operator allows you to make sure that several keywords are all used in the article
- search. By default the space character acts as an AND operator, it is possible to replace the space character
+ search. By default the space character acts as an AND operator, it is possible to replace the space character
by AND to obtain the same result. For example the query: Apple Microsoft is equivalent to Apple AND Microsoft
- Logical OR Operator: This operator allows you to retrieve articles matching the keyword a or the keyword b.
- It is important to note that this operator has a higher precedence than the AND operator. For example the
- query: Apple OR Microsoft will return all articles matching the keyword Apple as well as all articles matching
+ It is important to note that this operator has a higher precedence than the AND operator. For example the
+ query: Apple OR Microsoft will return all articles matching the keyword Apple as well as all articles matching
the keyword Microsoft
- Logical NOT Operator: This operator allows you to remove from the results the articles corresponding to the
specified keywords. To use it, you need to add NOT in front of each word or phrase surrounded by quotes.
@@ -154,5 +154,3 @@ class SourceGnews:
"""
top_headlines_topic: Optional[TopHeadlinesTopic] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('top_headlines_topic'), 'exclude': lambda f: f is None }})
r"""This parameter allows you to change the category for the request."""
-
-
diff --git a/src/airbyte/models/shared/source_google_ads.py b/src/airbyte_api/models/shared/source_google_ads.py
similarity index 99%
rename from src/airbyte/models/shared/source_google_ads.py
rename to src/airbyte_api/models/shared/source_google_ads.py
index 172cccab..abda2b28 100644
--- a/src/airbyte/models/shared/source_google_ads.py
+++ b/src/airbyte_api/models/shared/source_google_ads.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -22,7 +22,7 @@ class GoogleCredentials:
r"""The token used to obtain a new Access Token. For detailed instructions on finding this value, refer to our documentation."""
access_token: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token'), 'exclude': lambda f: f is None }})
r"""The Access Token for making authenticated requests. For detailed instructions on finding this value, refer to our documentation."""
-
+
@@ -33,7 +33,7 @@ class CustomQueriesArray:
r"""A custom defined GAQL query for building the report. Avoid including the segments.date field; wherever possible, Airbyte will automatically include it for incremental syncs. For more information, refer to Google's documentation."""
table_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('table_name') }})
r"""The table name in your destination database for the chosen query."""
-
+
class CustomerStatus(str, Enum):
@@ -64,5 +64,3 @@ class SourceGoogleAds:
SOURCE_TYPE: Final[SourceGoogleAdsGoogleAds] = dataclasses.field(default=SourceGoogleAdsGoogleAds.GOOGLE_ADS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[date] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""UTC date in the format YYYY-MM-DD. Any data before this date will not be replicated. (Default value of two years ago is used if not set)"""
-
-
diff --git a/src/airbyte/models/shared/source_google_analytics_data_api.py b/src/airbyte_api/models/shared/source_google_analytics_data_api.py
similarity index 99%
rename from src/airbyte/models/shared/source_google_analytics_data_api.py
rename to src/airbyte_api/models/shared/source_google_analytics_data_api.py
index 31028839..ca9a091b 100644
--- a/src/airbyte/models/shared/source_google_analytics_data_api.py
+++ b/src/airbyte_api/models/shared/source_google_analytics_data_api.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -18,7 +18,7 @@ class ServiceAccountKeyAuthentication:
credentials_json: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials_json') }})
r"""The JSON key linked to the service account used for authorization. For steps on obtaining this key, refer to the setup guide."""
AUTH_TYPE: Final[Optional[SourceGoogleAnalyticsDataAPISchemasAuthType]] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasAuthType.SERVICE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleAnalyticsDataAPIAuthType(str, Enum):
@@ -37,7 +37,7 @@ class AuthenticateViaGoogleOauth:
access_token: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token'), 'exclude': lambda f: f is None }})
r"""Access Token for making authenticated requests."""
AUTH_TYPE: Final[Optional[SourceGoogleAnalyticsDataAPIAuthType]] = dataclasses.field(default=SourceGoogleAnalyticsDataAPIAuthType.CLIENT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
@@ -47,7 +47,7 @@ class CohortReportSettings:
r"""Optional settings for a cohort report."""
accumulate: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('accumulate'), 'exclude': lambda f: f is None }})
r"""If true, accumulates the result from first touch day to the end day"""
-
+
@@ -56,7 +56,7 @@ class CohortReportSettings:
class DateRange:
end_date: date = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('endDate'), 'encoder': utils.dateisoformat(False), 'decoder': utils.datefromisoformat }})
start_date: date = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('startDate'), 'encoder': utils.dateisoformat(False), 'decoder': utils.datefromisoformat }})
-
+
class Dimension(str, Enum):
@@ -72,7 +72,7 @@ class Cohorts:
r"""Dimension used by the cohort. Required and only supports `firstSessionDate`"""
name: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('name'), 'exclude': lambda f: f is None }})
r"""Assigns a name to this cohort. If not set, cohorts are named by their zero based index cohort_0, cohort_1, etc."""
-
+
class SourceGoogleAnalyticsDataAPIGranularity(str, Enum):
@@ -92,7 +92,7 @@ class CohortsRange:
r"""The granularity used to interpret the startOffset and endOffset for the extended reporting date range for a cohort report."""
start_offset: Optional[int] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('startOffset'), 'exclude': lambda f: f is None }})
r"""Specifies the start date of the extended reporting date range for a cohort report."""
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayEnabled(str, Enum):
@@ -107,7 +107,7 @@ class SourceGoogleAnalyticsDataAPISchemasEnabled:
cohorts: Optional[List[Cohorts]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cohorts'), 'exclude': lambda f: f is None }})
cohorts_range: Optional[CohortsRange] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cohortsRange'), 'exclude': lambda f: f is None }})
ENABLED: Final[Optional[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayEnabled]] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayEnabled.TRUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('enabled'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleAnalyticsDataAPIEnabled(str, Enum):
@@ -118,7 +118,7 @@ class SourceGoogleAnalyticsDataAPIEnabled(str, Enum):
@dataclasses.dataclass
class SourceGoogleAnalyticsDataAPIDisabled:
ENABLED: Final[Optional[SourceGoogleAnalyticsDataAPIEnabled]] = dataclasses.field(default=SourceGoogleAnalyticsDataAPIEnabled.FALSE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('enabled'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayFilterName(str, Enum):
@@ -133,7 +133,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayValueType(str, Enum):
class SourceGoogleAnalyticsDataAPIDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasValueType(str, Enum):
@@ -145,7 +145,7 @@ class SourceGoogleAnalyticsDataAPISchemasValueType(str, Enum):
class SourceGoogleAnalyticsDataAPIInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterValueType(str, Enum):
@@ -157,7 +157,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterValueType(str, Enum):
@@ -169,7 +169,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterValueT
class SourceGoogleAnalyticsDataAPISchemasInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -179,7 +179,7 @@ class BetweenFilter:
from_value: Union[SourceGoogleAnalyticsDataAPIInt64Value, SourceGoogleAnalyticsDataAPIDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('fromValue') }})
to_value: Union[SourceGoogleAnalyticsDataAPISchemasInt64Value, SourceGoogleAnalyticsDataAPISchemasDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('toValue') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayFilterName.BETWEEN_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasFilterName(str, Enum):
@@ -202,7 +202,7 @@ class SourceGoogleAnalyticsDataAPIValueType(str, Enum):
class DoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPIValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPIValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class ValueType(str, Enum):
@@ -214,7 +214,7 @@ class ValueType(str, Enum):
class Int64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[ValueType] = dataclasses.field(default=ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -224,7 +224,7 @@ class NumericFilter:
operation: List[SourceGoogleAnalyticsDataAPISchemasValidEnums] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('operation') }})
value: Union[Int64Value, DoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasFilterName.NUMERIC_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPIFilterName(str, Enum):
@@ -237,7 +237,7 @@ class InListFilter:
values: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('values') }})
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPIFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPIFilterName.IN_LIST_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class FilterName(str, Enum):
@@ -260,7 +260,7 @@ class StringFilter:
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[FilterName] = dataclasses.field(default=FilterName.STRING_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
match_type: Optional[List[SourceGoogleAnalyticsDataAPIValidEnums]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('matchType'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayFilterType(str, Enum):
@@ -274,7 +274,7 @@ class Filter:
field_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('field_name') }})
filter_: Union[StringFilter, InListFilter, NumericFilter, BetweenFilter] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter') }})
FILTER_TYPE: Final[Optional[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayFilterType]] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayFilterType.FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilterFilterName(str, Enum):
@@ -289,7 +289,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilterValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterValueType(str, Enum):
@@ -301,7 +301,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilter4ToValueValueType(str, Enum):
@@ -313,7 +313,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilter4ToValueValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilter4ToValueValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilter4ValueType(str, Enum):
@@ -325,7 +325,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilter4ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilter4ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -335,7 +335,7 @@ class SourceGoogleAnalyticsDataAPISchemasBetweenFilter:
from_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('fromValue') }})
to_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('toValue') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilterFilterName.BETWEEN_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilterName(str, Enum):
@@ -358,7 +358,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3DoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ValueType(str, Enum):
@@ -370,7 +370,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3Int64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -380,7 +380,7 @@ class SourceGoogleAnalyticsDataAPISchemasNumericFilter:
operation: List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ValidEnums] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('operation') }})
value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3Int64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3DoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterFilterName.NUMERIC_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterName(str, Enum):
@@ -393,7 +393,7 @@ class SourceGoogleAnalyticsDataAPISchemasInListFilter:
values: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('values') }})
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3ExpressionFilterName.IN_LIST_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3FilterName(str, Enum):
@@ -416,7 +416,7 @@ class SourceGoogleAnalyticsDataAPISchemasStringFilter:
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3FilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter3FilterName.STRING_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
match_type: Optional[List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterValidEnums]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('matchType'), 'exclude': lambda f: f is None }})
-
+
@@ -425,7 +425,7 @@ class SourceGoogleAnalyticsDataAPISchemasStringFilter:
class SourceGoogleAnalyticsDataAPISchemasExpression:
field_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('field_name') }})
filter_: Union[SourceGoogleAnalyticsDataAPISchemasStringFilter, SourceGoogleAnalyticsDataAPISchemasInListFilter, SourceGoogleAnalyticsDataAPISchemasNumericFilter, SourceGoogleAnalyticsDataAPISchemasBetweenFilter] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasFilterType(str, Enum):
@@ -438,7 +438,7 @@ class NotExpression:
r"""The FilterExpression is NOT of notExpression."""
expression: Optional[SourceGoogleAnalyticsDataAPISchemasExpression] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('expression'), 'exclude': lambda f: f is None }})
FILTER_TYPE: Final[Optional[SourceGoogleAnalyticsDataAPISchemasFilterType]] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasFilterType.NOT_EXPRESSION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterName(str, Enum):
@@ -453,7 +453,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterFilterValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterValueType(str, Enum):
@@ -465,7 +465,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterFilter4ToValueValueType(str, Enum):
@@ -477,7 +477,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterFilter4ToValueValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterFilter4ToValueValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterFilter4ValueType(str, Enum):
@@ -489,7 +489,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterFilter4ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterFilter4ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -499,7 +499,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterBetwee
from_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('fromValue') }})
to_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('toValue') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsFilterName.BETWEEN_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2FilterName(str, Enum):
@@ -522,7 +522,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2DoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ExpressionsValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ValueType(str, Enum):
@@ -534,7 +534,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2Int64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -544,7 +544,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterNumeri
operation: List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterValidEnums] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('operation') }})
value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2Int64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2DoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2FilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2FilterName.NUMERIC_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterFilterName(str, Enum):
@@ -557,7 +557,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterInList
values: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('values') }})
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilterFilterName.IN_LIST_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterFilterName(str, Enum):
@@ -580,7 +580,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterString
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterFilterName.STRING_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
match_type: Optional[List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter2ValidEnums]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('matchType'), 'exclude': lambda f: f is None }})
-
+
@@ -589,7 +589,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterString
class SourceGoogleAnalyticsDataAPIExpression:
field_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('field_name') }})
filter_: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterStringFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterInListFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterNumericFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterBetweenFilter] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter') }})
-
+
class SourceGoogleAnalyticsDataAPIFilterType(str, Enum):
@@ -602,7 +602,7 @@ class OrGroup:
r"""The FilterExpressions in orGroup have an OR relationship."""
expressions: List[SourceGoogleAnalyticsDataAPIExpression] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('expressions') }})
FILTER_TYPE: Final[SourceGoogleAnalyticsDataAPIFilterType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPIFilterType.OR_GROUP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilterFilterName(str, Enum):
@@ -617,7 +617,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilterValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterValueType(str, Enum):
@@ -629,7 +629,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilter4ToValueValueType(str, Enum):
@@ -641,7 +641,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilter4ToValueValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilter4ToValueValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilter4ValueType(str, Enum):
@@ -653,7 +653,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilter4ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilter4ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -663,7 +663,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayBetweenFilter:
from_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('fromValue') }})
to_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('toValue') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilterFilterName.BETWEEN_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilterName(str, Enum):
@@ -686,7 +686,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1DoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ValueType(str, Enum):
@@ -698,7 +698,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimens
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1Int64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -708,7 +708,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayNumericFilter:
operation: List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsValidEnums] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('operation') }})
value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1Int64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1DoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterFilterName.NUMERIC_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterName(str, Enum):
@@ -721,7 +721,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayInListFilter:
values: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('values') }})
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ExpressionsFilterName.IN_LIST_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1FilterName(str, Enum):
@@ -744,7 +744,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayStringFilter:
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1FilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1FilterName.STRING_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
match_type: Optional[List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDimensionFilterDimensionsFilter1ValidEnums]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('matchType'), 'exclude': lambda f: f is None }})
-
+
@@ -753,7 +753,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayStringFilter:
class Expression:
field_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('field_name') }})
filter_: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayStringFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayInListFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayNumericFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayBetweenFilter] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter') }})
-
+
class FilterType(str, Enum):
@@ -766,7 +766,7 @@ class AndGroup:
r"""The FilterExpressions in andGroup have an AND relationship."""
expressions: List[Expression] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('expressions') }})
FILTER_TYPE: Final[FilterType] = dataclasses.field(default=FilterType.AND_GROUP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterFilterName(str, Enum):
@@ -781,7 +781,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4ValueType(str, Enum):
@@ -793,7 +793,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterFilter4ValueType(str, Enum):
@@ -805,7 +805,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterFilter4ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterFilter4ValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterFilterValueType(str, Enum):
@@ -817,7 +817,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterFilterValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -827,7 +827,7 @@ class SourceGoogleAnalyticsDataAPIBetweenFilter:
from_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('fromValue') }})
to_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('toValue') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterFilterName.BETWEEN_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterName(str, Enum):
@@ -850,7 +850,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterValueType(str, Enum):
@@ -862,7 +862,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterValueType
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -872,7 +872,7 @@ class SourceGoogleAnalyticsDataAPINumericFilter:
operation: List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterValidEnums] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('operation') }})
value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterName.NUMERIC_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterFilterName(str, Enum):
@@ -885,7 +885,7 @@ class SourceGoogleAnalyticsDataAPIInListFilter:
values: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('values') }})
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterFilterName.IN_LIST_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterFilterName(str, Enum):
@@ -908,7 +908,7 @@ class SourceGoogleAnalyticsDataAPIStringFilter:
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterFilterName.STRING_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
match_type: Optional[List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayValidEnums]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('matchType'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterType(str, Enum):
@@ -922,7 +922,7 @@ class SourceGoogleAnalyticsDataAPIFilter:
field_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('field_name') }})
filter_: Union[SourceGoogleAnalyticsDataAPIStringFilter, SourceGoogleAnalyticsDataAPIInListFilter, SourceGoogleAnalyticsDataAPINumericFilter, SourceGoogleAnalyticsDataAPIBetweenFilter] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter') }})
FILTER_TYPE: Final[Optional[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterType]] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter4FilterType.FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilterFilterName(str, Enum):
@@ -937,7 +937,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilterValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterValueType(str, Enum):
@@ -949,7 +949,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilter4ToValueValueType(str, Enum):
@@ -961,7 +961,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilter4ToValueValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilter4ToValueValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilter4ValueType(str, Enum):
@@ -973,7 +973,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilter4ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilter4ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -983,7 +983,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
from_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('fromValue') }})
to_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('toValue') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilterFilterName.BETWEEN_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilterName(str, Enum):
@@ -1006,7 +1006,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3DoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ValueType(str, Enum):
@@ -1018,7 +1018,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3Int64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -1028,7 +1028,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
operation: List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionValidEnums] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('operation') }})
value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3Int64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3DoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterFilterName.NUMERIC_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterName(str, Enum):
@@ -1041,7 +1041,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
values: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('values') }})
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ExpressionFilterName.IN_LIST_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3FilterName(str, Enum):
@@ -1064,7 +1064,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3FilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3FilterName.STRING_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
match_type: Optional[List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3ValidEnums]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('matchType'), 'exclude': lambda f: f is None }})
-
+
@@ -1073,7 +1073,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterExpression:
field_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('field_name') }})
filter_: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3StringFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3InListFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3NumericFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3BetweenFilter] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3FilterType(str, Enum):
@@ -1086,7 +1086,7 @@ class SourceGoogleAnalyticsDataAPINotExpression:
r"""The FilterExpression is NOT of notExpression."""
expression: Optional[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterExpression] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('expression'), 'exclude': lambda f: f is None }})
FILTER_TYPE: Final[Optional[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3FilterType]] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter3FilterType.NOT_EXPRESSION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilterFilterName(str, Enum):
@@ -1101,7 +1101,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilterValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterValueType(str, Enum):
@@ -1113,7 +1113,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilter4ToValueValueType(str, Enum):
@@ -1125,7 +1125,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilter4ToValueValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilter4ToValueValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilter4ValueType(str, Enum):
@@ -1137,7 +1137,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilter4ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilter4ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -1147,7 +1147,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
from_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('fromValue') }})
to_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('toValue') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilterFilterName.BETWEEN_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilterName(str, Enum):
@@ -1170,7 +1170,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2DoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ValueType(str, Enum):
@@ -1182,7 +1182,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2Int64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -1192,7 +1192,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
operation: List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsValidEnums] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('operation') }})
value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2Int64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2DoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterFilterName.NUMERIC_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterName(str, Enum):
@@ -1205,7 +1205,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
values: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('values') }})
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ExpressionsFilterName.IN_LIST_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2FilterName(str, Enum):
@@ -1228,7 +1228,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2FilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2FilterName.STRING_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
match_type: Optional[List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter2ValidEnums]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('matchType'), 'exclude': lambda f: f is None }})
-
+
@@ -1237,7 +1237,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterExpression:
field_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('field_name') }})
filter_: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterStringFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterInListFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterNumericFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterBetweenFilter] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterFilterType(str, Enum):
@@ -1250,7 +1250,7 @@ class SourceGoogleAnalyticsDataAPIOrGroup:
r"""The FilterExpressions in orGroup have an OR relationship."""
expressions: List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterExpression] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('expressions') }})
FILTER_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterFilterType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterFilterType.OR_GROUP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterName(str, Enum):
@@ -1265,7 +1265,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ValueType(str, Enum):
@@ -1277,7 +1277,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilterValueType(str, Enum):
@@ -1289,7 +1289,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1DoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilterValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterValueType(str, Enum):
@@ -1301,7 +1301,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1Int64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -1311,7 +1311,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterBetweenFi
from_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('fromValue') }})
to_value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1Int64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1DoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('toValue') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterName.BETWEEN_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1FilterName(str, Enum):
@@ -1334,7 +1334,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsDoubleValue:
value: float = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilter3ValueValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilter3ValueValueType.DOUBLE_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilter3ValueType(str, Enum):
@@ -1346,7 +1346,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFi
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsInt64Value:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
VALUE_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilter3ValueType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilter3ValueType.INT64_VALUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value_type') }})
-
+
@@ -1356,7 +1356,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterNumericFi
operation: List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilterValidEnums] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('operation') }})
value: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsInt64Value, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsDoubleValue] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1FilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1FilterName.NUMERIC_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilterFilterName(str, Enum):
@@ -1369,7 +1369,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterInListFil
values: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('values') }})
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilterFilterName.IN_LIST_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilterName(str, Enum):
@@ -1392,7 +1392,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterStringFil
case_sensitive: Optional[bool] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('caseSensitive'), 'exclude': lambda f: f is None }})
FILTER_NAME: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilterName] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ExpressionsFilterFilterName.STRING_FILTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_name') }})
match_type: Optional[List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterMetricsFilter1ValidEnums]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('matchType'), 'exclude': lambda f: f is None }})
-
+
@@ -1401,7 +1401,7 @@ class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterStringFil
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayExpression:
field_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('field_name') }})
filter_: Union[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterStringFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterInListFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterNumericFilter, SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterBetweenFilter] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter') }})
-
+
class SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterFilterType(str, Enum):
@@ -1414,7 +1414,7 @@ class SourceGoogleAnalyticsDataAPIAndGroup:
r"""The FilterExpressions in andGroup have an AND relationship."""
expressions: List[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayExpression] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('expressions') }})
FILTER_TYPE: Final[SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterFilterType] = dataclasses.field(default=SourceGoogleAnalyticsDataAPISchemasCustomReportsArrayMetricFilterFilterType.AND_GROUP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filter_type') }})
-
+
@@ -1433,7 +1433,7 @@ class SourceGoogleAnalyticsDataAPICustomReportConfig:
r"""Dimensions filter"""
metric_filter: Optional[Union[SourceGoogleAnalyticsDataAPIAndGroup, SourceGoogleAnalyticsDataAPIOrGroup, SourceGoogleAnalyticsDataAPINotExpression, SourceGoogleAnalyticsDataAPIFilter]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('metricFilter'), 'exclude': lambda f: f is None }})
r"""Metrics filter"""
-
+
class SourceGoogleAnalyticsDataAPIGoogleAnalyticsDataAPI(str, Enum):
@@ -1458,5 +1458,3 @@ class SourceGoogleAnalyticsDataAPI:
SOURCE_TYPE: Final[SourceGoogleAnalyticsDataAPIGoogleAnalyticsDataAPI] = dataclasses.field(default=SourceGoogleAnalyticsDataAPIGoogleAnalyticsDataAPI.GOOGLE_ANALYTICS_DATA_API, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
window_in_days: Optional[int] = dataclasses.field(default=1, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('window_in_days'), 'exclude': lambda f: f is None }})
r"""The interval in days for each data request made to the Google Analytics API. A larger value speeds up data sync, but increases the chance of data sampling, which may result in inaccuracies. We recommend a value of 1 to minimize sampling, unless speed is an absolute priority over accuracy. Acceptable values range from 1 to 364. Does not apply to custom Cohort reports. More information is available in the documentation."""
-
-
diff --git a/src/airbyte/models/shared/source_google_analytics_v4_service_account_only.py b/src/airbyte_api/models/shared/source_google_analytics_v4_service_account_only.py
similarity index 99%
rename from src/airbyte/models/shared/source_google_analytics_v4_service_account_only.py
rename to src/airbyte_api/models/shared/source_google_analytics_v4_service_account_only.py
index 5199d328..fe4accf0 100644
--- a/src/airbyte/models/shared/source_google_analytics_v4_service_account_only.py
+++ b/src/airbyte_api/models/shared/source_google_analytics_v4_service_account_only.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -18,7 +18,7 @@ class SourceGoogleAnalyticsV4ServiceAccountOnlyServiceAccountKeyAuthentication:
credentials_json: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials_json') }})
r"""The JSON key of the service account to use for authorization"""
AUTH_TYPE: Final[Optional[SourceGoogleAnalyticsV4ServiceAccountOnlyAuthType]] = dataclasses.field(default=SourceGoogleAnalyticsV4ServiceAccountOnlyAuthType.SERVICE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class GoogleAnalyticsV4ServiceAccountOnly(str, Enum):
@@ -41,5 +41,3 @@ class SourceGoogleAnalyticsV4ServiceAccountOnly:
SOURCE_TYPE: Final[GoogleAnalyticsV4ServiceAccountOnly] = dataclasses.field(default=GoogleAnalyticsV4ServiceAccountOnly.GOOGLE_ANALYTICS_V4_SERVICE_ACCOUNT_ONLY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
window_in_days: Optional[int] = dataclasses.field(default=1, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('window_in_days'), 'exclude': lambda f: f is None }})
r"""The time increment used by the connector when requesting data from the Google Analytics API. More information is available in the the docs. The bigger this value is, the faster the sync will be, but the more likely that sampling will be applied to your data, potentially causing inaccuracies in the returned results. We recommend setting this to 1 unless you have a hard requirement to make the sync faster at the expense of accuracy. The minimum allowed value for this field is 1, and the maximum is 364."""
-
-
diff --git a/src/airbyte/models/shared/source_google_directory.py b/src/airbyte_api/models/shared/source_google_directory.py
similarity index 98%
rename from src/airbyte/models/shared/source_google_directory.py
rename to src/airbyte_api/models/shared/source_google_directory.py
index 1808de11..1c853e89 100644
--- a/src/airbyte/models/shared/source_google_directory.py
+++ b/src/airbyte_api/models/shared/source_google_directory.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -22,7 +22,7 @@ class ServiceAccountKey:
r"""The email of the user, which has permissions to access the Google Workspace Admin APIs."""
CREDENTIALS_TITLE: Final[Optional[SourceGoogleDirectorySchemasCredentialsTitle]] = dataclasses.field(default=SourceGoogleDirectorySchemasCredentialsTitle.SERVICE_ACCOUNTS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials_title'), 'exclude': lambda f: f is None }})
r"""Authentication Scenario"""
-
+
class SourceGoogleDirectoryCredentialsTitle(str, Enum):
@@ -42,7 +42,7 @@ class SignInViaGoogleOAuth:
r"""The Token for obtaining a new access token."""
CREDENTIALS_TITLE: Final[Optional[SourceGoogleDirectoryCredentialsTitle]] = dataclasses.field(default=SourceGoogleDirectoryCredentialsTitle.WEB_SERVER_APP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials_title'), 'exclude': lambda f: f is None }})
r"""Authentication Scenario"""
-
+
class GoogleDirectory(str, Enum):
@@ -55,5 +55,3 @@ class SourceGoogleDirectory:
credentials: Optional[Union[SignInViaGoogleOAuth, ServiceAccountKey]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
r"""Google APIs use the OAuth 2.0 protocol for authentication and authorization. The Source supports Web server application and Service accounts scenarios."""
SOURCE_TYPE: Final[GoogleDirectory] = dataclasses.field(default=GoogleDirectory.GOOGLE_DIRECTORY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_google_drive.py b/src/airbyte_api/models/shared/source_google_drive.py
similarity index 99%
rename from src/airbyte/models/shared/source_google_drive.py
rename to src/airbyte_api/models/shared/source_google_drive.py
index 88a03f9c..05bebf6a 100644
--- a/src/airbyte/models/shared/source_google_drive.py
+++ b/src/airbyte_api/models/shared/source_google_drive.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class SourceGoogleDriveServiceAccountKeyAuthentication:
service_account_info: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('service_account_info') }})
r"""The JSON key of the service account to use for authorization. Read more here."""
AUTH_TYPE: Final[Optional[SourceGoogleDriveSchemasAuthType]] = dataclasses.field(default=SourceGoogleDriveSchemasAuthType.SERVICE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleDriveAuthType(str, Enum):
@@ -36,7 +36,7 @@ class SourceGoogleDriveAuthenticateViaGoogleOAuth:
refresh_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token') }})
r"""Refresh Token for the Google Drive API"""
AUTH_TYPE: Final[Optional[SourceGoogleDriveAuthType]] = dataclasses.field(default=SourceGoogleDriveAuthType.CLIENT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleDriveGoogleDrive(str, Enum):
@@ -54,7 +54,7 @@ class SourceGoogleDriveMode(str, Enum):
class SourceGoogleDriveLocal:
r"""Process files locally, supporting `fast` and `ocr` modes. This is the default option."""
MODE: Final[Optional[SourceGoogleDriveMode]] = dataclasses.field(default=SourceGoogleDriveMode.LOCAL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleDriveParsingStrategy(str, Enum):
@@ -76,7 +76,7 @@ class SourceGoogleDriveDocumentFileTypeFormatExperimental:
r"""If true, skip files that cannot be parsed and pass the error message along as the _ab_source_file_parse_error field. If false, fail the sync."""
strategy: Optional[SourceGoogleDriveParsingStrategy] = dataclasses.field(default=SourceGoogleDriveParsingStrategy.AUTO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('strategy'), 'exclude': lambda f: f is None }})
r"""The strategy used to parse documents. `fast` extracts text directly from the document which doesn't work for all files. `ocr_only` is more reliable, but slower. `hi_res` is the most reliable, but requires an API key and a hosted instance of unstructured and can't be used with local mode. See the unstructured.io documentation for more details: https://unstructured-io.github.io/unstructured/core/partition.html#partition-pdf"""
-
+
class SourceGoogleDriveSchemasStreamsFormatFiletype(str, Enum):
@@ -89,7 +89,7 @@ class SourceGoogleDriveParquetFormat:
decimal_as_float: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('decimal_as_float'), 'exclude': lambda f: f is None }})
r"""Whether to convert decimal fields to floats. There is a loss of precision when converting decimals to floats, so this is not recommended."""
FILETYPE: Final[Optional[SourceGoogleDriveSchemasStreamsFormatFiletype]] = dataclasses.field(default=SourceGoogleDriveSchemasStreamsFormatFiletype.PARQUET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleDriveSchemasStreamsFiletype(str, Enum):
@@ -100,7 +100,7 @@ class SourceGoogleDriveSchemasStreamsFiletype(str, Enum):
@dataclasses.dataclass
class SourceGoogleDriveJsonlFormat:
FILETYPE: Final[Optional[SourceGoogleDriveSchemasStreamsFiletype]] = dataclasses.field(default=SourceGoogleDriveSchemasStreamsFiletype.JSONL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleDriveSchemasFiletype(str, Enum):
@@ -116,7 +116,7 @@ class SourceGoogleDriveUserProvided:
column_names: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('column_names') }})
r"""The column names that will be used while emitting the CSV records"""
HEADER_DEFINITION_TYPE: Final[Optional[SourceGoogleDriveSchemasStreamsHeaderDefinitionType]] = dataclasses.field(default=SourceGoogleDriveSchemasStreamsHeaderDefinitionType.USER_PROVIDED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleDriveSchemasHeaderDefinitionType(str, Enum):
@@ -127,7 +127,7 @@ class SourceGoogleDriveSchemasHeaderDefinitionType(str, Enum):
@dataclasses.dataclass
class SourceGoogleDriveAutogenerated:
HEADER_DEFINITION_TYPE: Final[Optional[SourceGoogleDriveSchemasHeaderDefinitionType]] = dataclasses.field(default=SourceGoogleDriveSchemasHeaderDefinitionType.AUTOGENERATED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleDriveHeaderDefinitionType(str, Enum):
@@ -138,7 +138,7 @@ class SourceGoogleDriveHeaderDefinitionType(str, Enum):
@dataclasses.dataclass
class SourceGoogleDriveFromCSV:
HEADER_DEFINITION_TYPE: Final[Optional[SourceGoogleDriveHeaderDefinitionType]] = dataclasses.field(default=SourceGoogleDriveHeaderDefinitionType.FROM_CSV, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
@@ -170,7 +170,7 @@ class SourceGoogleDriveCSVFormat:
r"""Whether strings can be interpreted as null values. If true, strings that match the null_values set will be interpreted as null. If false, strings that match the null_values set will be interpreted as the string itself."""
true_values: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('true_values'), 'exclude': lambda f: f is None }})
r"""A set of case-sensitive strings that should be interpreted as true values."""
-
+
class SourceGoogleDriveFiletype(str, Enum):
@@ -183,7 +183,7 @@ class SourceGoogleDriveAvroFormat:
double_as_string: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('double_as_string'), 'exclude': lambda f: f is None }})
r"""Whether to convert double fields to strings. This is recommended if you have decimal numbers with a high degree of precision because there can be a loss precision when handling floating point numbers."""
FILETYPE: Final[Optional[SourceGoogleDriveFiletype]] = dataclasses.field(default=SourceGoogleDriveFiletype.AVRO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceGoogleDriveValidationPolicy(str, Enum):
@@ -212,7 +212,7 @@ class SourceGoogleDriveFileBasedStreamConfig:
r"""When enabled, syncs will not validate or structure records against the stream's schema."""
validation_policy: Optional[SourceGoogleDriveValidationPolicy] = dataclasses.field(default=SourceGoogleDriveValidationPolicy.EMIT_RECORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('validation_policy'), 'exclude': lambda f: f is None }})
r"""The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema."""
-
+
@@ -231,5 +231,3 @@ class SourceGoogleDrive:
SOURCE_TYPE: Final[SourceGoogleDriveGoogleDrive] = dataclasses.field(default=SourceGoogleDriveGoogleDrive.GOOGLE_DRIVE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_google_pagespeed_insights.py b/src/airbyte_api/models/shared/source_google_pagespeed_insights.py
similarity index 98%
rename from src/airbyte/models/shared/source_google_pagespeed_insights.py
rename to src/airbyte_api/models/shared/source_google_pagespeed_insights.py
index 10f4989c..efacb1db 100644
--- a/src/airbyte/models/shared/source_google_pagespeed_insights.py
+++ b/src/airbyte_api/models/shared/source_google_pagespeed_insights.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional
@@ -34,5 +34,3 @@ class SourceGooglePagespeedInsights:
api_key: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key'), 'exclude': lambda f: f is None }})
r"""Google PageSpeed API Key. See here. The key is optional - however the API is heavily rate limited when using without API Key. Creating and using the API key therefore is recommended. The key is case sensitive."""
SOURCE_TYPE: Final[GooglePagespeedInsights] = dataclasses.field(default=GooglePagespeedInsights.GOOGLE_PAGESPEED_INSIGHTS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_google_search_console.py b/src/airbyte_api/models/shared/source_google_search_console.py
similarity index 99%
rename from src/airbyte/models/shared/source_google_search_console.py
rename to src/airbyte_api/models/shared/source_google_search_console.py
index d195921e..a6f81bbf 100644
--- a/src/airbyte/models/shared/source_google_search_console.py
+++ b/src/airbyte_api/models/shared/source_google_search_console.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -21,7 +21,7 @@ class SourceGoogleSearchConsoleServiceAccountKeyAuthentication:
service_account_info: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('service_account_info') }})
r"""The JSON key of the service account to use for authorization. Read more here."""
AUTH_TYPE: Final[SourceGoogleSearchConsoleSchemasAuthType] = dataclasses.field(default=SourceGoogleSearchConsoleSchemasAuthType.SERVICE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceGoogleSearchConsoleAuthType(str, Enum):
@@ -40,7 +40,7 @@ class SourceGoogleSearchConsoleOAuth:
access_token: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token'), 'exclude': lambda f: f is None }})
r"""Access token for making authenticated requests. Read more here."""
AUTH_TYPE: Final[SourceGoogleSearchConsoleAuthType] = dataclasses.field(default=SourceGoogleSearchConsoleAuthType.CLIENT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceGoogleSearchConsoleValidEnums(str, Enum):
@@ -59,7 +59,7 @@ class SourceGoogleSearchConsoleCustomReportConfig:
r"""A list of available dimensions. Please note, that for technical reasons `date` is the default dimension which will be included in your query whether you specify it or not. Primary key will consist of your custom dimensions and the default dimension along with `site_url` and `search_type`."""
name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('name') }})
r"""The name of the custom report, this name would be used as stream name"""
-
+
class DataFreshness(str, Enum):
@@ -88,5 +88,3 @@ class SourceGoogleSearchConsole:
SOURCE_TYPE: Final[SourceGoogleSearchConsoleGoogleSearchConsole] = dataclasses.field(default=SourceGoogleSearchConsoleGoogleSearchConsole.GOOGLE_SEARCH_CONSOLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[date] = dataclasses.field(default=dateutil.parser.parse('2021-01-01').date(), metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""UTC date in the format YYYY-MM-DD. Any data before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_google_sheets.py b/src/airbyte_api/models/shared/source_google_sheets.py
similarity index 98%
rename from src/airbyte/models/shared/source_google_sheets.py
rename to src/airbyte_api/models/shared/source_google_sheets.py
index df954c0b..56184156 100644
--- a/src/airbyte/models/shared/source_google_sheets.py
+++ b/src/airbyte_api/models/shared/source_google_sheets.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -17,7 +17,7 @@ class SourceGoogleSheetsServiceAccountKeyAuthentication:
service_account_info: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('service_account_info') }})
r"""The JSON key of the service account to use for authorization. Read more here."""
AUTH_TYPE: Final[SourceGoogleSheetsSchemasAuthType] = dataclasses.field(default=SourceGoogleSheetsSchemasAuthType.SERVICE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceGoogleSheetsAuthType(str, Enum):
@@ -34,7 +34,7 @@ class SourceGoogleSheetsAuthenticateViaGoogleOAuth:
refresh_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token') }})
r"""Enter your Google application's refresh token. See Google's documentation for more information."""
AUTH_TYPE: Final[SourceGoogleSheetsAuthType] = dataclasses.field(default=SourceGoogleSheetsAuthType.CLIENT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceGoogleSheetsGoogleSheets(str, Enum):
@@ -51,5 +51,3 @@ class SourceGoogleSheets:
names_conversion: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('names_conversion'), 'exclude': lambda f: f is None }})
r"""Enables the conversion of column names to a standardized, SQL-compliant format. For example, 'My Name' -> 'my_name'. Enable this option if your destination is SQL-based."""
SOURCE_TYPE: Final[SourceGoogleSheetsGoogleSheets] = dataclasses.field(default=SourceGoogleSheetsGoogleSheets.GOOGLE_SHEETS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_google_webfonts.py b/src/airbyte_api/models/shared/source_google_webfonts.py
similarity index 97%
rename from src/airbyte/models/shared/source_google_webfonts.py
rename to src/airbyte_api/models/shared/source_google_webfonts.py
index 2feb1ff7..fc25e714 100644
--- a/src/airbyte/models/shared/source_google_webfonts.py
+++ b/src/airbyte_api/models/shared/source_google_webfonts.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -23,5 +23,3 @@ class SourceGoogleWebfonts:
sort: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sort'), 'exclude': lambda f: f is None }})
r"""Optional, to find how to sort"""
SOURCE_TYPE: Final[GoogleWebfonts] = dataclasses.field(default=GoogleWebfonts.GOOGLE_WEBFONTS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_google_workspace_admin_reports.py b/src/airbyte_api/models/shared/source_google_workspace_admin_reports.py
similarity index 97%
rename from src/airbyte/models/shared/source_google_workspace_admin_reports.py
rename to src/airbyte_api/models/shared/source_google_workspace_admin_reports.py
index 63c4aac1..4fa9d9c4 100644
--- a/src/airbyte/models/shared/source_google_workspace_admin_reports.py
+++ b/src/airbyte_api/models/shared/source_google_workspace_admin_reports.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class SourceGoogleWorkspaceAdminReports:
lookback: Optional[int] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('lookback'), 'exclude': lambda f: f is None }})
r"""Sets the range of time shown in the report. Reports API allows from up to 180 days ago."""
SOURCE_TYPE: Final[GoogleWorkspaceAdminReports] = dataclasses.field(default=GoogleWorkspaceAdminReports.GOOGLE_WORKSPACE_ADMIN_REPORTS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_greenhouse.py b/src/airbyte_api/models/shared/source_greenhouse.py
similarity index 96%
rename from src/airbyte/models/shared/source_greenhouse.py
rename to src/airbyte_api/models/shared/source_greenhouse.py
index f1c8a41f..a8625f49 100644
--- a/src/airbyte/models/shared/source_greenhouse.py
+++ b/src/airbyte_api/models/shared/source_greenhouse.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceGreenhouse:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""Greenhouse API Key. See the docs for more information on how to generate this key."""
SOURCE_TYPE: Final[Greenhouse] = dataclasses.field(default=Greenhouse.GREENHOUSE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_gridly.py b/src/airbyte_api/models/shared/source_gridly.py
similarity index 96%
rename from src/airbyte/models/shared/source_gridly.py
rename to src/airbyte_api/models/shared/source_gridly.py
index 0cbc7333..28d08f54 100644
--- a/src/airbyte/models/shared/source_gridly.py
+++ b/src/airbyte_api/models/shared/source_gridly.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -18,5 +18,3 @@ class SourceGridly:
grid_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('grid_id') }})
r"""ID of a grid, or can be ID of a branch"""
SOURCE_TYPE: Final[Gridly] = dataclasses.field(default=Gridly.GRIDLY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_harvest.py b/src/airbyte_api/models/shared/source_harvest.py
similarity index 98%
rename from src/airbyte/models/shared/source_harvest.py
rename to src/airbyte_api/models/shared/source_harvest.py
index 5bf0e0fb..69adc57b 100644
--- a/src/airbyte/models/shared/source_harvest.py
+++ b/src/airbyte_api/models/shared/source_harvest.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -21,7 +21,7 @@ class SourceHarvestAuthenticateWithPersonalAccessToken:
r"""Log into Harvest and then create new personal access token."""
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
AUTH_TYPE: Final[Optional[SourceHarvestSchemasAuthType]] = dataclasses.field(default=SourceHarvestSchemasAuthType.TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceHarvestAuthType(str, Enum):
@@ -40,7 +40,7 @@ class AuthenticateViaHarvestOAuth:
r"""Refresh Token to renew the expired Access Token."""
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
AUTH_TYPE: Final[Optional[SourceHarvestAuthType]] = dataclasses.field(default=SourceHarvestAuthType.CLIENT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceHarvestHarvest(str, Enum):
@@ -59,5 +59,3 @@ class SourceHarvest:
replication_end_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('replication_end_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00Z. Any data after this date will not be replicated."""
SOURCE_TYPE: Final[SourceHarvestHarvest] = dataclasses.field(default=SourceHarvestHarvest.HARVEST, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_hubplanner.py b/src/airbyte_api/models/shared/source_hubplanner.py
similarity index 96%
rename from src/airbyte/models/shared/source_hubplanner.py
rename to src/airbyte_api/models/shared/source_hubplanner.py
index eb5b5fab..77219dd7 100644
--- a/src/airbyte/models/shared/source_hubplanner.py
+++ b/src/airbyte_api/models/shared/source_hubplanner.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceHubplanner:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""Hubplanner API key. See https://github.com/hubplanner/API#authentication for more details."""
SOURCE_TYPE: Final[Hubplanner] = dataclasses.field(default=Hubplanner.HUBPLANNER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_hubspot.py b/src/airbyte_api/models/shared/source_hubspot.py
similarity index 98%
rename from src/airbyte/models/shared/source_hubspot.py
rename to src/airbyte_api/models/shared/source_hubspot.py
index fd17d02b..3b8f8710 100644
--- a/src/airbyte/models/shared/source_hubspot.py
+++ b/src/airbyte_api/models/shared/source_hubspot.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -21,7 +21,7 @@ class PrivateApp:
r"""HubSpot Access token. See the Hubspot docs if you need help finding this token."""
CREDENTIALS_TITLE: Final[SourceHubspotSchemasAuthType] = dataclasses.field(default=SourceHubspotSchemasAuthType.PRIVATE_APP_CREDENTIALS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials_title') }})
r"""Name of the credentials set"""
-
+
class SourceHubspotAuthType(str, Enum):
@@ -40,7 +40,7 @@ class SourceHubspotOAuth:
r"""Refresh token to renew an expired access token. See the Hubspot docs if you need help finding this token."""
CREDENTIALS_TITLE: Final[SourceHubspotAuthType] = dataclasses.field(default=SourceHubspotAuthType.O_AUTH_CREDENTIALS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials_title') }})
r"""Name of the credentials"""
-
+
class SourceHubspotHubspot(str, Enum):
@@ -57,5 +57,3 @@ class SourceHubspot:
enable_experimental_streams: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('enable_experimental_streams'), 'exclude': lambda f: f is None }})
r"""If enabled then experimental streams become available for sync."""
SOURCE_TYPE: Final[SourceHubspotHubspot] = dataclasses.field(default=SourceHubspotHubspot.HUBSPOT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_insightly.py b/src/airbyte_api/models/shared/source_insightly.py
similarity index 97%
rename from src/airbyte/models/shared/source_insightly.py
rename to src/airbyte_api/models/shared/source_insightly.py
index f5e0eab9..f9c3a1c2 100644
--- a/src/airbyte/models/shared/source_insightly.py
+++ b/src/airbyte_api/models/shared/source_insightly.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -19,5 +19,3 @@ class SourceInsightly:
token: Optional[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token') }})
r"""Your Insightly API token."""
SOURCE_TYPE: Final[Insightly] = dataclasses.field(default=Insightly.INSIGHTLY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_instagram.py b/src/airbyte_api/models/shared/source_instagram.py
similarity index 98%
rename from src/airbyte/models/shared/source_instagram.py
rename to src/airbyte_api/models/shared/source_instagram.py
index e7db1b0c..e3095ea1 100644
--- a/src/airbyte/models/shared/source_instagram.py
+++ b/src/airbyte_api/models/shared/source_instagram.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,5 +25,3 @@ class SourceInstagram:
SOURCE_TYPE: Final[SourceInstagramInstagram] = dataclasses.field(default=SourceInstagramInstagram.INSTAGRAM, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""The date from which you'd like to replicate data for User Insights, in the format YYYY-MM-DDT00:00:00Z. All data generated after this date will be replicated. If left blank, the start date will be set to 2 years before the present date."""
-
-
diff --git a/src/airbyte/models/shared/source_instatus.py b/src/airbyte_api/models/shared/source_instatus.py
similarity index 95%
rename from src/airbyte/models/shared/source_instatus.py
rename to src/airbyte_api/models/shared/source_instatus.py
index e4f52e1c..0abaaaa3 100644
--- a/src/airbyte/models/shared/source_instatus.py
+++ b/src/airbyte_api/models/shared/source_instatus.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceInstatus:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""Instatus REST API key"""
SOURCE_TYPE: Final[Instatus] = dataclasses.field(default=Instatus.INSTATUS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_intercom.py b/src/airbyte_api/models/shared/source_intercom.py
similarity index 98%
rename from src/airbyte/models/shared/source_intercom.py
rename to src/airbyte_api/models/shared/source_intercom.py
index 286b4c1e..e00adfff 100644
--- a/src/airbyte/models/shared/source_intercom.py
+++ b/src/airbyte_api/models/shared/source_intercom.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,5 +25,3 @@ class SourceIntercom:
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""Client Secret for your Intercom application."""
SOURCE_TYPE: Final[SourceIntercomIntercom] = dataclasses.field(default=SourceIntercomIntercom.INTERCOM, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_ip2whois.py b/src/airbyte_api/models/shared/source_ip2whois.py
similarity index 97%
rename from src/airbyte/models/shared/source_ip2whois.py
rename to src/airbyte_api/models/shared/source_ip2whois.py
index 55aa92fd..6a255bbc 100644
--- a/src/airbyte/models/shared/source_ip2whois.py
+++ b/src/airbyte_api/models/shared/source_ip2whois.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -19,5 +19,3 @@ class SourceIp2whois:
domain: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('domain'), 'exclude': lambda f: f is None }})
r"""Domain name. See here."""
SOURCE_TYPE: Final[Optional[Ip2whois]] = dataclasses.field(default=Ip2whois.IP2WHOIS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/source_iterable.py b/src/airbyte_api/models/shared/source_iterable.py
similarity index 97%
rename from src/airbyte/models/shared/source_iterable.py
rename to src/airbyte_api/models/shared/source_iterable.py
index b69c3dd8..ae2790be 100644
--- a/src/airbyte/models/shared/source_iterable.py
+++ b/src/airbyte_api/models/shared/source_iterable.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -21,5 +21,3 @@ class SourceIterable:
start_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""The date from which you'd like to replicate data for Iterable, in the format YYYY-MM-DDT00:00:00Z. All data generated after this date will be replicated."""
SOURCE_TYPE: Final[Iterable] = dataclasses.field(default=Iterable.ITERABLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_jira.py b/src/airbyte_api/models/shared/source_jira.py
similarity index 99%
rename from src/airbyte/models/shared/source_jira.py
rename to src/airbyte_api/models/shared/source_jira.py
index 9fd3e552..d7a381f3 100644
--- a/src/airbyte/models/shared/source_jira.py
+++ b/src/airbyte_api/models/shared/source_jira.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -44,5 +44,3 @@ class SourceJira:
SOURCE_TYPE: Final[Jira] = dataclasses.field(default=Jira.JIRA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""The date from which you want to replicate data from Jira, use the format YYYY-MM-DDT00:00:00Z. Note that this field only applies to certain streams, and only data generated on or after the start date will be replicated. Or leave it empty if you want to replicate all data. For more information, refer to the documentation."""
-
-
diff --git a/src/airbyte/models/shared/source_k6_cloud.py b/src/airbyte_api/models/shared/source_k6_cloud.py
similarity index 96%
rename from src/airbyte/models/shared/source_k6_cloud.py
rename to src/airbyte_api/models/shared/source_k6_cloud.py
index a1a8d444..9e57a594 100644
--- a/src/airbyte/models/shared/source_k6_cloud.py
+++ b/src/airbyte_api/models/shared/source_k6_cloud.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceK6Cloud:
api_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_token') }})
r"""Your API Token. See here. The key is case sensitive."""
SOURCE_TYPE: Final[K6Cloud] = dataclasses.field(default=K6Cloud.K6_CLOUD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_klarna.py b/src/airbyte_api/models/shared/source_klarna.py
similarity index 98%
rename from src/airbyte/models/shared/source_klarna.py
rename to src/airbyte_api/models/shared/source_klarna.py
index 72cfd356..9f3c0b5d 100644
--- a/src/airbyte/models/shared/source_klarna.py
+++ b/src/airbyte_api/models/shared/source_klarna.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -29,5 +29,3 @@ class SourceKlarna:
playground: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('playground'), 'exclude': lambda f: f is None }})
r"""Propertie defining if connector is used against playground or production environment"""
SOURCE_TYPE: Final[Klarna] = dataclasses.field(default=Klarna.KLARNA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_klaviyo.py b/src/airbyte_api/models/shared/source_klaviyo.py
similarity index 97%
rename from src/airbyte/models/shared/source_klaviyo.py
rename to src/airbyte_api/models/shared/source_klaviyo.py
index 89e94728..fe26f9de 100644
--- a/src/airbyte/models/shared/source_klaviyo.py
+++ b/src/airbyte_api/models/shared/source_klaviyo.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -21,5 +21,3 @@ class SourceKlaviyo:
SOURCE_TYPE: Final[Klaviyo] = dataclasses.field(default=Klaviyo.KLAVIYO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00Z. Any data before this date will not be replicated. This field is optional - if not provided, all data will be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_kyve.py b/src/airbyte_api/models/shared/source_kyve.py
similarity index 98%
rename from src/airbyte/models/shared/source_kyve.py
rename to src/airbyte_api/models/shared/source_kyve.py
index 63e43bbb..ca49a29a 100644
--- a/src/airbyte/models/shared/source_kyve.py
+++ b/src/airbyte_api/models/shared/source_kyve.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -25,5 +25,3 @@ class SourceKyve:
SOURCE_TYPE: Final[Kyve] = dataclasses.field(default=Kyve.KYVE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
url_base: Optional[str] = dataclasses.field(default='https://api.kyve.network', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('url_base'), 'exclude': lambda f: f is None }})
r"""URL to the KYVE Chain API."""
-
-
diff --git a/src/airbyte/models/shared/source_launchdarkly.py b/src/airbyte_api/models/shared/source_launchdarkly.py
similarity index 96%
rename from src/airbyte/models/shared/source_launchdarkly.py
rename to src/airbyte_api/models/shared/source_launchdarkly.py
index 29a85f6d..371046b5 100644
--- a/src/airbyte/models/shared/source_launchdarkly.py
+++ b/src/airbyte_api/models/shared/source_launchdarkly.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceLaunchdarkly:
access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token') }})
r"""Your Access token. See here."""
SOURCE_TYPE: Final[Launchdarkly] = dataclasses.field(default=Launchdarkly.LAUNCHDARKLY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_lemlist.py b/src/airbyte_api/models/shared/source_lemlist.py
similarity index 95%
rename from src/airbyte/models/shared/source_lemlist.py
rename to src/airbyte_api/models/shared/source_lemlist.py
index 2f18a84e..36516531 100644
--- a/src/airbyte/models/shared/source_lemlist.py
+++ b/src/airbyte_api/models/shared/source_lemlist.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceLemlist:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""Lemlist API key,"""
SOURCE_TYPE: Final[Lemlist] = dataclasses.field(default=Lemlist.LEMLIST, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_lever_hiring.py b/src/airbyte_api/models/shared/source_lever_hiring.py
similarity index 98%
rename from src/airbyte/models/shared/source_lever_hiring.py
rename to src/airbyte_api/models/shared/source_lever_hiring.py
index 6193e793..d9c27794 100644
--- a/src/airbyte/models/shared/source_lever_hiring.py
+++ b/src/airbyte_api/models/shared/source_lever_hiring.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -17,7 +17,7 @@ class AuthenticateViaLeverAPIKey:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""The Api Key of your Lever Hiring account."""
AUTH_TYPE: Final[Optional[SourceLeverHiringSchemasAuthType]] = dataclasses.field(default=SourceLeverHiringSchemasAuthType.API_KEY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceLeverHiringAuthType(str, Enum):
@@ -34,7 +34,7 @@ class AuthenticateViaLeverOAuth:
r"""The Client ID of your Lever Hiring developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Lever Hiring developer application."""
-
+
class SourceLeverHiringEnvironment(str, Enum):
@@ -56,5 +56,3 @@ class SourceLeverHiring:
environment: Optional[SourceLeverHiringEnvironment] = dataclasses.field(default=SourceLeverHiringEnvironment.SANDBOX, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('environment'), 'exclude': lambda f: f is None }})
r"""The environment in which you'd like to replicate data for Lever. This is used to determine which Lever API endpoint to use."""
SOURCE_TYPE: Final[SourceLeverHiringLeverHiring] = dataclasses.field(default=SourceLeverHiringLeverHiring.LEVER_HIRING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_linkedin_ads.py b/src/airbyte_api/models/shared/source_linkedin_ads.py
similarity index 99%
rename from src/airbyte/models/shared/source_linkedin_ads.py
rename to src/airbyte_api/models/shared/source_linkedin_ads.py
index 82903fbc..0e986dff 100644
--- a/src/airbyte/models/shared/source_linkedin_ads.py
+++ b/src/airbyte_api/models/shared/source_linkedin_ads.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -50,7 +50,7 @@ class AdAnalyticsReportConfiguration:
r"""Choose a category to pivot your analytics report around. This selection will organize your data based on the chosen attribute, allowing you to analyze trends and performance from different perspectives."""
time_granularity: TimeGranularity = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('time_granularity') }})
r"""Choose how to group the data in your report by time. The options are:
- 'ALL': A single result summarizing the entire time range.
- 'DAILY': Group results by each day.
- 'MONTHLY': Group results by each month.
- 'YEARLY': Group results by each year.
Selecting a time grouping helps you analyze trends and patterns over different time periods."""
-
+
class SourceLinkedinAdsSchemasAuthMethod(str, Enum):
@@ -63,7 +63,7 @@ class AccessToken:
access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token') }})
r"""The access token generated for your developer application. Refer to our documentation for more information."""
AUTH_METHOD: Final[Optional[SourceLinkedinAdsSchemasAuthMethod]] = dataclasses.field(default=SourceLinkedinAdsSchemasAuthMethod.ACCESS_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method'), 'exclude': lambda f: f is None }})
-
+
class SourceLinkedinAdsAuthMethod(str, Enum):
@@ -80,7 +80,7 @@ class SourceLinkedinAdsOAuth20:
refresh_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token') }})
r"""The key to refresh the expired access token. Refer to our documentation for more information."""
AUTH_METHOD: Final[Optional[SourceLinkedinAdsAuthMethod]] = dataclasses.field(default=SourceLinkedinAdsAuthMethod.O_AUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method'), 'exclude': lambda f: f is None }})
-
+
class SourceLinkedinAdsLinkedinAds(str, Enum):
@@ -97,5 +97,3 @@ class SourceLinkedinAds:
ad_analytics_reports: Optional[List[AdAnalyticsReportConfiguration]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('ad_analytics_reports'), 'exclude': lambda f: f is None }})
credentials: Optional[Union[SourceLinkedinAdsOAuth20, AccessToken]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
SOURCE_TYPE: Final[SourceLinkedinAdsLinkedinAds] = dataclasses.field(default=SourceLinkedinAdsLinkedinAds.LINKEDIN_ADS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_linkedin_pages.py b/src/airbyte_api/models/shared/source_linkedin_pages.py
similarity index 98%
rename from src/airbyte/models/shared/source_linkedin_pages.py
rename to src/airbyte_api/models/shared/source_linkedin_pages.py
index f9640a4d..a8fac70b 100644
--- a/src/airbyte/models/shared/source_linkedin_pages.py
+++ b/src/airbyte_api/models/shared/source_linkedin_pages.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -17,7 +17,7 @@ class SourceLinkedinPagesAccessToken:
access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token') }})
r"""The token value generated using the LinkedIn Developers OAuth Token Tools. See the docs to obtain yours."""
AUTH_METHOD: Final[Optional[SourceLinkedinPagesSchemasAuthMethod]] = dataclasses.field(default=SourceLinkedinPagesSchemasAuthMethod.ACCESS_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method'), 'exclude': lambda f: f is None }})
-
+
class SourceLinkedinPagesAuthMethod(str, Enum):
@@ -34,7 +34,7 @@ class SourceLinkedinPagesOAuth20:
refresh_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token') }})
r"""The token value generated using the LinkedIn Developers OAuth Token Tools. See the docs to obtain yours."""
AUTH_METHOD: Final[Optional[SourceLinkedinPagesAuthMethod]] = dataclasses.field(default=SourceLinkedinPagesAuthMethod.O_AUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method'), 'exclude': lambda f: f is None }})
-
+
class LinkedinPages(str, Enum):
@@ -48,5 +48,3 @@ class SourceLinkedinPages:
r"""Specify the Organization ID"""
credentials: Optional[Union[SourceLinkedinPagesOAuth20, SourceLinkedinPagesAccessToken]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
SOURCE_TYPE: Final[LinkedinPages] = dataclasses.field(default=LinkedinPages.LINKEDIN_PAGES, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_lokalise.py b/src/airbyte_api/models/shared/source_lokalise.py
similarity index 97%
rename from src/airbyte/models/shared/source_lokalise.py
rename to src/airbyte_api/models/shared/source_lokalise.py
index bc3619fd..d8d0f445 100644
--- a/src/airbyte/models/shared/source_lokalise.py
+++ b/src/airbyte_api/models/shared/source_lokalise.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class SourceLokalise:
project_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('project_id') }})
r"""Lokalise project ID. Available at Project Settings > General."""
SOURCE_TYPE: Final[Lokalise] = dataclasses.field(default=Lokalise.LOKALISE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_mailchimp.py b/src/airbyte_api/models/shared/source_mailchimp.py
similarity index 98%
rename from src/airbyte/models/shared/source_mailchimp.py
rename to src/airbyte_api/models/shared/source_mailchimp.py
index 86d13cb2..a727f4b9 100644
--- a/src/airbyte/models/shared/source_mailchimp.py
+++ b/src/airbyte_api/models/shared/source_mailchimp.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class APIKey:
apikey: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('apikey') }})
r"""Mailchimp API Key. See the docs for information on how to generate this key."""
AUTH_TYPE: Final[SourceMailchimpSchemasAuthType] = dataclasses.field(default=SourceMailchimpSchemasAuthType.APIKEY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceMailchimpAuthType(str, Enum):
@@ -36,7 +36,7 @@ class SourceMailchimpOAuth20:
r"""The Client ID of your OAuth application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your OAuth application."""
-
+
class SourceMailchimpMailchimp(str, Enum):
@@ -51,5 +51,3 @@ class SourceMailchimp:
SOURCE_TYPE: Final[SourceMailchimpMailchimp] = dataclasses.field(default=SourceMailchimpMailchimp.MAILCHIMP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""The date from which you want to start syncing data for Incremental streams. Only records that have been created or modified since this date will be synced. If left blank, all data will by synced."""
-
-
diff --git a/src/airbyte/models/shared/source_mailgun.py b/src/airbyte_api/models/shared/source_mailgun.py
similarity index 97%
rename from src/airbyte/models/shared/source_mailgun.py
rename to src/airbyte_api/models/shared/source_mailgun.py
index 500e7382..20d2c234 100644
--- a/src/airbyte/models/shared/source_mailgun.py
+++ b/src/airbyte_api/models/shared/source_mailgun.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -23,5 +23,3 @@ class SourceMailgun:
SOURCE_TYPE: Final[Mailgun] = dataclasses.field(default=Mailgun.MAILGUN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2020-10-01 00:00:00. Any data before this date will not be replicated. If omitted, defaults to 3 days ago."""
-
-
diff --git a/src/airbyte/models/shared/source_mailjet_sms.py b/src/airbyte_api/models/shared/source_mailjet_sms.py
similarity index 97%
rename from src/airbyte/models/shared/source_mailjet_sms.py
rename to src/airbyte_api/models/shared/source_mailjet_sms.py
index 95efa0ac..ae28b7f9 100644
--- a/src/airbyte/models/shared/source_mailjet_sms.py
+++ b/src/airbyte_api/models/shared/source_mailjet_sms.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class SourceMailjetSms:
SOURCE_TYPE: Final[MailjetSms] = dataclasses.field(default=MailjetSms.MAILJET_SMS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[int] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'exclude': lambda f: f is None }})
r"""Retrieve SMS messages created after the specified timestamp. Required format - Unix timestamp."""
-
-
diff --git a/src/airbyte/models/shared/source_marketo.py b/src/airbyte_api/models/shared/source_marketo.py
similarity index 98%
rename from src/airbyte/models/shared/source_marketo.py
rename to src/airbyte_api/models/shared/source_marketo.py
index 4416a023..37341a75 100644
--- a/src/airbyte/models/shared/source_marketo.py
+++ b/src/airbyte_api/models/shared/source_marketo.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,5 +25,3 @@ class SourceMarketo:
start_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""UTC date and time in the format 2017-01-25T00:00:00Z. Any data before this date will not be replicated."""
SOURCE_TYPE: Final[Marketo] = dataclasses.field(default=Marketo.MARKETO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_metabase.py b/src/airbyte_api/models/shared/source_metabase.py
similarity index 96%
rename from src/airbyte/models/shared/source_metabase.py
rename to src/airbyte_api/models/shared/source_metabase.py
index 6cecc487..897ac3f6 100644
--- a/src/airbyte/models/shared/source_metabase.py
+++ b/src/airbyte_api/models/shared/source_metabase.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -18,7 +18,7 @@ class SourceMetabase:
r"""URL to your metabase instance API"""
password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('password'), 'exclude': lambda f: f is None }})
session_token: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('session_token'), 'exclude': lambda f: f is None }})
- r"""To generate your session token, you need to run the following command: ``` curl -X POST \
+ r"""To generate your session token, you need to run the following command: ``` curl -X POST \
-H \"Content-Type: application/json\" \
-d '{\"username\": \"person@metabase.com\", \"password\": \"fakepassword\"}' \
http://localhost:3000/api/session
@@ -27,5 +27,3 @@ class SourceMetabase:
"""
SOURCE_TYPE: Final[Metabase] = dataclasses.field(default=Metabase.METABASE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
username: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('username'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/source_microsoft_sharepoint.py b/src/airbyte_api/models/shared/source_microsoft_sharepoint.py
similarity index 99%
rename from src/airbyte/models/shared/source_microsoft_sharepoint.py
rename to src/airbyte_api/models/shared/source_microsoft_sharepoint.py
index 346c0a19..ee633cea 100644
--- a/src/airbyte/models/shared/source_microsoft_sharepoint.py
+++ b/src/airbyte_api/models/shared/source_microsoft_sharepoint.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -28,7 +28,7 @@ class ServiceKeyAuthentication:
user_principal_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('user_principal_name') }})
r"""Special characters such as a period, comma, space, and the at sign (@) are converted to underscores (_). More details: https://learn.microsoft.com/en-us/sharepoint/list-onedrive-urls"""
AUTH_TYPE: Final[Optional[SourceMicrosoftSharepointSchemasAuthType]] = dataclasses.field(default=SourceMicrosoftSharepointSchemasAuthType.SERVICE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceMicrosoftSharepointAuthType(str, Enum):
@@ -50,7 +50,7 @@ class AuthenticateViaMicrosoftOAuth:
tenant_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tenant_id') }})
r"""Tenant ID of the Microsoft SharePoint user"""
AUTH_TYPE: Final[Optional[SourceMicrosoftSharepointAuthType]] = dataclasses.field(default=SourceMicrosoftSharepointAuthType.CLIENT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceMicrosoftSharepointMicrosoftSharepoint(str, Enum):
@@ -68,7 +68,7 @@ class SourceMicrosoftSharepointMode(str, Enum):
class SourceMicrosoftSharepointLocal:
r"""Process files locally, supporting `fast` and `ocr` modes. This is the default option."""
MODE: Final[Optional[SourceMicrosoftSharepointMode]] = dataclasses.field(default=SourceMicrosoftSharepointMode.LOCAL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class SourceMicrosoftSharepointParsingStrategy(str, Enum):
@@ -90,7 +90,7 @@ class SourceMicrosoftSharepointDocumentFileTypeFormatExperimental:
r"""If true, skip files that cannot be parsed and pass the error message along as the _ab_source_file_parse_error field. If false, fail the sync."""
strategy: Optional[SourceMicrosoftSharepointParsingStrategy] = dataclasses.field(default=SourceMicrosoftSharepointParsingStrategy.AUTO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('strategy'), 'exclude': lambda f: f is None }})
r"""The strategy used to parse documents. `fast` extracts text directly from the document which doesn't work for all files. `ocr_only` is more reliable, but slower. `hi_res` is the most reliable, but requires an API key and a hosted instance of unstructured and can't be used with local mode. See the unstructured.io documentation for more details: https://unstructured-io.github.io/unstructured/core/partition.html#partition-pdf"""
-
+
class SourceMicrosoftSharepointSchemasStreamsFormatFiletype(str, Enum):
@@ -103,7 +103,7 @@ class SourceMicrosoftSharepointParquetFormat:
decimal_as_float: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('decimal_as_float'), 'exclude': lambda f: f is None }})
r"""Whether to convert decimal fields to floats. There is a loss of precision when converting decimals to floats, so this is not recommended."""
FILETYPE: Final[Optional[SourceMicrosoftSharepointSchemasStreamsFormatFiletype]] = dataclasses.field(default=SourceMicrosoftSharepointSchemasStreamsFormatFiletype.PARQUET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceMicrosoftSharepointSchemasStreamsFiletype(str, Enum):
@@ -114,7 +114,7 @@ class SourceMicrosoftSharepointSchemasStreamsFiletype(str, Enum):
@dataclasses.dataclass
class SourceMicrosoftSharepointJsonlFormat:
FILETYPE: Final[Optional[SourceMicrosoftSharepointSchemasStreamsFiletype]] = dataclasses.field(default=SourceMicrosoftSharepointSchemasStreamsFiletype.JSONL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceMicrosoftSharepointSchemasFiletype(str, Enum):
@@ -130,7 +130,7 @@ class SourceMicrosoftSharepointUserProvided:
column_names: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('column_names') }})
r"""The column names that will be used while emitting the CSV records"""
HEADER_DEFINITION_TYPE: Final[Optional[SourceMicrosoftSharepointSchemasStreamsHeaderDefinitionType]] = dataclasses.field(default=SourceMicrosoftSharepointSchemasStreamsHeaderDefinitionType.USER_PROVIDED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceMicrosoftSharepointSchemasHeaderDefinitionType(str, Enum):
@@ -141,7 +141,7 @@ class SourceMicrosoftSharepointSchemasHeaderDefinitionType(str, Enum):
@dataclasses.dataclass
class SourceMicrosoftSharepointAutogenerated:
HEADER_DEFINITION_TYPE: Final[Optional[SourceMicrosoftSharepointSchemasHeaderDefinitionType]] = dataclasses.field(default=SourceMicrosoftSharepointSchemasHeaderDefinitionType.AUTOGENERATED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceMicrosoftSharepointHeaderDefinitionType(str, Enum):
@@ -152,7 +152,7 @@ class SourceMicrosoftSharepointHeaderDefinitionType(str, Enum):
@dataclasses.dataclass
class SourceMicrosoftSharepointFromCSV:
HEADER_DEFINITION_TYPE: Final[Optional[SourceMicrosoftSharepointHeaderDefinitionType]] = dataclasses.field(default=SourceMicrosoftSharepointHeaderDefinitionType.FROM_CSV, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
@@ -184,7 +184,7 @@ class SourceMicrosoftSharepointCSVFormat:
r"""Whether strings can be interpreted as null values. If true, strings that match the null_values set will be interpreted as null. If false, strings that match the null_values set will be interpreted as the string itself."""
true_values: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('true_values'), 'exclude': lambda f: f is None }})
r"""A set of case-sensitive strings that should be interpreted as true values."""
-
+
class SourceMicrosoftSharepointFiletype(str, Enum):
@@ -197,7 +197,7 @@ class SourceMicrosoftSharepointAvroFormat:
double_as_string: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('double_as_string'), 'exclude': lambda f: f is None }})
r"""Whether to convert double fields to strings. This is recommended if you have decimal numbers with a high degree of precision because there can be a loss precision when handling floating point numbers."""
FILETYPE: Final[Optional[SourceMicrosoftSharepointFiletype]] = dataclasses.field(default=SourceMicrosoftSharepointFiletype.AVRO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceMicrosoftSharepointValidationPolicy(str, Enum):
@@ -226,7 +226,7 @@ class SourceMicrosoftSharepointFileBasedStreamConfig:
r"""When enabled, syncs will not validate or structure records against the stream's schema."""
validation_policy: Optional[SourceMicrosoftSharepointValidationPolicy] = dataclasses.field(default=SourceMicrosoftSharepointValidationPolicy.EMIT_RECORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('validation_policy'), 'exclude': lambda f: f is None }})
r"""The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema."""
-
+
@@ -245,5 +245,3 @@ class SourceMicrosoftSharepoint:
SOURCE_TYPE: Final[SourceMicrosoftSharepointMicrosoftSharepoint] = dataclasses.field(default=SourceMicrosoftSharepointMicrosoftSharepoint.MICROSOFT_SHAREPOINT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_microsoft_teams.py b/src/airbyte_api/models/shared/source_microsoft_teams.py
similarity index 98%
rename from src/airbyte/models/shared/source_microsoft_teams.py
rename to src/airbyte_api/models/shared/source_microsoft_teams.py
index ddcfe8d8..bdf94e52 100644
--- a/src/airbyte/models/shared/source_microsoft_teams.py
+++ b/src/airbyte_api/models/shared/source_microsoft_teams.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -21,7 +21,7 @@ class AuthenticateViaMicrosoft:
tenant_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tenant_id') }})
r"""A globally unique identifier (GUID) that is different than your organization name or domain. Follow these steps to obtain: open one of the Teams where you belong inside the Teams Application -> Click on the … next to the Team title -> Click on Get link to team -> Copy the link to the team and grab the tenant ID form the URL"""
AUTH_TYPE: Final[Optional[SourceMicrosoftTeamsSchemasAuthType]] = dataclasses.field(default=SourceMicrosoftTeamsSchemasAuthType.TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceMicrosoftTeamsAuthType(str, Enum):
@@ -40,7 +40,7 @@ class AuthenticateViaMicrosoftOAuth20:
tenant_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tenant_id') }})
r"""A globally unique identifier (GUID) that is different than your organization name or domain. Follow these steps to obtain: open one of the Teams where you belong inside the Teams Application -> Click on the … next to the Team title -> Click on Get link to team -> Copy the link to the team and grab the tenant ID form the URL"""
AUTH_TYPE: Final[Optional[SourceMicrosoftTeamsAuthType]] = dataclasses.field(default=SourceMicrosoftTeamsAuthType.CLIENT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceMicrosoftTeamsMicrosoftTeams(str, Enum):
@@ -55,5 +55,3 @@ class SourceMicrosoftTeams:
credentials: Optional[Union[AuthenticateViaMicrosoftOAuth20, AuthenticateViaMicrosoft]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
r"""Choose how to authenticate to Microsoft"""
SOURCE_TYPE: Final[SourceMicrosoftTeamsMicrosoftTeams] = dataclasses.field(default=SourceMicrosoftTeamsMicrosoftTeams.MICROSOFT_TEAMS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_mixpanel.py b/src/airbyte_api/models/shared/source_mixpanel.py
similarity index 99%
rename from src/airbyte/models/shared/source_mixpanel.py
rename to src/airbyte_api/models/shared/source_mixpanel.py
index 7bca42f4..1a638bf8 100644
--- a/src/airbyte/models/shared/source_mixpanel.py
+++ b/src/airbyte_api/models/shared/source_mixpanel.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -18,7 +18,7 @@ class ProjectSecret:
api_secret: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_secret') }})
r"""Mixpanel project secret. See the docs for more information on how to obtain this."""
OPTION_TITLE: Final[Optional[SourceMixpanelSchemasOptionTitle]] = dataclasses.field(default=SourceMixpanelSchemasOptionTitle.PROJECT_SECRET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('option_title'), 'exclude': lambda f: f is None }})
-
+
class SourceMixpanelOptionTitle(str, Enum):
@@ -35,7 +35,7 @@ class ServiceAccount:
username: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('username') }})
r"""Mixpanel Service Account Username. See the docs for more information on how to obtain this."""
OPTION_TITLE: Final[Optional[SourceMixpanelOptionTitle]] = dataclasses.field(default=SourceMixpanelOptionTitle.SERVICE_ACCOUNT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('option_title'), 'exclude': lambda f: f is None }})
-
+
class SourceMixpanelRegion(str, Enum):
@@ -67,5 +67,3 @@ class SourceMixpanel:
SOURCE_TYPE: Final[Mixpanel] = dataclasses.field(default=Mixpanel.MIXPANEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[date] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""The date in the format YYYY-MM-DD. Any data before this date will not be replicated. If this option is not set, the connector will replicate data from up to one year ago by default."""
-
-
diff --git a/src/airbyte/models/shared/source_monday.py b/src/airbyte_api/models/shared/source_monday.py
similarity index 98%
rename from src/airbyte/models/shared/source_monday.py
rename to src/airbyte_api/models/shared/source_monday.py
index 0697048c..474690dc 100644
--- a/src/airbyte/models/shared/source_monday.py
+++ b/src/airbyte_api/models/shared/source_monday.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -17,7 +17,7 @@ class APIToken:
api_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_token') }})
r"""API Token for making authenticated requests."""
AUTH_TYPE: Final[SourceMondaySchemasAuthType] = dataclasses.field(default=SourceMondaySchemasAuthType.API_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceMondayAuthType(str, Enum):
@@ -36,7 +36,7 @@ class SourceMondayOAuth20:
AUTH_TYPE: Final[SourceMondayAuthType] = dataclasses.field(default=SourceMondayAuthType.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
subdomain: Optional[str] = dataclasses.field(default='', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('subdomain'), 'exclude': lambda f: f is None }})
r"""Slug/subdomain of the account, or the first part of the URL that comes before .monday.com"""
-
+
class SourceMondayMonday(str, Enum):
@@ -48,5 +48,3 @@ class SourceMondayMonday(str, Enum):
class SourceMonday:
credentials: Optional[Union[SourceMondayOAuth20, APIToken]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
SOURCE_TYPE: Final[SourceMondayMonday] = dataclasses.field(default=SourceMondayMonday.MONDAY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_mongodb_internal_poc.py b/src/airbyte_api/models/shared/source_mongodb_internal_poc.py
similarity index 98%
rename from src/airbyte/models/shared/source_mongodb_internal_poc.py
rename to src/airbyte_api/models/shared/source_mongodb_internal_poc.py
index 85ad3676..5b3700ce 100644
--- a/src/airbyte/models/shared/source_mongodb_internal_poc.py
+++ b/src/airbyte_api/models/shared/source_mongodb_internal_poc.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -25,5 +25,3 @@ class SourceMongodbInternalPoc:
SOURCE_TYPE: Final[MongodbInternalPoc] = dataclasses.field(default=MongodbInternalPoc.MONGODB_INTERNAL_POC, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
user: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('user'), 'exclude': lambda f: f is None }})
r"""The username which is used to access the database."""
-
-
diff --git a/src/airbyte/models/shared/source_mongodb_v2.py b/src/airbyte_api/models/shared/source_mongodb_v2.py
similarity index 99%
rename from src/airbyte/models/shared/source_mongodb_v2.py
rename to src/airbyte_api/models/shared/source_mongodb_v2.py
index a1717a56..e0705dea 100644
--- a/src/airbyte/models/shared/source_mongodb_v2.py
+++ b/src/airbyte_api/models/shared/source_mongodb_v2.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Any, Dict, Final, Optional, Union
@@ -30,7 +30,7 @@ class SelfManagedReplicaSet:
r"""When enabled, syncs will validate and structure records against the stream's schema."""
username: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('username'), 'exclude': lambda f: f is None }})
r"""The username which is used to access the database."""
-
+
class SourceMongodbV2ClusterType(str, Enum):
@@ -56,7 +56,7 @@ class MongoDBAtlasReplicaSet:
CLUSTER_TYPE: Final[SourceMongodbV2ClusterType] = dataclasses.field(default=SourceMongodbV2ClusterType.ATLAS_REPLICA_SET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('cluster_type') }})
schema_enforced: Optional[bool] = dataclasses.field(default=True, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('schema_enforced'), 'exclude': lambda f: f is None }})
r"""When enabled, syncs will validate and structure records against the stream's schema."""
-
+
class MongodbV2(str, Enum):
@@ -75,5 +75,3 @@ class SourceMongodbV2:
queue_size: Optional[int] = dataclasses.field(default=10000, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('queue_size'), 'exclude': lambda f: f is None }})
r"""The size of the internal queue. This may interfere with memory consumption and efficiency of the connector, please be careful."""
SOURCE_TYPE: Final[MongodbV2] = dataclasses.field(default=MongodbV2.MONGODB_V2, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_mssql.py b/src/airbyte_api/models/shared/source_mssql.py
similarity index 99%
rename from src/airbyte/models/shared/source_mssql.py
rename to src/airbyte_api/models/shared/source_mssql.py
index 7d98dc18..ac768334 100644
--- a/src/airbyte/models/shared/source_mssql.py
+++ b/src/airbyte_api/models/shared/source_mssql.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional, Union
@@ -16,7 +16,7 @@ class SourceMssqlSchemasMethod(str, Enum):
class ScanChangesWithUserDefinedCursor:
r"""Incrementally detects new inserts and updates using the cursor column chosen when configuring a connection (e.g. created_at, updated_at)."""
METHOD: Final[SourceMssqlSchemasMethod] = dataclasses.field(default=SourceMssqlSchemasMethod.STANDARD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class SourceMssqlMethod(str, Enum):
@@ -30,7 +30,7 @@ class ReadChangesUsingChangeDataCaptureCDC:
initial_waiting_seconds: Optional[int] = dataclasses.field(default=300, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('initial_waiting_seconds'), 'exclude': lambda f: f is None }})
r"""The amount of time the connector will wait when it launches to determine if there is new data to sync or not. Defaults to 300 seconds. Valid range: 120 seconds to 1200 seconds. Read about initial waiting time."""
METHOD: Final[SourceMssqlMethod] = dataclasses.field(default=SourceMssqlMethod.CDC, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class SourceMssqlMssql(str, Enum):
@@ -49,7 +49,7 @@ class SourceMssqlEncryptedVerifyCertificate:
host_name_in_certificate: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('hostNameInCertificate'), 'exclude': lambda f: f is None }})
r"""Specifies the host name of the server. The value of this property must match the subject property of the certificate."""
SSL_METHOD: Final[SourceMssqlSchemasSSLMethodSSLMethodSSLMethod] = dataclasses.field(default=SourceMssqlSchemasSSLMethodSSLMethodSSLMethod.ENCRYPTED_VERIFY_CERTIFICATE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('ssl_method') }})
-
+
class SourceMssqlSchemasSslMethodSslMethod(str, Enum):
@@ -61,7 +61,7 @@ class SourceMssqlSchemasSslMethodSslMethod(str, Enum):
class SourceMssqlEncryptedTrustServerCertificate:
r"""Use the certificate provided by the server without verification. (For testing purposes only!)"""
SSL_METHOD: Final[SourceMssqlSchemasSslMethodSslMethod] = dataclasses.field(default=SourceMssqlSchemasSslMethodSslMethod.ENCRYPTED_TRUST_SERVER_CERTIFICATE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('ssl_method') }})
-
+
class SourceMssqlSchemasSslMethod(str, Enum):
@@ -73,7 +73,7 @@ class SourceMssqlSchemasSslMethod(str, Enum):
class Unencrypted:
r"""Data transfer will not be encrypted."""
SSL_METHOD: Final[SourceMssqlSchemasSslMethod] = dataclasses.field(default=SourceMssqlSchemasSslMethod.UNENCRYPTED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('ssl_method') }})
-
+
class SourceMssqlSchemasTunnelMethodTunnelMethod(str, Enum):
@@ -94,7 +94,7 @@ class SourceMssqlPasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class SourceMssqlSchemasTunnelMethod(str, Enum):
@@ -115,7 +115,7 @@ class SourceMssqlSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class SourceMssqlTunnelMethod(str, Enum):
@@ -128,7 +128,7 @@ class SourceMssqlTunnelMethod(str, Enum):
class SourceMssqlNoTunnel:
TUNNEL_METHOD: Final[SourceMssqlTunnelMethod] = dataclasses.field(default=SourceMssqlTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -156,5 +156,3 @@ class SourceMssql:
r"""The encryption method which is used when communicating with the database."""
tunnel_method: Optional[Union[SourceMssqlNoTunnel, SourceMssqlSSHKeyAuthentication, SourceMssqlPasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/source_my_hours.py b/src/airbyte_api/models/shared/source_my_hours.py
similarity index 97%
rename from src/airbyte/models/shared/source_my_hours.py
rename to src/airbyte_api/models/shared/source_my_hours.py
index 49080c98..41e157bd 100644
--- a/src/airbyte/models/shared/source_my_hours.py
+++ b/src/airbyte_api/models/shared/source_my_hours.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -23,5 +23,3 @@ class SourceMyHours:
logs_batch_size: Optional[int] = dataclasses.field(default=30, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('logs_batch_size'), 'exclude': lambda f: f is None }})
r"""Pagination size used for retrieving logs in days"""
SOURCE_TYPE: Final[MyHours] = dataclasses.field(default=MyHours.MY_HOURS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_mysql.py b/src/airbyte_api/models/shared/source_mysql.py
similarity index 99%
rename from src/airbyte/models/shared/source_mysql.py
rename to src/airbyte_api/models/shared/source_mysql.py
index 53de87e9..01efce48 100644
--- a/src/airbyte/models/shared/source_mysql.py
+++ b/src/airbyte_api/models/shared/source_mysql.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -16,7 +16,7 @@ class SourceMysqlSchemasMethod(str, Enum):
class SourceMysqlScanChangesWithUserDefinedCursor:
r"""Incrementally detects new inserts and updates using the cursor column chosen when configuring a connection (e.g. created_at, updated_at)."""
METHOD: Final[SourceMysqlSchemasMethod] = dataclasses.field(default=SourceMysqlSchemasMethod.STANDARD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class SourceMysqlMethod(str, Enum):
@@ -32,7 +32,7 @@ class ReadChangesUsingBinaryLogCDC:
METHOD: Final[SourceMysqlMethod] = dataclasses.field(default=SourceMysqlMethod.CDC, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
server_time_zone: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('server_time_zone'), 'exclude': lambda f: f is None }})
r"""Enter the configured MySQL server timezone. This should only be done if the configured timezone in your MySQL instance does not conform to IANNA standard."""
-
+
class SourceMysqlMysql(str, Enum):
@@ -55,7 +55,7 @@ class VerifyIdentity:
client_key_password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_key_password'), 'exclude': lambda f: f is None }})
r"""Password for keystorage. This field is optional. If you do not add it - the password will be generated automatically."""
MODE: Final[SourceMysqlSchemasSSLModeSSLModesMode] = dataclasses.field(default=SourceMysqlSchemasSSLModeSSLModesMode.VERIFY_IDENTITY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode') }})
-
+
class SourceMysqlSchemasSslModeMode(str, Enum):
@@ -75,7 +75,7 @@ class SourceMysqlVerifyCA:
client_key_password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_key_password'), 'exclude': lambda f: f is None }})
r"""Password for keystorage. This field is optional. If you do not add it - the password will be generated automatically."""
MODE: Final[SourceMysqlSchemasSslModeMode] = dataclasses.field(default=SourceMysqlSchemasSslModeMode.VERIFY_CA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode') }})
-
+
class SourceMysqlSchemasMode(str, Enum):
@@ -87,7 +87,7 @@ class SourceMysqlSchemasMode(str, Enum):
class Required:
r"""Always connect with SSL. If the MySQL server doesn’t support SSL, the connection will not be established. Certificate Authority (CA) and Hostname are not verified."""
MODE: Final[SourceMysqlSchemasMode] = dataclasses.field(default=SourceMysqlSchemasMode.REQUIRED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode') }})
-
+
class SourceMysqlMode(str, Enum):
@@ -99,7 +99,7 @@ class SourceMysqlMode(str, Enum):
class Preferred:
r"""Automatically attempt SSL connection. If the MySQL server does not support SSL, continue with a regular connection."""
MODE: Final[SourceMysqlMode] = dataclasses.field(default=SourceMysqlMode.PREFERRED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode') }})
-
+
class SourceMysqlSchemasTunnelMethodTunnelMethod(str, Enum):
@@ -120,7 +120,7 @@ class SourceMysqlPasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class SourceMysqlSchemasTunnelMethod(str, Enum):
@@ -141,7 +141,7 @@ class SourceMysqlSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class SourceMysqlTunnelMethod(str, Enum):
@@ -154,7 +154,7 @@ class SourceMysqlTunnelMethod(str, Enum):
class SourceMysqlNoTunnel:
TUNNEL_METHOD: Final[SourceMysqlTunnelMethod] = dataclasses.field(default=SourceMysqlTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -180,5 +180,3 @@ class SourceMysql:
r"""SSL connection modes. Read more in the docs."""
tunnel_method: Optional[Union[SourceMysqlNoTunnel, SourceMysqlSSHKeyAuthentication, SourceMysqlPasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/source_netsuite.py b/src/airbyte_api/models/shared/source_netsuite.py
similarity index 98%
rename from src/airbyte/models/shared/source_netsuite.py
rename to src/airbyte_api/models/shared/source_netsuite.py
index cf5d494e..9e945fa6 100644
--- a/src/airbyte/models/shared/source_netsuite.py
+++ b/src/airbyte_api/models/shared/source_netsuite.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional
@@ -31,5 +31,3 @@ class SourceNetsuite:
SOURCE_TYPE: Final[Netsuite] = dataclasses.field(default=Netsuite.NETSUITE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
window_in_days: Optional[int] = dataclasses.field(default=30, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('window_in_days'), 'exclude': lambda f: f is None }})
r"""The amount of days used to query the data with date chunks. Set smaller value, if you have lots of data."""
-
-
diff --git a/src/airbyte/models/shared/source_notion.py b/src/airbyte_api/models/shared/source_notion.py
similarity index 98%
rename from src/airbyte/models/shared/source_notion.py
rename to src/airbyte_api/models/shared/source_notion.py
index 4c36020c..3bb6987a 100644
--- a/src/airbyte/models/shared/source_notion.py
+++ b/src/airbyte_api/models/shared/source_notion.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class SourceNotionAccessToken:
token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token') }})
r"""The Access Token for your private Notion integration. See the docs for more information on how to obtain this token."""
AUTH_TYPE: Final[SourceNotionSchemasAuthType] = dataclasses.field(default=SourceNotionSchemasAuthType.TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceNotionAuthType(str, Enum):
@@ -36,7 +36,7 @@ class SourceNotionOAuth20:
client_secret: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret') }})
r"""The Client Secret of your Notion integration. See our docs for more information."""
AUTH_TYPE: Final[SourceNotionAuthType] = dataclasses.field(default=SourceNotionAuthType.O_AUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceNotionNotion(str, Enum):
@@ -51,5 +51,3 @@ class SourceNotion:
SOURCE_TYPE: Final[SourceNotionNotion] = dataclasses.field(default=SourceNotionNotion.NOTION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format YYYY-MM-DDTHH:MM:SS.000Z. During incremental sync, any data generated before this date will not be replicated. If left blank, the start date will be set to 2 years before the present date."""
-
-
diff --git a/src/airbyte/models/shared/source_nytimes.py b/src/airbyte_api/models/shared/source_nytimes.py
similarity index 98%
rename from src/airbyte/models/shared/source_nytimes.py
rename to src/airbyte_api/models/shared/source_nytimes.py
index e9c1e2af..a8348e7b 100644
--- a/src/airbyte/models/shared/source_nytimes.py
+++ b/src/airbyte_api/models/shared/source_nytimes.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -36,5 +36,3 @@ class SourceNytimes:
share_type: Optional[ShareTypeUsedForMostPopularSharedStream] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('share_type'), 'exclude': lambda f: f is None }})
r"""Share Type"""
SOURCE_TYPE: Final[Nytimes] = dataclasses.field(default=Nytimes.NYTIMES, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_okta.py b/src/airbyte_api/models/shared/source_okta.py
similarity index 98%
rename from src/airbyte/models/shared/source_okta.py
rename to src/airbyte_api/models/shared/source_okta.py
index f14153cc..8d381b56 100644
--- a/src/airbyte/models/shared/source_okta.py
+++ b/src/airbyte_api/models/shared/source_okta.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -17,7 +17,7 @@ class SourceOktaAPIToken:
api_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_token') }})
r"""An Okta token. See the docs for instructions on how to generate it."""
AUTH_TYPE: Final[SourceOktaSchemasAuthType] = dataclasses.field(default=SourceOktaSchemasAuthType.API_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceOktaAuthType(str, Enum):
@@ -34,7 +34,7 @@ class SourceOktaOAuth20:
refresh_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token') }})
r"""Refresh Token to obtain new Access Token, when it's expired."""
AUTH_TYPE: Final[SourceOktaAuthType] = dataclasses.field(default=SourceOktaAuthType.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class Okta(str, Enum):
@@ -50,5 +50,3 @@ class SourceOkta:
SOURCE_TYPE: Final[Okta] = dataclasses.field(default=Okta.OKTA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'exclude': lambda f: f is None }})
r"""UTC date and time in the format YYYY-MM-DDTHH:MM:SSZ. Any data before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_omnisend.py b/src/airbyte_api/models/shared/source_omnisend.py
similarity index 95%
rename from src/airbyte/models/shared/source_omnisend.py
rename to src/airbyte_api/models/shared/source_omnisend.py
index cc659620..9a34cb15 100644
--- a/src/airbyte/models/shared/source_omnisend.py
+++ b/src/airbyte_api/models/shared/source_omnisend.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceOmnisend:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""API Key"""
SOURCE_TYPE: Final[Omnisend] = dataclasses.field(default=Omnisend.OMNISEND, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_onesignal.py b/src/airbyte_api/models/shared/source_onesignal.py
similarity index 98%
rename from src/airbyte/models/shared/source_onesignal.py
rename to src/airbyte_api/models/shared/source_onesignal.py
index 4d33a0fe..282485c1 100644
--- a/src/airbyte/models/shared/source_onesignal.py
+++ b/src/airbyte_api/models/shared/source_onesignal.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -16,7 +16,7 @@ class Applications:
app_api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('app_api_key') }})
app_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('app_id') }})
app_name: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('app_name'), 'exclude': lambda f: f is None }})
-
+
class Onesignal(str, Enum):
@@ -35,5 +35,3 @@ class SourceOnesignal:
user_auth_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('user_auth_key') }})
r"""OneSignal User Auth Key, see the docs for more information on how to obtain this key."""
SOURCE_TYPE: Final[Onesignal] = dataclasses.field(default=Onesignal.ONESIGNAL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_oracle.py b/src/airbyte_api/models/shared/source_oracle.py
similarity index 99%
rename from src/airbyte/models/shared/source_oracle.py
rename to src/airbyte_api/models/shared/source_oracle.py
index d211d198..2e8ce085 100644
--- a/src/airbyte/models/shared/source_oracle.py
+++ b/src/airbyte_api/models/shared/source_oracle.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional, Union
@@ -17,7 +17,7 @@ class SystemIDSID:
r"""Use SID (Oracle System Identifier)"""
sid: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sid') }})
CONNECTION_TYPE: Final[Optional[SourceOracleConnectionType]] = dataclasses.field(default=SourceOracleConnectionType.SID, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('connection_type'), 'exclude': lambda f: f is None }})
-
+
class ConnectionType(str, Enum):
@@ -30,7 +30,7 @@ class ServiceName:
r"""Use service name"""
service_name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('service_name') }})
CONNECTION_TYPE: Final[Optional[ConnectionType]] = dataclasses.field(default=ConnectionType.SERVICE_NAME, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('connection_type'), 'exclude': lambda f: f is None }})
-
+
class SourceOracleEncryptionMethod(str, Enum):
@@ -44,7 +44,7 @@ class TLSEncryptedVerifyCertificate:
ssl_certificate: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('ssl_certificate') }})
r"""Privacy Enhanced Mail (PEM) files are concatenated certificate containers frequently used in certificate installations."""
ENCRYPTION_METHOD: Final[SourceOracleEncryptionMethod] = dataclasses.field(default=SourceOracleEncryptionMethod.ENCRYPTED_VERIFY_CERTIFICATE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('encryption_method') }})
-
+
class EncryptionAlgorithm(str, Enum):
@@ -64,7 +64,7 @@ class NativeNetworkEncryptionNNE:
encryption_algorithm: Optional[EncryptionAlgorithm] = dataclasses.field(default=EncryptionAlgorithm.AES256, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('encryption_algorithm'), 'exclude': lambda f: f is None }})
r"""This parameter defines what encryption algorithm is used."""
ENCRYPTION_METHOD: Final[EncryptionMethod] = dataclasses.field(default=EncryptionMethod.CLIENT_NNE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('encryption_method') }})
-
+
class SourceOracleOracle(str, Enum):
@@ -88,7 +88,7 @@ class SourceOraclePasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class SourceOracleSchemasTunnelMethod(str, Enum):
@@ -109,7 +109,7 @@ class SourceOracleSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class SourceOracleTunnelMethod(str, Enum):
@@ -122,7 +122,7 @@ class SourceOracleTunnelMethod(str, Enum):
class SourceOracleNoTunnel:
TUNNEL_METHOD: Final[SourceOracleTunnelMethod] = dataclasses.field(default=SourceOracleTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -144,7 +144,7 @@ class SourceOracle:
port: Optional[int] = dataclasses.field(default=1521, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('port'), 'exclude': lambda f: f is None }})
r"""Port of the database.
Oracle Corporations recommends the following port numbers:
- 1521 - Default listening port for client connections to the listener.
+ 1521 - Default listening port for client connections to the listener.
2484 - Recommended and officially registered listening port for client connections to the listener using TCP/IP with SSL
"""
schemas: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('schemas'), 'exclude': lambda f: f is None }})
@@ -152,5 +152,3 @@ class SourceOracle:
SOURCE_TYPE: Final[SourceOracleOracle] = dataclasses.field(default=SourceOracleOracle.ORACLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
tunnel_method: Optional[Union[SourceOracleNoTunnel, SourceOracleSSHKeyAuthentication, SourceOraclePasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/source_orb.py b/src/airbyte_api/models/shared/source_orb.py
similarity index 98%
rename from src/airbyte/models/shared/source_orb.py
rename to src/airbyte_api/models/shared/source_orb.py
index 7c6a9bfd..40e351de 100644
--- a/src/airbyte/models/shared/source_orb.py
+++ b/src/airbyte_api/models/shared/source_orb.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional
@@ -29,5 +29,3 @@ class SourceOrb:
r"""Property key names to extract from all events, in order to enrich ledger entries corresponding to an event deduction."""
subscription_usage_grouping_key: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('subscription_usage_grouping_key'), 'exclude': lambda f: f is None }})
r"""Property key name to group subscription usage by."""
-
-
diff --git a/src/airbyte/models/shared/source_orbit.py b/src/airbyte_api/models/shared/source_orbit.py
similarity index 97%
rename from src/airbyte/models/shared/source_orbit.py
rename to src/airbyte_api/models/shared/source_orbit.py
index 2e87002d..5275f09d 100644
--- a/src/airbyte/models/shared/source_orbit.py
+++ b/src/airbyte_api/models/shared/source_orbit.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class SourceOrbit:
SOURCE_TYPE: Final[Orbit] = dataclasses.field(default=Orbit.ORBIT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'exclude': lambda f: f is None }})
r"""Date in the format 2022-06-26. Only load members whose last activities are after this date."""
-
-
diff --git a/src/airbyte/models/shared/source_outbrain_amplify.py b/src/airbyte_api/models/shared/source_outbrain_amplify.py
similarity index 99%
rename from src/airbyte/models/shared/source_outbrain_amplify.py
rename to src/airbyte_api/models/shared/source_outbrain_amplify.py
index 3e973a49..00c91b01 100644
--- a/src/airbyte/models/shared/source_outbrain_amplify.py
+++ b/src/airbyte_api/models/shared/source_outbrain_amplify.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -19,7 +19,7 @@ class SourceOutbrainAmplifyUsernamePassword:
username: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('username') }})
r"""Add Username for authentication."""
TYPE: Final[BothUsernameAndPasswordIsRequiredForAuthenticationRequest] = dataclasses.field(default=BothUsernameAndPasswordIsRequiredForAuthenticationRequest.USERNAME_PASSWORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('type') }})
-
+
class AccessTokenIsRequiredForAuthenticationRequests(str, Enum):
@@ -32,7 +32,7 @@ class SourceOutbrainAmplifyAccessToken:
access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token') }})
r"""Access Token for making authenticated requests."""
TYPE: Final[AccessTokenIsRequiredForAuthenticationRequests] = dataclasses.field(default=AccessTokenIsRequiredForAuthenticationRequests.ACCESS_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('type') }})
-
+
class GranularityForGeoLocationRegion(str, Enum):
@@ -65,5 +65,3 @@ class SourceOutbrainAmplify:
report_granularity: Optional[GranularityForPeriodicReports] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('report_granularity'), 'exclude': lambda f: f is None }})
r"""The granularity used for periodic data in reports. See the docs."""
SOURCE_TYPE: Final[OutbrainAmplify] = dataclasses.field(default=OutbrainAmplify.OUTBRAIN_AMPLIFY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_outreach.py b/src/airbyte_api/models/shared/source_outreach.py
similarity index 98%
rename from src/airbyte/models/shared/source_outreach.py
rename to src/airbyte_api/models/shared/source_outreach.py
index 82ecf217..fab8d144 100644
--- a/src/airbyte/models/shared/source_outreach.py
+++ b/src/airbyte_api/models/shared/source_outreach.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -25,5 +25,3 @@ class SourceOutreach:
start_date: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date') }})
r"""The date from which you'd like to replicate data for Outreach API, in the format YYYY-MM-DDT00:00:00Z. All data generated after this date will be replicated."""
SOURCE_TYPE: Final[Outreach] = dataclasses.field(default=Outreach.OUTREACH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_paypal_transaction.py b/src/airbyte_api/models/shared/source_paypal_transaction.py
similarity index 98%
rename from src/airbyte/models/shared/source_paypal_transaction.py
rename to src/airbyte_api/models/shared/source_paypal_transaction.py
index 630fda79..b54326ea 100644
--- a/src/airbyte/models/shared/source_paypal_transaction.py
+++ b/src/airbyte_api/models/shared/source_paypal_transaction.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -29,5 +29,3 @@ class SourcePaypalTransaction:
SOURCE_TYPE: Final[PaypalTransaction] = dataclasses.field(default=PaypalTransaction.PAYPAL_TRANSACTION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
time_window: Optional[int] = dataclasses.field(default=7, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('time_window'), 'exclude': lambda f: f is None }})
r"""The number of days per request. Must be a number between 1 and 31."""
-
-
diff --git a/src/airbyte/models/shared/source_paystack.py b/src/airbyte_api/models/shared/source_paystack.py
similarity index 98%
rename from src/airbyte/models/shared/source_paystack.py
rename to src/airbyte_api/models/shared/source_paystack.py
index 13dab288..3df7868d 100644
--- a/src/airbyte/models/shared/source_paystack.py
+++ b/src/airbyte_api/models/shared/source_paystack.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -23,5 +23,3 @@ class SourcePaystack:
lookback_window_days: Optional[int] = dataclasses.field(default=0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('lookback_window_days'), 'exclude': lambda f: f is None }})
r"""When set, the connector will always reload data from the past N days, where N is the value set here. This is useful if your data is updated after creation."""
SOURCE_TYPE: Final[Paystack] = dataclasses.field(default=Paystack.PAYSTACK, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_pendo.py b/src/airbyte_api/models/shared/source_pendo.py
similarity index 95%
rename from src/airbyte/models/shared/source_pendo.py
rename to src/airbyte_api/models/shared/source_pendo.py
index 0e8f44cc..69bc4b60 100644
--- a/src/airbyte/models/shared/source_pendo.py
+++ b/src/airbyte_api/models/shared/source_pendo.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -16,5 +16,3 @@ class Pendo(str, Enum):
class SourcePendo:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
SOURCE_TYPE: Final[Pendo] = dataclasses.field(default=Pendo.PENDO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_persistiq.py b/src/airbyte_api/models/shared/source_persistiq.py
similarity index 96%
rename from src/airbyte/models/shared/source_persistiq.py
rename to src/airbyte_api/models/shared/source_persistiq.py
index 823a2b0c..66e0030d 100644
--- a/src/airbyte/models/shared/source_persistiq.py
+++ b/src/airbyte_api/models/shared/source_persistiq.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourcePersistiq:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""PersistIq API Key. See the docs for more information on where to find that key."""
SOURCE_TYPE: Final[Persistiq] = dataclasses.field(default=Persistiq.PERSISTIQ, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_pexels_api.py b/src/airbyte_api/models/shared/source_pexels_api.py
similarity index 98%
rename from src/airbyte/models/shared/source_pexels_api.py
rename to src/airbyte_api/models/shared/source_pexels_api.py
index 9a1a9438..321ce0fb 100644
--- a/src/airbyte/models/shared/source_pexels_api.py
+++ b/src/airbyte_api/models/shared/source_pexels_api.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -27,5 +27,3 @@ class SourcePexelsAPI:
size: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('size'), 'exclude': lambda f: f is None }})
r"""Optional, Minimum photo size. The current supported sizes are large(24MP), medium(12MP) or small(4MP)."""
SOURCE_TYPE: Final[PexelsAPI] = dataclasses.field(default=PexelsAPI.PEXELS_API, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_pinterest.py b/src/airbyte_api/models/shared/source_pinterest.py
similarity index 99%
rename from src/airbyte/models/shared/source_pinterest.py
rename to src/airbyte_api/models/shared/source_pinterest.py
index aa974d7e..99470d55 100644
--- a/src/airbyte/models/shared/source_pinterest.py
+++ b/src/airbyte_api/models/shared/source_pinterest.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -22,7 +22,7 @@ class OAuth20:
refresh_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token') }})
r"""Refresh Token to obtain new Access Token, when it's expired."""
AUTH_METHOD: Final[SourcePinterestAuthMethod] = dataclasses.field(default=SourcePinterestAuthMethod.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method') }})
-
+
class SourcePinterestValidEnums(str, Enum):
@@ -225,7 +225,7 @@ class ReportConfig:
r"""A date in the format YYYY-MM-DD. If you have not set a date, it would be defaulted to latest allowed date by report api (913 days from today)."""
view_window_days: Optional[ViewWindowDays] = dataclasses.field(default=ViewWindowDays.THIRTY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('view_window_days'), 'exclude': lambda f: f is None }})
r"""Number of days to use as the conversion attribution window for a view action."""
-
+
class SourcePinterestPinterest(str, Enum):
@@ -249,5 +249,3 @@ class SourcePinterest:
r"""A date in the format YYYY-MM-DD. If you have not set a date, it would be defaulted to latest allowed date by api (89 days from today)."""
status: Optional[List[Status]] = dataclasses.field(default=UNSET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('status'), 'exclude': lambda f: f is SourcePinterest.UNSET }})
r"""For the ads, ad_groups, and campaigns streams, specifying a status will filter out records that do not match the specified ones. If a status is not specified, the source will default to records with a status of either ACTIVE or PAUSED."""
-
-
diff --git a/src/airbyte/models/shared/source_pipedrive.py b/src/airbyte_api/models/shared/source_pipedrive.py
similarity index 97%
rename from src/airbyte/models/shared/source_pipedrive.py
rename to src/airbyte_api/models/shared/source_pipedrive.py
index b1c588cd..965674c8 100644
--- a/src/airbyte/models/shared/source_pipedrive.py
+++ b/src/airbyte_api/models/shared/source_pipedrive.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class SourcePipedrive:
replication_start_date: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('replication_start_date') }})
r"""UTC date and time in the format 2017-01-25T00:00:00Z. Any data before this date will not be replicated. When specified and not None, then stream will behave as incremental"""
SOURCE_TYPE: Final[Pipedrive] = dataclasses.field(default=Pipedrive.PIPEDRIVE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_pocket.py b/src/airbyte_api/models/shared/source_pocket.py
similarity index 99%
rename from src/airbyte/models/shared/source_pocket.py
rename to src/airbyte_api/models/shared/source_pocket.py
index df678049..990651bb 100644
--- a/src/airbyte/models/shared/source_pocket.py
+++ b/src/airbyte_api/models/shared/source_pocket.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -61,5 +61,3 @@ class SourcePocket:
r"""Select the state of the items to retrieve."""
tag: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tag'), 'exclude': lambda f: f is None }})
r"""Return only items tagged with this tag name. Use _untagged_ for retrieving only untagged items."""
-
-
diff --git a/src/airbyte/models/shared/source_pokeapi.py b/src/airbyte_api/models/shared/source_pokeapi.py
similarity index 99%
rename from src/airbyte/models/shared/source_pokeapi.py
rename to src/airbyte_api/models/shared/source_pokeapi.py
index a7594730..189f25fe 100644
--- a/src/airbyte/models/shared/source_pokeapi.py
+++ b/src/airbyte_api/models/shared/source_pokeapi.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -918,5 +918,3 @@ class SourcePokeapi:
pokemon_name: PokemonName = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('pokemon_name') }})
r"""Pokemon requested from the API."""
SOURCE_TYPE: Final[Pokeapi] = dataclasses.field(default=Pokeapi.POKEAPI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_polygon_stock_api.py b/src/airbyte_api/models/shared/source_polygon_stock_api.py
similarity index 98%
rename from src/airbyte/models/shared/source_polygon_stock_api.py
rename to src/airbyte_api/models/shared/source_polygon_stock_api.py
index e4069169..47e897e5 100644
--- a/src/airbyte/models/shared/source_polygon_stock_api.py
+++ b/src/airbyte_api/models/shared/source_polygon_stock_api.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -34,5 +34,3 @@ class SourcePolygonStockAPI:
sort: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sort'), 'exclude': lambda f: f is None }})
r"""Sort the results by timestamp. asc will return results in ascending order (oldest at the top), desc will return results in descending order (newest at the top)."""
SOURCE_TYPE: Final[PolygonStockAPI] = dataclasses.field(default=PolygonStockAPI.POLYGON_STOCK_API, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_postgres.py b/src/airbyte_api/models/shared/source_postgres.py
similarity index 99%
rename from src/airbyte/models/shared/source_postgres.py
rename to src/airbyte_api/models/shared/source_postgres.py
index ecc2281b..1ae4e4ce 100644
--- a/src/airbyte/models/shared/source_postgres.py
+++ b/src/airbyte_api/models/shared/source_postgres.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Any, Dict, Final, List, Optional, Union
@@ -16,7 +16,7 @@ class SourcePostgresSchemasReplicationMethodMethod(str, Enum):
class SourcePostgresScanChangesWithUserDefinedCursor:
r"""Incrementally detects new inserts and updates using the cursor column chosen when configuring a connection (e.g. created_at, updated_at)."""
METHOD: Final[SourcePostgresSchemasReplicationMethodMethod] = dataclasses.field(default=SourcePostgresSchemasReplicationMethodMethod.STANDARD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class SourcePostgresSchemasMethod(str, Enum):
@@ -28,7 +28,7 @@ class SourcePostgresSchemasMethod(str, Enum):
class DetectChangesWithXminSystemColumn:
r"""Recommended - Incrementally reads new inserts and updates via Postgres Xmin system column. Only recommended for tables up to 500GB."""
METHOD: Final[SourcePostgresSchemasMethod] = dataclasses.field(default=SourcePostgresSchemasMethod.XMIN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('method') }})
-
+
class LSNCommitBehaviour(str, Enum):
@@ -65,7 +65,7 @@ class ReadChangesUsingWriteAheadLogCDC:
r"""A logical decoding plugin installed on the PostgreSQL server."""
queue_size: Optional[int] = dataclasses.field(default=10000, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('queue_size'), 'exclude': lambda f: f is None }})
r"""The size of the internal queue. This may interfere with memory consumption and efficiency of the connector, please be careful."""
-
+
class SourcePostgresPostgres(str, Enum):
@@ -90,7 +90,7 @@ class SourcePostgresVerifyFull:
client_key_password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_key_password'), 'exclude': lambda f: f is None }})
r"""Password for keystorage. If you do not add it - the password will be generated automatically."""
MODE: Final[SourcePostgresSchemasSSLModeSSLModes6Mode] = dataclasses.field(default=SourcePostgresSchemasSSLModeSSLModes6Mode.VERIFY_FULL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode') }})
-
+
class SourcePostgresSchemasSSLModeSSLModes5Mode(str, Enum):
@@ -112,7 +112,7 @@ class SourcePostgresVerifyCa:
client_key_password: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_key_password'), 'exclude': lambda f: f is None }})
r"""Password for keystorage. If you do not add it - the password will be generated automatically."""
MODE: Final[SourcePostgresSchemasSSLModeSSLModes5Mode] = dataclasses.field(default=SourcePostgresSchemasSSLModeSSLModes5Mode.VERIFY_CA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode') }})
-
+
class SourcePostgresSchemasSSLModeSSLModesMode(str, Enum):
@@ -126,7 +126,7 @@ class SourcePostgresRequire:
UNSET='__SPEAKEASY_UNSET__'
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
MODE: Final[SourcePostgresSchemasSSLModeSSLModesMode] = dataclasses.field(default=SourcePostgresSchemasSSLModeSSLModesMode.REQUIRE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode') }})
-
+
class SourcePostgresSchemasSslModeMode(str, Enum):
@@ -140,7 +140,7 @@ class SourcePostgresPrefer:
UNSET='__SPEAKEASY_UNSET__'
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
MODE: Final[SourcePostgresSchemasSslModeMode] = dataclasses.field(default=SourcePostgresSchemasSslModeMode.PREFER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode') }})
-
+
class SourcePostgresSchemasMode(str, Enum):
@@ -154,7 +154,7 @@ class SourcePostgresAllow:
UNSET='__SPEAKEASY_UNSET__'
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
MODE: Final[SourcePostgresSchemasMode] = dataclasses.field(default=SourcePostgresSchemasMode.ALLOW, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode') }})
-
+
class SourcePostgresMode(str, Enum):
@@ -168,7 +168,7 @@ class SourcePostgresDisable:
UNSET='__SPEAKEASY_UNSET__'
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
MODE: Final[SourcePostgresMode] = dataclasses.field(default=SourcePostgresMode.DISABLE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode') }})
-
+
class SourcePostgresSchemasTunnelMethodTunnelMethod(str, Enum):
@@ -189,7 +189,7 @@ class SourcePostgresPasswordAuthentication:
r"""Connect through a jump server tunnel host using username and password authentication"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class SourcePostgresSchemasTunnelMethod(str, Enum):
@@ -210,7 +210,7 @@ class SourcePostgresSSHKeyAuthentication:
r"""Connect through a jump server tunnel host using username and ssh key"""
tunnel_port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_port'), 'exclude': lambda f: f is None }})
r"""Port on the proxy/jump server that accepts inbound ssh connections."""
-
+
class SourcePostgresTunnelMethod(str, Enum):
@@ -223,7 +223,7 @@ class SourcePostgresTunnelMethod(str, Enum):
class SourcePostgresNoTunnel:
TUNNEL_METHOD: Final[SourcePostgresTunnelMethod] = dataclasses.field(default=SourcePostgresTunnelMethod.NO_TUNNEL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method') }})
r"""No ssh tunnel needed to connect to database"""
-
+
@@ -253,5 +253,3 @@ class SourcePostgres:
"""
tunnel_method: Optional[Union[SourcePostgresNoTunnel, SourcePostgresSSHKeyAuthentication, SourcePostgresPasswordAuthentication]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tunnel_method'), 'exclude': lambda f: f is None }})
r"""Whether to initiate an SSH tunnel before connecting to the database, and if so, which kind of authentication to use."""
-
-
diff --git a/src/airbyte/models/shared/source_posthog.py b/src/airbyte_api/models/shared/source_posthog.py
similarity index 98%
rename from src/airbyte/models/shared/source_posthog.py
rename to src/airbyte_api/models/shared/source_posthog.py
index daf6a2f6..8d5ad723 100644
--- a/src/airbyte/models/shared/source_posthog.py
+++ b/src/airbyte_api/models/shared/source_posthog.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,5 +25,3 @@ class SourcePosthog:
events_time_step: Optional[int] = dataclasses.field(default=30, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('events_time_step'), 'exclude': lambda f: f is None }})
r"""Set lower value in case of failing long running sync of events stream."""
SOURCE_TYPE: Final[Posthog] = dataclasses.field(default=Posthog.POSTHOG, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_postmarkapp.py b/src/airbyte_api/models/shared/source_postmarkapp.py
similarity index 96%
rename from src/airbyte/models/shared/source_postmarkapp.py
rename to src/airbyte_api/models/shared/source_postmarkapp.py
index 4ee3e20a..71cb1489 100644
--- a/src/airbyte/models/shared/source_postmarkapp.py
+++ b/src/airbyte_api/models/shared/source_postmarkapp.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class SourcePostmarkapp:
x_postmark_server_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('X-Postmark-Server-Token') }})
r"""API Key for server"""
SOURCE_TYPE: Final[Postmarkapp] = dataclasses.field(default=Postmarkapp.POSTMARKAPP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_prestashop.py b/src/airbyte_api/models/shared/source_prestashop.py
similarity index 97%
rename from src/airbyte/models/shared/source_prestashop.py
rename to src/airbyte_api/models/shared/source_prestashop.py
index 99ddca17..c0c281b8 100644
--- a/src/airbyte/models/shared/source_prestashop.py
+++ b/src/airbyte_api/models/shared/source_prestashop.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -22,5 +22,3 @@ class SourcePrestashop:
url: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('url') }})
r"""Shop URL without trailing slash."""
SOURCE_TYPE: Final[Prestashop] = dataclasses.field(default=Prestashop.PRESTASHOP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_punk_api.py b/src/airbyte_api/models/shared/source_punk_api.py
similarity index 97%
rename from src/airbyte/models/shared/source_punk_api.py
rename to src/airbyte_api/models/shared/source_punk_api.py
index 3a2f594c..f9b33f6c 100644
--- a/src/airbyte/models/shared/source_punk_api.py
+++ b/src/airbyte_api/models/shared/source_punk_api.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class SourcePunkAPI:
id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('id'), 'exclude': lambda f: f is None }})
r"""To extract specific data with Unique ID"""
SOURCE_TYPE: Final[PunkAPI] = dataclasses.field(default=PunkAPI.PUNK_API, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_pypi.py b/src/airbyte_api/models/shared/source_pypi.py
similarity index 97%
rename from src/airbyte/models/shared/source_pypi.py
rename to src/airbyte_api/models/shared/source_pypi.py
index d52de1d6..8b45f08b 100644
--- a/src/airbyte/models/shared/source_pypi.py
+++ b/src/airbyte_api/models/shared/source_pypi.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -19,5 +19,3 @@ class SourcePypi:
SOURCE_TYPE: Final[Pypi] = dataclasses.field(default=Pypi.PYPI, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
version: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('version'), 'exclude': lambda f: f is None }})
r"""Version of the project/package. Use it to find a particular release instead of all releases."""
-
-
diff --git a/src/airbyte/models/shared/source_qualaroo.py b/src/airbyte_api/models/shared/source_qualaroo.py
similarity index 98%
rename from src/airbyte/models/shared/source_qualaroo.py
rename to src/airbyte_api/models/shared/source_qualaroo.py
index beca0e55..d828c40c 100644
--- a/src/airbyte/models/shared/source_qualaroo.py
+++ b/src/airbyte_api/models/shared/source_qualaroo.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional
@@ -23,5 +23,3 @@ class SourceQualaroo:
SOURCE_TYPE: Final[Qualaroo] = dataclasses.field(default=Qualaroo.QUALAROO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
survey_ids: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('survey_ids'), 'exclude': lambda f: f is None }})
r"""IDs of the surveys from which you'd like to replicate data. If left empty, data from all surveys to which you have access will be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_quickbooks.py b/src/airbyte_api/models/shared/source_quickbooks.py
similarity index 98%
rename from src/airbyte/models/shared/source_quickbooks.py
rename to src/airbyte_api/models/shared/source_quickbooks.py
index 2ffdbf72..8b9c6bbc 100644
--- a/src/airbyte/models/shared/source_quickbooks.py
+++ b/src/airbyte_api/models/shared/source_quickbooks.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -29,7 +29,7 @@ class SourceQuickbooksOAuth20:
token_expiry_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token_expiry_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""The date-time when the access token should be refreshed."""
AUTH_TYPE: Final[Optional[SourceQuickbooksAuthType]] = dataclasses.field(default=SourceQuickbooksAuthType.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class Quickbooks(str, Enum):
@@ -45,5 +45,3 @@ class SourceQuickbooks:
sandbox: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sandbox'), 'exclude': lambda f: f is None }})
r"""Determines whether to use the sandbox or production environment."""
SOURCE_TYPE: Final[Quickbooks] = dataclasses.field(default=Quickbooks.QUICKBOOKS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_railz.py b/src/airbyte_api/models/shared/source_railz.py
similarity index 96%
rename from src/airbyte/models/shared/source_railz.py
rename to src/airbyte_api/models/shared/source_railz.py
index 9ef37661..df719134 100644
--- a/src/airbyte/models/shared/source_railz.py
+++ b/src/airbyte_api/models/shared/source_railz.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -21,5 +21,3 @@ class SourceRailz:
start_date: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date') }})
r"""Start date"""
SOURCE_TYPE: Final[Railz] = dataclasses.field(default=Railz.RAILZ, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_recharge.py b/src/airbyte_api/models/shared/source_recharge.py
similarity index 98%
rename from src/airbyte/models/shared/source_recharge.py
rename to src/airbyte_api/models/shared/source_recharge.py
index 0ccf0db7..6e48c2d8 100644
--- a/src/airbyte/models/shared/source_recharge.py
+++ b/src/airbyte_api/models/shared/source_recharge.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -23,5 +23,3 @@ class SourceRecharge:
SOURCE_TYPE: Final[Recharge] = dataclasses.field(default=Recharge.RECHARGE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
use_orders_deprecated_api: Optional[bool] = dataclasses.field(default=True, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('use_orders_deprecated_api'), 'exclude': lambda f: f is None }})
r"""Define whether or not the `Orders` stream should use the deprecated `2021-01` API version, or use `2021-11`, otherwise."""
-
-
diff --git a/src/airbyte/models/shared/source_recreation.py b/src/airbyte_api/models/shared/source_recreation.py
similarity index 96%
rename from src/airbyte/models/shared/source_recreation.py
rename to src/airbyte_api/models/shared/source_recreation.py
index 83b3a240..ed9c597f 100644
--- a/src/airbyte/models/shared/source_recreation.py
+++ b/src/airbyte_api/models/shared/source_recreation.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -18,5 +18,3 @@ class SourceRecreation:
r"""API Key"""
query_campsites: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('query_campsites'), 'exclude': lambda f: f is None }})
SOURCE_TYPE: Final[Recreation] = dataclasses.field(default=Recreation.RECREATION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_recruitee.py b/src/airbyte_api/models/shared/source_recruitee.py
similarity index 97%
rename from src/airbyte/models/shared/source_recruitee.py
rename to src/airbyte_api/models/shared/source_recruitee.py
index 0507f332..15296f39 100644
--- a/src/airbyte/models/shared/source_recruitee.py
+++ b/src/airbyte_api/models/shared/source_recruitee.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class SourceRecruitee:
company_id: int = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('company_id') }})
r"""Recruitee Company ID. You can also find this ID on the Recruitee API tokens page."""
SOURCE_TYPE: Final[Recruitee] = dataclasses.field(default=Recruitee.RECRUITEE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_redshift.py b/src/airbyte_api/models/shared/source_redshift.py
similarity index 98%
rename from src/airbyte/models/shared/source_redshift.py
rename to src/airbyte_api/models/shared/source_redshift.py
index 76e6df68..bebdca03 100644
--- a/src/airbyte/models/shared/source_redshift.py
+++ b/src/airbyte_api/models/shared/source_redshift.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, List, Optional
@@ -29,5 +29,3 @@ class SourceRedshift:
schemas: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('schemas'), 'exclude': lambda f: f is None }})
r"""The list of schemas to sync from. Specify one or more explicitly or keep empty to process all schemas. Schema names are case sensitive."""
SOURCE_TYPE: Final[SourceRedshiftRedshift] = dataclasses.field(default=SourceRedshiftRedshift.REDSHIFT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_retently.py b/src/airbyte_api/models/shared/source_retently.py
similarity index 98%
rename from src/airbyte/models/shared/source_retently.py
rename to src/airbyte_api/models/shared/source_retently.py
index b3dbea5f..d3f192c7 100644
--- a/src/airbyte/models/shared/source_retently.py
+++ b/src/airbyte_api/models/shared/source_retently.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Any, Dict, Final, Optional, Union
@@ -19,7 +19,7 @@ class AuthenticateWithAPIToken:
r"""Retently API Token. See the docs for more information on how to obtain this key."""
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
AUTH_TYPE: Final[Optional[SourceRetentlySchemasAuthType]] = dataclasses.field(default=SourceRetentlySchemasAuthType.TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceRetentlyAuthType(str, Enum):
@@ -38,7 +38,7 @@ class AuthenticateViaRetentlyOAuth:
r"""Retently Refresh Token which can be used to fetch new Bearer Tokens when the current one expires."""
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
AUTH_TYPE: Final[Optional[SourceRetentlyAuthType]] = dataclasses.field(default=SourceRetentlyAuthType.CLIENT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceRetentlyRetently(str, Enum):
@@ -51,5 +51,3 @@ class SourceRetently:
credentials: Optional[Union[AuthenticateViaRetentlyOAuth, AuthenticateWithAPIToken]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
r"""Choose how to authenticate to Retently"""
SOURCE_TYPE: Final[Optional[SourceRetentlyRetently]] = dataclasses.field(default=SourceRetentlyRetently.RETENTLY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/source_rki_covid.py b/src/airbyte_api/models/shared/source_rki_covid.py
similarity index 96%
rename from src/airbyte/models/shared/source_rki_covid.py
rename to src/airbyte_api/models/shared/source_rki_covid.py
index f164619f..5dbd8626 100644
--- a/src/airbyte/models/shared/source_rki_covid.py
+++ b/src/airbyte_api/models/shared/source_rki_covid.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceRkiCovid:
start_date: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date') }})
r"""UTC date in the format 2017-01-25. Any data before this date will not be replicated."""
SOURCE_TYPE: Final[RkiCovid] = dataclasses.field(default=RkiCovid.RKI_COVID, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_rss.py b/src/airbyte_api/models/shared/source_rss.py
similarity index 95%
rename from src/airbyte/models/shared/source_rss.py
rename to src/airbyte_api/models/shared/source_rss.py
index 0318dc53..cc6b0bc1 100644
--- a/src/airbyte/models/shared/source_rss.py
+++ b/src/airbyte_api/models/shared/source_rss.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceRss:
url: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('url') }})
r"""RSS Feed URL"""
SOURCE_TYPE: Final[Rss] = dataclasses.field(default=Rss.RSS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_s3.py b/src/airbyte_api/models/shared/source_s3.py
similarity index 99%
rename from src/airbyte/models/shared/source_s3.py
rename to src/airbyte_api/models/shared/source_s3.py
index 08ec9539..6a8692a1 100644
--- a/src/airbyte/models/shared/source_s3.py
+++ b/src/airbyte_api/models/shared/source_s3.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -30,7 +30,7 @@ class Jsonl:
r"""Whether newline characters are allowed in JSON values. Turning this on may affect performance. Leave blank to default to False."""
unexpected_field_behavior: Optional[UnexpectedFieldBehavior] = dataclasses.field(default=UnexpectedFieldBehavior.INFER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('unexpected_field_behavior'), 'exclude': lambda f: f is None }})
r"""How JSON fields outside of explicit_schema (if given) are treated. Check PyArrow documentation for details"""
-
+
class SourceS3SchemasFiletype(str, Enum):
@@ -42,7 +42,7 @@ class SourceS3SchemasFiletype(str, Enum):
class Avro:
r"""This connector utilises fastavro for Avro parsing."""
FILETYPE: Final[Optional[SourceS3SchemasFiletype]] = dataclasses.field(default=SourceS3SchemasFiletype.AVRO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceS3Filetype(str, Enum):
@@ -60,7 +60,7 @@ class Parquet:
columns: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('columns'), 'exclude': lambda f: f is None }})
r"""If you only want to sync a subset of the columns from the file(s), add the columns you want here as a comma-delimited list. Leave it empty to sync all columns."""
FILETYPE: Final[Optional[SourceS3Filetype]] = dataclasses.field(default=SourceS3Filetype.PARQUET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceS3SchemasFormatFileFormatFiletype(str, Enum):
@@ -92,7 +92,7 @@ class Csv:
r"""Whether newline characters are allowed in CSV values. Turning this on may affect performance. Leave blank to default to False."""
quote_char: Optional[str] = dataclasses.field(default='"', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('quote_char'), 'exclude': lambda f: f is None }})
r"""The character used for quoting CSV values. To disallow quoting, make this field blank."""
-
+
@@ -114,7 +114,7 @@ class S3AmazonWebServices:
r"""Specifies the Amazon Resource Name (ARN) of an IAM role that you want to use to perform operations requested using this profile. Set the External ID to the Airbyte workspace ID, which can be found in the URL of this page."""
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00Z. Any file modified before this date will not be replicated."""
-
+
class SourceS3S3(str, Enum):
@@ -132,7 +132,7 @@ class SourceS3Mode(str, Enum):
class SourceS3Local:
r"""Process files locally, supporting `fast` and `ocr` modes. This is the default option."""
MODE: Final[Optional[SourceS3Mode]] = dataclasses.field(default=SourceS3Mode.LOCAL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('mode'), 'exclude': lambda f: f is None }})
-
+
class SourceS3ParsingStrategy(str, Enum):
@@ -154,7 +154,7 @@ class SourceS3DocumentFileTypeFormatExperimental:
r"""If true, skip files that cannot be parsed and pass the error message along as the _ab_source_file_parse_error field. If false, fail the sync."""
strategy: Optional[SourceS3ParsingStrategy] = dataclasses.field(default=SourceS3ParsingStrategy.AUTO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('strategy'), 'exclude': lambda f: f is None }})
r"""The strategy used to parse documents. `fast` extracts text directly from the document which doesn't work for all files. `ocr_only` is more reliable, but slower. `hi_res` is the most reliable, but requires an API key and a hosted instance of unstructured and can't be used with local mode. See the unstructured.io documentation for more details: https://unstructured-io.github.io/unstructured/core/partition.html#partition-pdf"""
-
+
class SourceS3SchemasStreamsFormatFormat4Filetype(str, Enum):
@@ -167,7 +167,7 @@ class SourceS3ParquetFormat:
decimal_as_float: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('decimal_as_float'), 'exclude': lambda f: f is None }})
r"""Whether to convert decimal fields to floats. There is a loss of precision when converting decimals to floats, so this is not recommended."""
FILETYPE: Final[Optional[SourceS3SchemasStreamsFormatFormat4Filetype]] = dataclasses.field(default=SourceS3SchemasStreamsFormatFormat4Filetype.PARQUET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceS3SchemasStreamsFormatFormatFiletype(str, Enum):
@@ -178,7 +178,7 @@ class SourceS3SchemasStreamsFormatFormatFiletype(str, Enum):
@dataclasses.dataclass
class SourceS3JsonlFormat:
FILETYPE: Final[Optional[SourceS3SchemasStreamsFormatFormatFiletype]] = dataclasses.field(default=SourceS3SchemasStreamsFormatFormatFiletype.JSONL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceS3SchemasStreamsFormatFiletype(str, Enum):
@@ -194,7 +194,7 @@ class SourceS3UserProvided:
column_names: List[str] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('column_names') }})
r"""The column names that will be used while emitting the CSV records"""
HEADER_DEFINITION_TYPE: Final[Optional[SourceS3SchemasStreamsHeaderDefinitionType]] = dataclasses.field(default=SourceS3SchemasStreamsHeaderDefinitionType.USER_PROVIDED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceS3SchemasHeaderDefinitionType(str, Enum):
@@ -205,7 +205,7 @@ class SourceS3SchemasHeaderDefinitionType(str, Enum):
@dataclasses.dataclass
class SourceS3Autogenerated:
HEADER_DEFINITION_TYPE: Final[Optional[SourceS3SchemasHeaderDefinitionType]] = dataclasses.field(default=SourceS3SchemasHeaderDefinitionType.AUTOGENERATED, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceS3HeaderDefinitionType(str, Enum):
@@ -216,7 +216,7 @@ class SourceS3HeaderDefinitionType(str, Enum):
@dataclasses.dataclass
class SourceS3FromCSV:
HEADER_DEFINITION_TYPE: Final[Optional[SourceS3HeaderDefinitionType]] = dataclasses.field(default=SourceS3HeaderDefinitionType.FROM_CSV, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('header_definition_type'), 'exclude': lambda f: f is None }})
-
+
class SourceS3InferenceType(str, Enum):
@@ -255,7 +255,7 @@ class SourceS3CSVFormat:
r"""Whether strings can be interpreted as null values. If true, strings that match the null_values set will be interpreted as null. If false, strings that match the null_values set will be interpreted as the string itself."""
true_values: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('true_values'), 'exclude': lambda f: f is None }})
r"""A set of case-sensitive strings that should be interpreted as true values."""
-
+
class SourceS3SchemasStreamsFiletype(str, Enum):
@@ -268,7 +268,7 @@ class SourceS3AvroFormat:
double_as_string: Optional[bool] = dataclasses.field(default=False, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('double_as_string'), 'exclude': lambda f: f is None }})
r"""Whether to convert double fields to strings. This is recommended if you have decimal numbers with a high degree of precision because there can be a loss precision when handling floating point numbers."""
FILETYPE: Final[Optional[SourceS3SchemasStreamsFiletype]] = dataclasses.field(default=SourceS3SchemasStreamsFiletype.AVRO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('filetype'), 'exclude': lambda f: f is None }})
-
+
class SourceS3ValidationPolicy(str, Enum):
@@ -299,7 +299,7 @@ class SourceS3FileBasedStreamConfig:
r"""When enabled, syncs will not validate or structure records against the stream's schema."""
validation_policy: Optional[SourceS3ValidationPolicy] = dataclasses.field(default=SourceS3ValidationPolicy.EMIT_RECORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('validation_policy'), 'exclude': lambda f: f is None }})
r"""The name of the validation policy that dictates sync behavior when a record does not adhere to the stream schema."""
-
+
@@ -334,5 +334,3 @@ class SourceS3:
SOURCE_TYPE: Final[SourceS3S3] = dataclasses.field(default=SourceS3S3.S3, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00.000000Z. Any file modified before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_salesforce.py b/src/airbyte_api/models/shared/source_salesforce.py
similarity index 99%
rename from src/airbyte/models/shared/source_salesforce.py
rename to src/airbyte_api/models/shared/source_salesforce.py
index 9a033223..d19dcf4d 100644
--- a/src/airbyte/models/shared/source_salesforce.py
+++ b/src/airbyte_api/models/shared/source_salesforce.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -31,7 +31,7 @@ class SearchCriteria(str, Enum):
class StreamsCriteria:
value: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('value') }})
criteria: Optional[SearchCriteria] = dataclasses.field(default=SearchCriteria.CONTAINS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('criteria'), 'exclude': lambda f: f is None }})
-
+
@@ -54,5 +54,3 @@ class SourceSalesforce:
r"""Enter the date (or date-time) in the YYYY-MM-DD or YYYY-MM-DDTHH:mm:ssZ format. Airbyte will replicate the data updated on and after this date. If this field is blank, Airbyte will replicate the data for last two years."""
streams_criteria: Optional[List[StreamsCriteria]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('streams_criteria'), 'exclude': lambda f: f is None }})
r"""Add filters to select only required stream based on `SObject` name. Use this field to filter which tables are displayed by this connector. This is useful if your Salesforce account has a large number of tables (>1000), in which case you may find it easier to navigate the UI and speed up the connector's performance if you restrict the tables displayed by this connector."""
-
-
diff --git a/src/airbyte/models/shared/source_salesloft.py b/src/airbyte_api/models/shared/source_salesloft.py
similarity index 98%
rename from src/airbyte/models/shared/source_salesloft.py
rename to src/airbyte_api/models/shared/source_salesloft.py
index f004dcfc..5c356cb8 100644
--- a/src/airbyte/models/shared/source_salesloft.py
+++ b/src/airbyte_api/models/shared/source_salesloft.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class AuthenticateViaAPIKey:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""API Key for making authenticated requests. More instruction on how to find this value in our docs"""
AUTH_TYPE: Final[SourceSalesloftSchemasAuthType] = dataclasses.field(default=SourceSalesloftSchemasAuthType.API_KEY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceSalesloftAuthType(str, Enum):
@@ -40,7 +40,7 @@ class AuthenticateViaOAuth:
token_expiry_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token_expiry_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""The date-time when the access token should be refreshed."""
AUTH_TYPE: Final[SourceSalesloftAuthType] = dataclasses.field(default=SourceSalesloftAuthType.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class Salesloft(str, Enum):
@@ -54,5 +54,3 @@ class SourceSalesloft:
start_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""The date from which you'd like to replicate data for Salesloft API, in the format YYYY-MM-DDT00:00:00Z. All data generated after this date will be replicated."""
SOURCE_TYPE: Final[Salesloft] = dataclasses.field(default=Salesloft.SALESLOFT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_sap_fieldglass.py b/src/airbyte_api/models/shared/source_sap_fieldglass.py
similarity index 95%
rename from src/airbyte/models/shared/source_sap_fieldglass.py
rename to src/airbyte_api/models/shared/source_sap_fieldglass.py
index 9cadd72b..e9ee1108 100644
--- a/src/airbyte/models/shared/source_sap_fieldglass.py
+++ b/src/airbyte_api/models/shared/source_sap_fieldglass.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceSapFieldglass:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""API Key"""
SOURCE_TYPE: Final[SapFieldglass] = dataclasses.field(default=SapFieldglass.SAP_FIELDGLASS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_secoda.py b/src/airbyte_api/models/shared/source_secoda.py
similarity index 96%
rename from src/airbyte/models/shared/source_secoda.py
rename to src/airbyte_api/models/shared/source_secoda.py
index 1efe6500..4f49f4bb 100644
--- a/src/airbyte/models/shared/source_secoda.py
+++ b/src/airbyte_api/models/shared/source_secoda.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceSecoda:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""Your API Access Key. See here. The key is case sensitive."""
SOURCE_TYPE: Final[Secoda] = dataclasses.field(default=Secoda.SECODA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_sendgrid.py b/src/airbyte_api/models/shared/source_sendgrid.py
similarity index 97%
rename from src/airbyte/models/shared/source_sendgrid.py
rename to src/airbyte_api/models/shared/source_sendgrid.py
index 71e9ebdc..3947661f 100644
--- a/src/airbyte/models/shared/source_sendgrid.py
+++ b/src/airbyte_api/models/shared/source_sendgrid.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -21,5 +21,3 @@ class SourceSendgrid:
SOURCE_TYPE: Final[Sendgrid] = dataclasses.field(default=Sendgrid.SENDGRID, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_time: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_time'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""Start time in ISO8601 format. Any data before this time point will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_sendinblue.py b/src/airbyte_api/models/shared/source_sendinblue.py
similarity index 96%
rename from src/airbyte/models/shared/source_sendinblue.py
rename to src/airbyte_api/models/shared/source_sendinblue.py
index bfc79885..41963254 100644
--- a/src/airbyte/models/shared/source_sendinblue.py
+++ b/src/airbyte_api/models/shared/source_sendinblue.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceSendinblue:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""Your API Key. See here."""
SOURCE_TYPE: Final[Sendinblue] = dataclasses.field(default=Sendinblue.SENDINBLUE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_senseforce.py b/src/airbyte_api/models/shared/source_senseforce.py
similarity index 99%
rename from src/airbyte/models/shared/source_senseforce.py
rename to src/airbyte_api/models/shared/source_senseforce.py
index 167ac137..eb6264d7 100644
--- a/src/airbyte/models/shared/source_senseforce.py
+++ b/src/airbyte_api/models/shared/source_senseforce.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -26,5 +26,3 @@ class SourceSenseforce:
slice_range: Optional[int] = dataclasses.field(default=10, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('slice_range'), 'exclude': lambda f: f is None }})
r"""The time increment used by the connector when requesting data from the Senseforce API. The bigger the value is, the less requests will be made and faster the sync will be. On the other hand, the more seldom the state is persisted and the more likely one could run into rate limites. Furthermore, consider that large chunks of time might take a long time for the Senseforce query to return data - meaning it could take in effect longer than with more smaller time slices. If there are a lot of data per day, set this setting to 1. If there is only very little data per day, you might change the setting to 10 or more."""
SOURCE_TYPE: Final[Senseforce] = dataclasses.field(default=Senseforce.SENSEFORCE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_sentry.py b/src/airbyte_api/models/shared/source_sentry.py
similarity index 98%
rename from src/airbyte/models/shared/source_sentry.py
rename to src/airbyte_api/models/shared/source_sentry.py
index d7947bfe..c8846323 100644
--- a/src/airbyte/models/shared/source_sentry.py
+++ b/src/airbyte_api/models/shared/source_sentry.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Any, Final, List, Optional
@@ -25,5 +25,3 @@ class SourceSentry:
hostname: Optional[str] = dataclasses.field(default='sentry.io', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('hostname'), 'exclude': lambda f: f is None }})
r"""Host name of Sentry API server.For self-hosted, specify your host name here. Otherwise, leave it empty."""
SOURCE_TYPE: Final[Sentry] = dataclasses.field(default=Sentry.SENTRY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_sftp.py b/src/airbyte_api/models/shared/source_sftp.py
similarity index 98%
rename from src/airbyte/models/shared/source_sftp.py
rename to src/airbyte_api/models/shared/source_sftp.py
index 8b6393a3..a4f5b11d 100644
--- a/src/airbyte/models/shared/source_sftp.py
+++ b/src/airbyte_api/models/shared/source_sftp.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -19,7 +19,7 @@ class SourceSftpSSHKeyAuthentication:
r"""OS-level user account ssh key credentials in RSA PEM format ( created with ssh-keygen -t rsa -m PEM -f myuser_rsa )"""
AUTH_METHOD: Final[SourceSftpSchemasAuthMethod] = dataclasses.field(default=SourceSftpSchemasAuthMethod.SSH_KEY_AUTH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method') }})
r"""Connect through ssh key"""
-
+
class SourceSftpAuthMethod(str, Enum):
@@ -34,7 +34,7 @@ class SourceSftpPasswordAuthentication:
r"""OS-level password for logging into the jump server host"""
AUTH_METHOD: Final[SourceSftpAuthMethod] = dataclasses.field(default=SourceSftpAuthMethod.SSH_PASSWORD_AUTH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method') }})
r"""Connect through password authentication"""
-
+
class Sftp(str, Enum):
@@ -59,5 +59,3 @@ class SourceSftp:
port: Optional[int] = dataclasses.field(default=22, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('port'), 'exclude': lambda f: f is None }})
r"""The server port"""
SOURCE_TYPE: Final[Sftp] = dataclasses.field(default=Sftp.SFTP, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_sftp_bulk.py b/src/airbyte_api/models/shared/source_sftp_bulk.py
similarity index 99%
rename from src/airbyte/models/shared/source_sftp_bulk.py
rename to src/airbyte_api/models/shared/source_sftp_bulk.py
index c93589e9..749190cd 100644
--- a/src/airbyte/models/shared/source_sftp_bulk.py
+++ b/src/airbyte_api/models/shared/source_sftp_bulk.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -46,5 +46,3 @@ class SourceSftpBulk:
separator: Optional[str] = dataclasses.field(default=',', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('separator'), 'exclude': lambda f: f is None }})
r"""The separator used in the CSV files. Define None if you want to use the Sniffer functionality"""
SOURCE_TYPE: Final[SftpBulk] = dataclasses.field(default=SftpBulk.SFTP_BULK, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_shopify.py b/src/airbyte_api/models/shared/source_shopify.py
similarity index 98%
rename from src/airbyte/models/shared/source_shopify.py
rename to src/airbyte_api/models/shared/source_shopify.py
index 922b7eac..77b6e9b0 100644
--- a/src/airbyte/models/shared/source_shopify.py
+++ b/src/airbyte_api/models/shared/source_shopify.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -20,7 +20,7 @@ class APIPassword:
api_password: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_password') }})
r"""The API Password for your private application in the `Shopify` store."""
AUTH_METHOD: Final[SourceShopifySchemasAuthMethod] = dataclasses.field(default=SourceShopifySchemasAuthMethod.API_PASSWORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method') }})
-
+
class SourceShopifyAuthMethod(str, Enum):
@@ -38,7 +38,7 @@ class SourceShopifyOAuth20:
r"""The Client ID of the Shopify developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of the Shopify developer application."""
-
+
class SourceShopifyShopify(str, Enum):
@@ -55,5 +55,3 @@ class SourceShopify:
SOURCE_TYPE: Final[SourceShopifyShopify] = dataclasses.field(default=SourceShopifyShopify.SHOPIFY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[date] = dataclasses.field(default=dateutil.parser.parse('2020-01-01').date(), metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""The date you would like to replicate data from. Format: YYYY-MM-DD. Any data before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_shortio.py b/src/airbyte_api/models/shared/source_shortio.py
similarity index 97%
rename from src/airbyte/models/shared/source_shortio.py
rename to src/airbyte_api/models/shared/source_shortio.py
index 72787f92..493d7a76 100644
--- a/src/airbyte/models/shared/source_shortio.py
+++ b/src/airbyte_api/models/shared/source_shortio.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -20,5 +20,3 @@ class SourceShortio:
start_date: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date') }})
r"""UTC date and time in the format 2017-01-25T00:00:00Z. Any data before this date will not be replicated."""
SOURCE_TYPE: Final[Shortio] = dataclasses.field(default=Shortio.SHORTIO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_slack.py b/src/airbyte_api/models/shared/source_slack.py
similarity index 99%
rename from src/airbyte/models/shared/source_slack.py
rename to src/airbyte_api/models/shared/source_slack.py
index e816645f..ff7cf7e7 100644
--- a/src/airbyte/models/shared/source_slack.py
+++ b/src/airbyte_api/models/shared/source_slack.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class SourceSlackAPIToken:
api_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_token') }})
r"""A Slack bot token. See the docs for instructions on how to generate it."""
OPTION_TITLE: Final[SourceSlackSchemasOptionTitle] = dataclasses.field(default=SourceSlackSchemasOptionTitle.API_TOKEN_CREDENTIALS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('option_title') }})
-
+
class SourceSlackOptionTitle(str, Enum):
@@ -36,7 +36,7 @@ class SignInViaSlackOAuth:
client_secret: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret') }})
r"""Slack client_secret. See our docs if you need help finding this secret."""
OPTION_TITLE: Final[SourceSlackOptionTitle] = dataclasses.field(default=SourceSlackOptionTitle.DEFAULT_O_AUTH2_0_AUTHORIZATION, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('option_title') }})
-
+
class SourceSlackSlack(str, Enum):
@@ -57,5 +57,3 @@ class SourceSlack:
lookback_window: Optional[int] = dataclasses.field(default=0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('lookback_window'), 'exclude': lambda f: f is None }})
r"""How far into the past to look for messages in threads, default is 0 days"""
SOURCE_TYPE: Final[SourceSlackSlack] = dataclasses.field(default=SourceSlackSlack.SLACK, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_smaily.py b/src/airbyte_api/models/shared/source_smaily.py
similarity index 97%
rename from src/airbyte/models/shared/source_smaily.py
rename to src/airbyte_api/models/shared/source_smaily.py
index 53992cf5..6668c4ff 100644
--- a/src/airbyte/models/shared/source_smaily.py
+++ b/src/airbyte_api/models/shared/source_smaily.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -21,5 +21,3 @@ class SourceSmaily:
api_username: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_username') }})
r"""API user username. See https://smaily.com/help/api/general/create-api-user/"""
SOURCE_TYPE: Final[Smaily] = dataclasses.field(default=Smaily.SMAILY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_smartengage.py b/src/airbyte_api/models/shared/source_smartengage.py
similarity index 95%
rename from src/airbyte/models/shared/source_smartengage.py
rename to src/airbyte_api/models/shared/source_smartengage.py
index 619a7cfa..3197398a 100644
--- a/src/airbyte/models/shared/source_smartengage.py
+++ b/src/airbyte_api/models/shared/source_smartengage.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceSmartengage:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""API Key"""
SOURCE_TYPE: Final[Smartengage] = dataclasses.field(default=Smartengage.SMARTENGAGE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_smartsheets.py b/src/airbyte_api/models/shared/source_smartsheets.py
similarity index 99%
rename from src/airbyte/models/shared/source_smartsheets.py
rename to src/airbyte_api/models/shared/source_smartsheets.py
index 5ff2546e..d78d02da 100644
--- a/src/airbyte/models/shared/source_smartsheets.py
+++ b/src/airbyte_api/models/shared/source_smartsheets.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class APIAccessToken:
access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token') }})
r"""The access token to use for accessing your data from Smartsheets. This access token must be generated by a user with at least read access to the data you'd like to replicate. Generate an access token in the Smartsheets main menu by clicking Account > Apps & Integrations > API Access. See the setup guide for information on how to obtain this token."""
AUTH_TYPE: Final[Optional[SourceSmartsheetsSchemasAuthType]] = dataclasses.field(default=SourceSmartsheetsSchemasAuthType.ACCESS_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceSmartsheetsAuthType(str, Enum):
@@ -40,7 +40,7 @@ class SourceSmartsheetsOAuth20:
token_expiry_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token_expiry_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""The date-time when the access token should be refreshed."""
AUTH_TYPE: Final[Optional[SourceSmartsheetsAuthType]] = dataclasses.field(default=SourceSmartsheetsAuthType.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class Validenums(str, Enum):
@@ -77,5 +77,3 @@ class SourceSmartsheets:
SOURCE_TYPE: Final[SourceSmartsheetsSmartsheets] = dataclasses.field(default=SourceSmartsheetsSmartsheets.SMARTSHEETS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_datetime: Optional[datetime] = dataclasses.field(default=dateutil.parser.isoparse('2020-01-01T00:00:00+00:00'), metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_datetime'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""Only rows modified after this date/time will be replicated. This should be an ISO 8601 string, for instance: `2000-01-01T13:00:00`"""
-
-
diff --git a/src/airbyte/models/shared/source_snapchat_marketing.py b/src/airbyte_api/models/shared/source_snapchat_marketing.py
similarity index 98%
rename from src/airbyte/models/shared/source_snapchat_marketing.py
rename to src/airbyte_api/models/shared/source_snapchat_marketing.py
index 305cab14..6bcac619 100644
--- a/src/airbyte/models/shared/source_snapchat_marketing.py
+++ b/src/airbyte_api/models/shared/source_snapchat_marketing.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -27,5 +27,3 @@ class SourceSnapchatMarketing:
SOURCE_TYPE: Final[SourceSnapchatMarketingSnapchatMarketing] = dataclasses.field(default=SourceSnapchatMarketingSnapchatMarketing.SNAPCHAT_MARKETING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[date] = dataclasses.field(default=dateutil.parser.parse('2022-01-01').date(), metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""Date in the format 2022-01-01. Any data before this date will not be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_snowflake.py b/src/airbyte_api/models/shared/source_snowflake.py
similarity index 99%
rename from src/airbyte/models/shared/source_snowflake.py
rename to src/airbyte_api/models/shared/source_snowflake.py
index 5c0f5e18..8f8cff69 100644
--- a/src/airbyte/models/shared/source_snowflake.py
+++ b/src/airbyte_api/models/shared/source_snowflake.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional, Union
@@ -19,7 +19,7 @@ class SourceSnowflakeUsernameAndPassword:
username: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('username') }})
r"""The username you created to allow Airbyte to access the database."""
AUTH_TYPE: Final[SourceSnowflakeSchemasAuthType] = dataclasses.field(default=SourceSnowflakeSchemasAuthType.USERNAME_PASSWORD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceSnowflakeAuthType(str, Enum):
@@ -38,7 +38,7 @@ class SourceSnowflakeOAuth20:
AUTH_TYPE: Final[SourceSnowflakeAuthType] = dataclasses.field(default=SourceSnowflakeAuthType.O_AUTH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
refresh_token: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token'), 'exclude': lambda f: f is None }})
r"""Refresh Token for making authenticated requests."""
-
+
class SourceSnowflakeSnowflake(str, Enum):
@@ -62,5 +62,3 @@ class SourceSnowflake:
schema: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('schema'), 'exclude': lambda f: f is None }})
r"""The source Snowflake schema tables. Leave empty to access tables from multiple schemas."""
SOURCE_TYPE: Final[SourceSnowflakeSnowflake] = dataclasses.field(default=SourceSnowflakeSnowflake.SNOWFLAKE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_sonar_cloud.py b/src/airbyte_api/models/shared/source_sonar_cloud.py
similarity index 98%
rename from src/airbyte/models/shared/source_sonar_cloud.py
rename to src/airbyte_api/models/shared/source_sonar_cloud.py
index 4cd0ab80..32eb2ded 100644
--- a/src/airbyte/models/shared/source_sonar_cloud.py
+++ b/src/airbyte_api/models/shared/source_sonar_cloud.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -26,5 +26,3 @@ class SourceSonarCloud:
SOURCE_TYPE: Final[SonarCloud] = dataclasses.field(default=SonarCloud.SONAR_CLOUD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[date] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""To retrieve issues created after the given date (inclusive)."""
-
-
diff --git a/src/airbyte/models/shared/source_spacex_api.py b/src/airbyte_api/models/shared/source_spacex_api.py
similarity index 96%
rename from src/airbyte/models/shared/source_spacex_api.py
rename to src/airbyte_api/models/shared/source_spacex_api.py
index c9903df9..b6b517bf 100644
--- a/src/airbyte/models/shared/source_spacex_api.py
+++ b/src/airbyte_api/models/shared/source_spacex_api.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -17,5 +17,3 @@ class SourceSpacexAPI:
id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('id'), 'exclude': lambda f: f is None }})
options: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('options'), 'exclude': lambda f: f is None }})
SOURCE_TYPE: Final[SpacexAPI] = dataclasses.field(default=SpacexAPI.SPACEX_API, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_square.py b/src/airbyte_api/models/shared/source_square.py
similarity index 98%
rename from src/airbyte/models/shared/source_square.py
rename to src/airbyte_api/models/shared/source_square.py
index bfb068d5..1508e0bd 100644
--- a/src/airbyte/models/shared/source_square.py
+++ b/src/airbyte_api/models/shared/source_square.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -19,7 +19,7 @@ class SourceSquareAPIKey:
api_key: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_key') }})
r"""The API key for a Square application"""
AUTH_TYPE: Final[SourceSquareSchemasAuthType] = dataclasses.field(default=SourceSquareSchemasAuthType.API_KEY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceSquareAuthType(str, Enum):
@@ -36,7 +36,7 @@ class OauthAuthentication:
refresh_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token') }})
r"""A refresh token generated using the above client ID and secret"""
AUTH_TYPE: Final[SourceSquareAuthType] = dataclasses.field(default=SourceSquareAuthType.O_AUTH, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type') }})
-
+
class SourceSquareSquare(str, Enum):
@@ -55,5 +55,3 @@ class SourceSquare:
SOURCE_TYPE: Final[SourceSquareSquare] = dataclasses.field(default=SourceSquareSquare.SQUARE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[date] = dataclasses.field(default=dateutil.parser.parse('2021-01-01').date(), metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""UTC date in the format YYYY-MM-DD. Any data before this date will not be replicated. If not set, all data will be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_strava.py b/src/airbyte_api/models/shared/source_strava.py
similarity index 98%
rename from src/airbyte/models/shared/source_strava.py
rename to src/airbyte_api/models/shared/source_strava.py
index 4f0f2205..6641b37e 100644
--- a/src/airbyte/models/shared/source_strava.py
+++ b/src/airbyte_api/models/shared/source_strava.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -31,5 +31,3 @@ class SourceStrava:
r"""UTC date and time. Any data before this date will not be replicated."""
AUTH_TYPE: Final[Optional[SourceStravaAuthType]] = dataclasses.field(default=SourceStravaAuthType.CLIENT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
SOURCE_TYPE: Final[SourceStravaStrava] = dataclasses.field(default=SourceStravaStrava.STRAVA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_stripe.py b/src/airbyte_api/models/shared/source_stripe.py
similarity index 99%
rename from src/airbyte/models/shared/source_stripe.py
rename to src/airbyte_api/models/shared/source_stripe.py
index de85b241..fdae5d4b 100644
--- a/src/airbyte/models/shared/source_stripe.py
+++ b/src/airbyte_api/models/shared/source_stripe.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -31,5 +31,3 @@ class SourceStripe:
SOURCE_TYPE: Final[Stripe] = dataclasses.field(default=Stripe.STRIPE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=dateutil.parser.isoparse('2017-01-25T00:00:00Z'), metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""UTC date and time in the format 2017-01-25T00:00:00Z. Only data generated after this date will be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_survey_sparrow.py b/src/airbyte_api/models/shared/source_survey_sparrow.py
similarity index 98%
rename from src/airbyte/models/shared/source_survey_sparrow.py
rename to src/airbyte_api/models/shared/source_survey_sparrow.py
index 2c8b7fda..df21ce57 100644
--- a/src/airbyte/models/shared/source_survey_sparrow.py
+++ b/src/airbyte_api/models/shared/source_survey_sparrow.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Any, Final, List, Optional, Union
@@ -15,7 +15,7 @@ class SourceSurveySparrowURLBase(str, Enum):
@dataclasses.dataclass
class GlobalAccount:
URL_BASE: Final[Optional[SourceSurveySparrowURLBase]] = dataclasses.field(default=SourceSurveySparrowURLBase.HTTPS_API_SURVEYSPARROW_COM_V3, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('url_base'), 'exclude': lambda f: f is None }})
-
+
class URLBase(str, Enum):
@@ -26,7 +26,7 @@ class URLBase(str, Enum):
@dataclasses.dataclass
class EUBasedAccount:
URL_BASE: Final[Optional[URLBase]] = dataclasses.field(default=URLBase.HTTPS_EU_API_SURVEYSPARROW_COM_V3, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('url_base'), 'exclude': lambda f: f is None }})
-
+
class SurveySparrow(str, Enum):
@@ -43,5 +43,3 @@ class SourceSurveySparrow:
SOURCE_TYPE: Final[SurveySparrow] = dataclasses.field(default=SurveySparrow.SURVEY_SPARROW, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
survey_id: Optional[List[Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('survey_id'), 'exclude': lambda f: f is None }})
r"""A List of your survey ids for survey-specific stream"""
-
-
diff --git a/src/airbyte/models/shared/source_surveymonkey.py b/src/airbyte_api/models/shared/source_surveymonkey.py
similarity index 98%
rename from src/airbyte/models/shared/source_surveymonkey.py
rename to src/airbyte_api/models/shared/source_surveymonkey.py
index 7f9a1221..7fb97d19 100644
--- a/src/airbyte/models/shared/source_surveymonkey.py
+++ b/src/airbyte_api/models/shared/source_surveymonkey.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -24,7 +24,7 @@ class SurveyMonkeyAuthorizationMethod:
r"""The Client ID of the SurveyMonkey developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of the SurveyMonkey developer application."""
-
+
class OriginDatacenterOfTheSurveyMonkeyAccount(str, Enum):
@@ -49,5 +49,3 @@ class SourceSurveymonkey:
SOURCE_TYPE: Final[SourceSurveymonkeySurveymonkey] = dataclasses.field(default=SourceSurveymonkeySurveymonkey.SURVEYMONKEY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
survey_ids: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('survey_ids'), 'exclude': lambda f: f is None }})
r"""IDs of the surveys from which you'd like to replicate data. If left empty, data from all boards to which you have access will be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_tempo.py b/src/airbyte_api/models/shared/source_tempo.py
similarity index 95%
rename from src/airbyte/models/shared/source_tempo.py
rename to src/airbyte_api/models/shared/source_tempo.py
index 4d4b5b54..1b3f96e5 100644
--- a/src/airbyte/models/shared/source_tempo.py
+++ b/src/airbyte_api/models/shared/source_tempo.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceTempo:
api_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_token') }})
r"""Tempo API Token. Go to Tempo>Settings, scroll down to Data Access and select API integration."""
SOURCE_TYPE: Final[Tempo] = dataclasses.field(default=Tempo.TEMPO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_the_guardian_api.py b/src/airbyte_api/models/shared/source_the_guardian_api.py
similarity index 98%
rename from src/airbyte/models/shared/source_the_guardian_api.py
rename to src/airbyte_api/models/shared/source_the_guardian_api.py
index b8305d8c..4523ff92 100644
--- a/src/airbyte/models/shared/source_the_guardian_api.py
+++ b/src/airbyte_api/models/shared/source_the_guardian_api.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -27,5 +27,3 @@ class SourceTheGuardianAPI:
SOURCE_TYPE: Final[TheGuardianAPI] = dataclasses.field(default=TheGuardianAPI.THE_GUARDIAN_API, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
tag: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('tag'), 'exclude': lambda f: f is None }})
r"""(Optional) A tag is a piece of data that is used by The Guardian to categorise content. Use this parameter to filter results by showing only the ones matching the entered tag. See here for a list of all tags, and here for the tags endpoint documentation."""
-
-
diff --git a/src/airbyte/models/shared/source_tiktok_marketing.py b/src/airbyte_api/models/shared/source_tiktok_marketing.py
similarity index 99%
rename from src/airbyte/models/shared/source_tiktok_marketing.py
rename to src/airbyte_api/models/shared/source_tiktok_marketing.py
index 8cf63450..af3bb29c 100644
--- a/src/airbyte/models/shared/source_tiktok_marketing.py
+++ b/src/airbyte_api/models/shared/source_tiktok_marketing.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -21,7 +21,7 @@ class SandboxAccessToken:
advertiser_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('advertiser_id') }})
r"""The Advertiser ID which generated for the developer's Sandbox application."""
AUTH_TYPE: Final[Optional[SourceTiktokMarketingSchemasAuthType]] = dataclasses.field(default=SourceTiktokMarketingSchemasAuthType.SANDBOX_ACCESS_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceTiktokMarketingAuthType(str, Enum):
@@ -40,7 +40,7 @@ class SourceTiktokMarketingOAuth20:
advertiser_id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('advertiser_id'), 'exclude': lambda f: f is None }})
r"""The Advertiser ID to filter reports and streams. Let this empty to retrieve all."""
AUTH_TYPE: Final[Optional[SourceTiktokMarketingAuthType]] = dataclasses.field(default=SourceTiktokMarketingAuthType.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceTiktokMarketingTiktokMarketing(str, Enum):
@@ -61,5 +61,3 @@ class SourceTiktokMarketing:
SOURCE_TYPE: Final[Optional[SourceTiktokMarketingTiktokMarketing]] = dataclasses.field(default=SourceTiktokMarketingTiktokMarketing.TIKTOK_MARKETING, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType'), 'exclude': lambda f: f is None }})
start_date: Optional[date] = dataclasses.field(default=dateutil.parser.parse('2016-09-01').date(), metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""The Start Date in format: YYYY-MM-DD. Any data before this date will not be replicated. If this parameter is not set, all data will be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_trello.py b/src/airbyte_api/models/shared/source_trello.py
similarity index 98%
rename from src/airbyte/models/shared/source_trello.py
rename to src/airbyte_api/models/shared/source_trello.py
index f3af6ad1..e2ee4bf9 100644
--- a/src/airbyte/models/shared/source_trello.py
+++ b/src/airbyte_api/models/shared/source_trello.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,5 +25,3 @@ class SourceTrello:
board_ids: Optional[List[str]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('board_ids'), 'exclude': lambda f: f is None }})
r"""IDs of the boards to replicate data from. If left empty, data from all boards to which you have access will be replicated. Please note that this is not the 8-character ID in the board's shortLink (URL of the board). Rather, what is required here is the 24-character ID usually returned by the API"""
SOURCE_TYPE: Final[Trello] = dataclasses.field(default=Trello.TRELLO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_trustpilot.py b/src/airbyte_api/models/shared/source_trustpilot.py
similarity index 98%
rename from src/airbyte/models/shared/source_trustpilot.py
rename to src/airbyte_api/models/shared/source_trustpilot.py
index 0b629e3b..0278545b 100644
--- a/src/airbyte/models/shared/source_trustpilot.py
+++ b/src/airbyte_api/models/shared/source_trustpilot.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -20,7 +20,7 @@ class SourceTrustpilotAPIKey:
client_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_id') }})
r"""The API key of the Trustpilot API application."""
AUTH_TYPE: Final[Optional[SourceTrustpilotSchemasAuthType]] = dataclasses.field(default=SourceTrustpilotSchemasAuthType.APIKEY, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceTrustpilotAuthType(str, Enum):
@@ -41,7 +41,7 @@ class SourceTrustpilotOAuth20:
token_expiry_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token_expiry_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""The date-time when the access token should be refreshed."""
AUTH_TYPE: Final[Optional[SourceTrustpilotAuthType]] = dataclasses.field(default=SourceTrustpilotAuthType.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class Trustpilot(str, Enum):
@@ -57,5 +57,3 @@ class SourceTrustpilot:
start_date: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date') }})
r"""For streams with sync. method incremental the start date time to be used"""
SOURCE_TYPE: Final[Trustpilot] = dataclasses.field(default=Trustpilot.TRUSTPILOT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_tvmaze_schedule.py b/src/airbyte_api/models/shared/source_tvmaze_schedule.py
similarity index 98%
rename from src/airbyte/models/shared/source_tvmaze_schedule.py
rename to src/airbyte_api/models/shared/source_tvmaze_schedule.py
index ed18e757..955f709d 100644
--- a/src/airbyte/models/shared/source_tvmaze_schedule.py
+++ b/src/airbyte_api/models/shared/source_tvmaze_schedule.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -26,5 +26,3 @@ class SourceTvmazeSchedule:
all countries plus global web channels (e.g. Netflix). Alternatively,
set to 'global' for just global web channels.
"""
-
-
diff --git a/src/airbyte/models/shared/source_twilio.py b/src/airbyte_api/models/shared/source_twilio.py
similarity index 97%
rename from src/airbyte/models/shared/source_twilio.py
rename to src/airbyte_api/models/shared/source_twilio.py
index ff931a09..b4dd9239 100644
--- a/src/airbyte/models/shared/source_twilio.py
+++ b/src/airbyte_api/models/shared/source_twilio.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,5 +25,3 @@ class SourceTwilio:
lookback_window: Optional[int] = dataclasses.field(default=0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('lookback_window'), 'exclude': lambda f: f is None }})
r"""How far into the past to look for records. (in minutes)"""
SOURCE_TYPE: Final[Twilio] = dataclasses.field(default=Twilio.TWILIO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_twilio_taskrouter.py b/src/airbyte_api/models/shared/source_twilio_taskrouter.py
similarity index 96%
rename from src/airbyte/models/shared/source_twilio_taskrouter.py
rename to src/airbyte_api/models/shared/source_twilio_taskrouter.py
index 5b050ab8..c7a8bfab 100644
--- a/src/airbyte/models/shared/source_twilio_taskrouter.py
+++ b/src/airbyte_api/models/shared/source_twilio_taskrouter.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -19,5 +19,3 @@ class SourceTwilioTaskrouter:
auth_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_token') }})
r"""Twilio Auth Token"""
SOURCE_TYPE: Final[TwilioTaskrouter] = dataclasses.field(default=TwilioTaskrouter.TWILIO_TASKROUTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_twitter.py b/src/airbyte_api/models/shared/source_twitter.py
similarity index 98%
rename from src/airbyte/models/shared/source_twitter.py
rename to src/airbyte_api/models/shared/source_twitter.py
index 6b513674..3f49e57d 100644
--- a/src/airbyte/models/shared/source_twitter.py
+++ b/src/airbyte_api/models/shared/source_twitter.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,5 +25,3 @@ class SourceTwitter:
SOURCE_TYPE: Final[Twitter] = dataclasses.field(default=Twitter.TWITTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""The start date for retrieving tweets cannot be more than 7 days in the past."""
-
-
diff --git a/src/airbyte/models/shared/source_typeform.py b/src/airbyte_api/models/shared/source_typeform.py
similarity index 98%
rename from src/airbyte/models/shared/source_typeform.py
rename to src/airbyte_api/models/shared/source_typeform.py
index 72670efe..8035f976 100644
--- a/src/airbyte/models/shared/source_typeform.py
+++ b/src/airbyte_api/models/shared/source_typeform.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class SourceTypeformPrivateToken:
access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token') }})
r"""Log into your Typeform account and then generate a personal Access Token."""
AUTH_TYPE: Final[Optional[SourceTypeformSchemasAuthType]] = dataclasses.field(default=SourceTypeformSchemasAuthType.ACCESS_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceTypeformAuthType(str, Enum):
@@ -40,7 +40,7 @@ class SourceTypeformOAuth20:
token_expiry_date: datetime = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('token_expiry_date'), 'encoder': utils.datetimeisoformat(False), 'decoder': dateutil.parser.isoparse }})
r"""The date-time when the access token should be refreshed."""
AUTH_TYPE: Final[Optional[SourceTypeformAuthType]] = dataclasses.field(default=SourceTypeformAuthType.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceTypeformTypeform(str, Enum):
@@ -56,5 +56,3 @@ class SourceTypeform:
SOURCE_TYPE: Final[SourceTypeformTypeform] = dataclasses.field(default=SourceTypeformTypeform.TYPEFORM, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""The date from which you'd like to replicate data for Typeform API, in the format YYYY-MM-DDT00:00:00Z. All data generated after this date will be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_us_census.py b/src/airbyte_api/models/shared/source_us_census.py
similarity index 97%
rename from src/airbyte/models/shared/source_us_census.py
rename to src/airbyte_api/models/shared/source_us_census.py
index cfe7b152..5830f016 100644
--- a/src/airbyte/models/shared/source_us_census.py
+++ b/src/airbyte_api/models/shared/source_us_census.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class SourceUsCensus:
query_params: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('query_params'), 'exclude': lambda f: f is None }})
r"""The query parameters portion of the GET request, without the api key"""
SOURCE_TYPE: Final[UsCensus] = dataclasses.field(default=UsCensus.US_CENSUS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_vantage.py b/src/airbyte_api/models/shared/source_vantage.py
similarity index 96%
rename from src/airbyte/models/shared/source_vantage.py
rename to src/airbyte_api/models/shared/source_vantage.py
index 196b1d1d..bf9b2099 100644
--- a/src/airbyte/models/shared/source_vantage.py
+++ b/src/airbyte_api/models/shared/source_vantage.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceVantage:
access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token') }})
r"""Your API Access token. See here."""
SOURCE_TYPE: Final[Vantage] = dataclasses.field(default=Vantage.VANTAGE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_webflow.py b/src/airbyte_api/models/shared/source_webflow.py
similarity index 97%
rename from src/airbyte/models/shared/source_webflow.py
rename to src/airbyte_api/models/shared/source_webflow.py
index 17d79913..527f0252 100644
--- a/src/airbyte/models/shared/source_webflow.py
+++ b/src/airbyte_api/models/shared/source_webflow.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -21,5 +21,3 @@ class SourceWebflow:
accept_version: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('accept_version'), 'exclude': lambda f: f is None }})
r"""The version of the Webflow API to use. See https://developers.webflow.com/#versioning"""
SOURCE_TYPE: Final[Webflow] = dataclasses.field(default=Webflow.WEBFLOW, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_whisky_hunter.py b/src/airbyte_api/models/shared/source_whisky_hunter.py
similarity index 95%
rename from src/airbyte/models/shared/source_whisky_hunter.py
rename to src/airbyte_api/models/shared/source_whisky_hunter.py
index 751152bb..9c5dad0f 100644
--- a/src/airbyte/models/shared/source_whisky_hunter.py
+++ b/src/airbyte_api/models/shared/source_whisky_hunter.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -15,5 +15,3 @@ class WhiskyHunter(str, Enum):
@dataclasses.dataclass
class SourceWhiskyHunter:
SOURCE_TYPE: Final[Optional[WhiskyHunter]] = dataclasses.field(default=WhiskyHunter.WHISKY_HUNTER, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/source_wikipedia_pageviews.py b/src/airbyte_api/models/shared/source_wikipedia_pageviews.py
similarity index 98%
rename from src/airbyte/models/shared/source_wikipedia_pageviews.py
rename to src/airbyte_api/models/shared/source_wikipedia_pageviews.py
index df8879a0..265557d3 100644
--- a/src/airbyte/models/shared/source_wikipedia_pageviews.py
+++ b/src/airbyte_api/models/shared/source_wikipedia_pageviews.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -29,5 +29,3 @@ class SourceWikipediaPageviews:
start: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start') }})
r"""The date of the first day to include, in YYYYMMDD or YYYYMMDDHH format."""
SOURCE_TYPE: Final[WikipediaPageviews] = dataclasses.field(default=WikipediaPageviews.WIKIPEDIA_PAGEVIEWS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_woocommerce.py b/src/airbyte_api/models/shared/source_woocommerce.py
similarity index 97%
rename from src/airbyte/models/shared/source_woocommerce.py
rename to src/airbyte_api/models/shared/source_woocommerce.py
index cc3f93c1..46f7ada5 100644
--- a/src/airbyte/models/shared/source_woocommerce.py
+++ b/src/airbyte_api/models/shared/source_woocommerce.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -24,5 +24,3 @@ class SourceWoocommerce:
start_date: date = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.dateisoformat(False), 'decoder': utils.datefromisoformat }})
r"""The date you would like to replicate data from. Format: YYYY-MM-DD"""
SOURCE_TYPE: Final[Woocommerce] = dataclasses.field(default=Woocommerce.WOOCOMMERCE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_xkcd.py b/src/airbyte_api/models/shared/source_xkcd.py
similarity index 94%
rename from src/airbyte/models/shared/source_xkcd.py
rename to src/airbyte_api/models/shared/source_xkcd.py
index bd0b2295..85845247 100644
--- a/src/airbyte/models/shared/source_xkcd.py
+++ b/src/airbyte_api/models/shared/source_xkcd.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -15,5 +15,3 @@ class Xkcd(str, Enum):
@dataclasses.dataclass
class SourceXkcd:
SOURCE_TYPE: Final[Optional[Xkcd]] = dataclasses.field(default=Xkcd.XKCD, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/source_yandex_metrica.py b/src/airbyte_api/models/shared/source_yandex_metrica.py
similarity index 98%
rename from src/airbyte/models/shared/source_yandex_metrica.py
rename to src/airbyte_api/models/shared/source_yandex_metrica.py
index 58271507..9dd1a53c 100644
--- a/src/airbyte/models/shared/source_yandex_metrica.py
+++ b/src/airbyte_api/models/shared/source_yandex_metrica.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import date
from enum import Enum
@@ -24,5 +24,3 @@ class SourceYandexMetrica:
end_date: Optional[date] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('end_date'), 'encoder': utils.dateisoformat(True), 'decoder': utils.datefromisoformat, 'exclude': lambda f: f is None }})
r"""Starting point for your data replication, in format of \\"YYYY-MM-DD\\". If not provided will sync till most recent date."""
SOURCE_TYPE: Final[YandexMetrica] = dataclasses.field(default=YandexMetrica.YANDEX_METRICA, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_yotpo.py b/src/airbyte_api/models/shared/source_yotpo.py
similarity index 98%
rename from src/airbyte/models/shared/source_yotpo.py
rename to src/airbyte_api/models/shared/source_yotpo.py
index 5fb944d8..cdbee3c6 100644
--- a/src/airbyte/models/shared/source_yotpo.py
+++ b/src/airbyte_api/models/shared/source_yotpo.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,5 +25,3 @@ class SourceYotpo:
email: Optional[str] = dataclasses.field(default='example@gmail.com', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('email'), 'exclude': lambda f: f is None }})
r"""Email address registered with yotpo."""
SOURCE_TYPE: Final[Yotpo] = dataclasses.field(default=Yotpo.YOTPO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_youtube_analytics.py b/src/airbyte_api/models/shared/source_youtube_analytics.py
similarity index 97%
rename from src/airbyte/models/shared/source_youtube_analytics.py
rename to src/airbyte_api/models/shared/source_youtube_analytics.py
index c974bf67..c57cff41 100644
--- a/src/airbyte/models/shared/source_youtube_analytics.py
+++ b/src/airbyte_api/models/shared/source_youtube_analytics.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Any, Dict, Final, Optional
@@ -19,7 +19,7 @@ class AuthenticateViaOAuth20:
refresh_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token') }})
r"""A refresh token generated using the above client ID and secret"""
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
-
+
class SourceYoutubeAnalyticsYoutubeAnalytics(str, Enum):
@@ -31,5 +31,3 @@ class SourceYoutubeAnalyticsYoutubeAnalytics(str, Enum):
class SourceYoutubeAnalytics:
credentials: AuthenticateViaOAuth20 = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials') }})
SOURCE_TYPE: Final[SourceYoutubeAnalyticsYoutubeAnalytics] = dataclasses.field(default=SourceYoutubeAnalyticsYoutubeAnalytics.YOUTUBE_ANALYTICS, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_zendesk_chat.py b/src/airbyte_api/models/shared/source_zendesk_chat.py
similarity index 98%
rename from src/airbyte/models/shared/source_zendesk_chat.py
rename to src/airbyte_api/models/shared/source_zendesk_chat.py
index 23dc8fbe..c6a43d18 100644
--- a/src/airbyte/models/shared/source_zendesk_chat.py
+++ b/src/airbyte_api/models/shared/source_zendesk_chat.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -19,7 +19,7 @@ class SourceZendeskChatAccessToken:
access_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('access_token') }})
r"""The Access Token to make authenticated requests."""
CREDENTIALS: Final[SourceZendeskChatSchemasCredentials] = dataclasses.field(default=SourceZendeskChatSchemasCredentials.ACCESS_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials') }})
-
+
class SourceZendeskChatCredentials(str, Enum):
@@ -38,7 +38,7 @@ class SourceZendeskChatOAuth20:
CREDENTIALS: Final[SourceZendeskChatCredentials] = dataclasses.field(default=SourceZendeskChatCredentials.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials') }})
refresh_token: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('refresh_token'), 'exclude': lambda f: f is None }})
r"""Refresh Token to obtain new Access Token, when it's expired."""
-
+
class SourceZendeskChatZendeskChat(str, Enum):
@@ -54,5 +54,3 @@ class SourceZendeskChat:
SOURCE_TYPE: Final[SourceZendeskChatZendeskChat] = dataclasses.field(default=SourceZendeskChatZendeskChat.ZENDESK_CHAT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
subdomain: Optional[str] = dataclasses.field(default='', metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('subdomain'), 'exclude': lambda f: f is None }})
r"""Required if you access Zendesk Chat from a Zendesk Support subdomain."""
-
-
diff --git a/src/airbyte/models/shared/source_zendesk_sell.py b/src/airbyte_api/models/shared/source_zendesk_sell.py
similarity index 95%
rename from src/airbyte/models/shared/source_zendesk_sell.py
rename to src/airbyte_api/models/shared/source_zendesk_sell.py
index 12d7a805..5331e9fd 100644
--- a/src/airbyte/models/shared/source_zendesk_sell.py
+++ b/src/airbyte_api/models/shared/source_zendesk_sell.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceZendeskSell:
api_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('api_token') }})
r"""The API token for authenticating to Zendesk Sell"""
SOURCE_TYPE: Final[ZendeskSell] = dataclasses.field(default=ZendeskSell.ZENDESK_SELL, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_zendesk_sunshine.py b/src/airbyte_api/models/shared/source_zendesk_sunshine.py
similarity index 98%
rename from src/airbyte/models/shared/source_zendesk_sunshine.py
rename to src/airbyte_api/models/shared/source_zendesk_sunshine.py
index 609e1de5..fbedee5e 100644
--- a/src/airbyte/models/shared/source_zendesk_sunshine.py
+++ b/src/airbyte_api/models/shared/source_zendesk_sunshine.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -21,7 +21,7 @@ class SourceZendeskSunshineAPIToken:
email: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('email') }})
r"""The user email for your Zendesk account"""
AUTH_METHOD: Final[Optional[SourceZendeskSunshineSchemasAuthMethod]] = dataclasses.field(default=SourceZendeskSunshineSchemasAuthMethod.API_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method'), 'exclude': lambda f: f is None }})
-
+
class SourceZendeskSunshineAuthMethod(str, Enum):
@@ -38,7 +38,7 @@ class SourceZendeskSunshineOAuth20:
client_secret: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret') }})
r"""The Client Secret of your OAuth application."""
AUTH_METHOD: Final[Optional[SourceZendeskSunshineAuthMethod]] = dataclasses.field(default=SourceZendeskSunshineAuthMethod.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_method'), 'exclude': lambda f: f is None }})
-
+
class SourceZendeskSunshineZendeskSunshine(str, Enum):
@@ -54,5 +54,3 @@ class SourceZendeskSunshine:
r"""The subdomain for your Zendesk Account."""
credentials: Optional[Union[SourceZendeskSunshineOAuth20, SourceZendeskSunshineAPIToken]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
SOURCE_TYPE: Final[SourceZendeskSunshineZendeskSunshine] = dataclasses.field(default=SourceZendeskSunshineZendeskSunshine.ZENDESK_SUNSHINE, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_zendesk_support.py b/src/airbyte_api/models/shared/source_zendesk_support.py
similarity index 99%
rename from src/airbyte/models/shared/source_zendesk_support.py
rename to src/airbyte_api/models/shared/source_zendesk_support.py
index 5a5267ab..f42a6000 100644
--- a/src/airbyte/models/shared/source_zendesk_support.py
+++ b/src/airbyte_api/models/shared/source_zendesk_support.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -23,7 +23,7 @@ class SourceZendeskSupportAPIToken:
r"""The user email for your Zendesk account."""
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
CREDENTIALS: Final[Optional[SourceZendeskSupportSchemasCredentials]] = dataclasses.field(default=SourceZendeskSupportSchemasCredentials.API_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
+
class SourceZendeskSupportCredentials(str, Enum):
@@ -42,7 +42,7 @@ class SourceZendeskSupportOAuth20:
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The OAuth client secret. See this guide for more information."""
CREDENTIALS: Final[Optional[SourceZendeskSupportCredentials]] = dataclasses.field(default=SourceZendeskSupportCredentials.OAUTH2_0, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
+
class SourceZendeskSupportZendeskSupport(str, Enum):
@@ -61,5 +61,3 @@ class SourceZendeskSupport:
SOURCE_TYPE: Final[SourceZendeskSupportZendeskSupport] = dataclasses.field(default=SourceZendeskSupportZendeskSupport.ZENDESK_SUPPORT, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_date: Optional[datetime] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_date'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is None }})
r"""The UTC date and time from which you'd like to replicate data, in the format YYYY-MM-DDT00:00:00Z. All data generated after this date will be replicated."""
-
-
diff --git a/src/airbyte/models/shared/source_zendesk_talk.py b/src/airbyte_api/models/shared/source_zendesk_talk.py
similarity index 99%
rename from src/airbyte/models/shared/source_zendesk_talk.py
rename to src/airbyte_api/models/shared/source_zendesk_talk.py
index 5bf34695..22a9343c 100644
--- a/src/airbyte/models/shared/source_zendesk_talk.py
+++ b/src/airbyte_api/models/shared/source_zendesk_talk.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -25,7 +25,7 @@ class SourceZendeskTalkOAuth20:
r"""Client ID"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""Client Secret"""
-
+
class SourceZendeskTalkAuthType(str, Enum):
@@ -42,7 +42,7 @@ class SourceZendeskTalkAPIToken:
r"""The user email for your Zendesk account."""
additional_properties: Optional[Dict[str, Any]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'exclude': lambda f: f is None }})
AUTH_TYPE: Final[Optional[SourceZendeskTalkAuthType]] = dataclasses.field(default=SourceZendeskTalkAuthType.API_TOKEN, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('auth_type'), 'exclude': lambda f: f is None }})
-
+
class SourceZendeskTalkZendeskTalk(str, Enum):
@@ -59,5 +59,3 @@ class SourceZendeskTalk:
credentials: Optional[Union[SourceZendeskTalkAPIToken, SourceZendeskTalkOAuth20]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
r"""Zendesk service provides two authentication methods. Choose between: `OAuth2.0` or `API token`."""
SOURCE_TYPE: Final[SourceZendeskTalkZendeskTalk] = dataclasses.field(default=SourceZendeskTalkZendeskTalk.ZENDESK_TALK, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/source_zenloop.py b/src/airbyte_api/models/shared/source_zenloop.py
similarity index 98%
rename from src/airbyte/models/shared/source_zenloop.py
rename to src/airbyte_api/models/shared/source_zenloop.py
index cf929340..65201147 100644
--- a/src/airbyte/models/shared/source_zenloop.py
+++ b/src/airbyte_api/models/shared/source_zenloop.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final, Optional
@@ -23,5 +23,3 @@ class SourceZenloop:
r"""Zenloop Survey Group ID. Can be found by pulling All Survey Groups via SurveyGroups stream. Leave empty to pull answers from all survey groups"""
survey_id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('survey_id'), 'exclude': lambda f: f is None }})
r"""Zenloop Survey ID. Can be found here. Leave empty to pull answers from all surveys"""
-
-
diff --git a/src/airbyte/models/shared/source_zoho_crm.py b/src/airbyte_api/models/shared/source_zoho_crm.py
similarity index 98%
rename from src/airbyte/models/shared/source_zoho_crm.py
rename to src/airbyte_api/models/shared/source_zoho_crm.py
index 4e60cb63..037462a1 100644
--- a/src/airbyte/models/shared/source_zoho_crm.py
+++ b/src/airbyte_api/models/shared/source_zoho_crm.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
import dateutil.parser
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from datetime import datetime
from enum import Enum
@@ -55,5 +55,3 @@ class SourceZohoCrm:
SOURCE_TYPE: Final[ZohoCrm] = dataclasses.field(default=ZohoCrm.ZOHO_CRM, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
start_datetime: Optional[datetime] = dataclasses.field(default=UNSET, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('start_datetime'), 'encoder': utils.datetimeisoformat(True), 'decoder': dateutil.parser.isoparse, 'exclude': lambda f: f is SourceZohoCrm.UNSET }})
r"""ISO 8601, for instance: `YYYY-MM-DD`, `YYYY-MM-DD HH:MM:SS+HH:MM`"""
-
-
diff --git a/src/airbyte/models/shared/source_zoom.py b/src/airbyte_api/models/shared/source_zoom.py
similarity index 95%
rename from src/airbyte/models/shared/source_zoom.py
rename to src/airbyte_api/models/shared/source_zoom.py
index 5bc7cb8e..aa0c4a85 100644
--- a/src/airbyte/models/shared/source_zoom.py
+++ b/src/airbyte_api/models/shared/source_zoom.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from enum import Enum
from typing import Final
@@ -17,5 +17,3 @@ class SourceZoom:
jwt_token: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('jwt_token') }})
r"""JWT Token"""
SOURCE_TYPE: Final[Zoom] = dataclasses.field(default=Zoom.ZOOM, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
-
-
diff --git a/src/airbyte/models/shared/sourcecreaterequest.py b/src/airbyte_api/models/shared/sourcecreaterequest.py
similarity index 99%
rename from src/airbyte/models/shared/sourcecreaterequest.py
rename to src/airbyte_api/models/shared/sourcecreaterequest.py
index f6feff68..fe87ab3f 100644
--- a/src/airbyte/models/shared/sourcecreaterequest.py
+++ b/src/airbyte_api/models/shared/sourcecreaterequest.py
@@ -195,7 +195,7 @@
from .source_zenloop import SourceZenloop
from .source_zoho_crm import SourceZohoCrm
from .source_zoom import SourceZoom
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional, Union
@@ -212,5 +212,3 @@ class SourceCreateRequest:
r"""The UUID of the connector definition. One of configuration.sourceType or definitionId must be provided."""
secret_id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('secretId'), 'exclude': lambda f: f is None }})
r"""Optional secretID obtained through the public API OAuth redirect flow."""
-
-
diff --git a/src/airbyte/models/shared/sourcepatchrequest.py b/src/airbyte_api/models/shared/sourcepatchrequest.py
similarity index 99%
rename from src/airbyte/models/shared/sourcepatchrequest.py
rename to src/airbyte_api/models/shared/sourcepatchrequest.py
index e6f34c24..fdb9315c 100644
--- a/src/airbyte/models/shared/sourcepatchrequest.py
+++ b/src/airbyte_api/models/shared/sourcepatchrequest.py
@@ -195,7 +195,7 @@
from .source_zenloop import SourceZenloop
from .source_zoho_crm import SourceZohoCrm
from .source_zoom import SourceZoom
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional, Union
@@ -209,5 +209,3 @@ class SourcePatchRequest:
secret_id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('secretId'), 'exclude': lambda f: f is None }})
r"""Optional secretID obtained through the public API OAuth redirect flow."""
workspace_id: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('workspaceId'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/sourceputrequest.py b/src/airbyte_api/models/shared/sourceputrequest.py
similarity index 99%
rename from src/airbyte/models/shared/sourceputrequest.py
rename to src/airbyte_api/models/shared/sourceputrequest.py
index 79079af5..6f7f5f25 100644
--- a/src/airbyte/models/shared/sourceputrequest.py
+++ b/src/airbyte_api/models/shared/sourceputrequest.py
@@ -195,7 +195,7 @@
from .source_zenloop import SourceZenloop
from .source_zoho_crm import SourceZohoCrm
from .source_zoom import SourceZoom
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Union
@@ -206,5 +206,3 @@ class SourcePutRequest:
configuration: Union[SourceAha, SourceAircall, SourceAirtable, SourceAmazonAds, SourceAmazonSellerPartner, SourceAmazonSqs, SourceAmplitude, SourceApifyDataset, SourceAppfollow, SourceAsana, SourceAuth0, SourceAwsCloudtrail, SourceAzureBlobStorage, SourceAzureTable, SourceBambooHr, SourceBigquery, SourceBingAds, SourceBraintree, SourceBraze, SourceCart, SourceChargebee, SourceChartmogul, SourceClickhouse, SourceClickupAPI, SourceClockify, SourceCloseCom, SourceCoda, SourceCoinAPI, SourceCoinmarketcap, SourceConfigcat, SourceConfluence, SourceConvex, SourceDatascope, SourceDelighted, SourceDixa, SourceDockerhub, SourceDremio, SourceDynamodb, Union[ContinuousFeed], SourceEmailoctopus, SourceExchangeRates, SourceFacebookMarketing, SourceFaker, SourceFauna, SourceFile, SourceFirebolt, SourceFreshcaller, SourceFreshdesk, SourceFreshsales, SourceGainsightPx, SourceGcs, SourceGetlago, SourceGithub, SourceGitlab, SourceGlassfrog, SourceGnews, SourceGoogleAds, SourceGoogleAnalyticsDataAPI, SourceGoogleAnalyticsV4ServiceAccountOnly, SourceGoogleDirectory, SourceGoogleDrive, SourceGooglePagespeedInsights, SourceGoogleSearchConsole, SourceGoogleSheets, SourceGoogleWebfonts, SourceGoogleWorkspaceAdminReports, SourceGreenhouse, SourceGridly, SourceHarvest, SourceHubplanner, SourceHubspot, SourceInsightly, SourceInstagram, SourceInstatus, SourceIntercom, SourceIp2whois, SourceIterable, SourceJira, SourceK6Cloud, SourceKlarna, SourceKlaviyo, SourceKyve, SourceLaunchdarkly, SourceLemlist, SourceLeverHiring, SourceLinkedinAds, SourceLinkedinPages, SourceLokalise, SourceMailchimp, SourceMailgun, SourceMailjetSms, SourceMarketo, SourceMetabase, SourceMicrosoftSharepoint, SourceMicrosoftTeams, SourceMixpanel, SourceMonday, SourceMongodbInternalPoc, SourceMongodbV2, SourceMssql, SourceMyHours, SourceMysql, SourceNetsuite, SourceNotion, SourceNytimes, SourceOkta, SourceOmnisend, SourceOnesignal, SourceOracle, SourceOrb, SourceOrbit, SourceOutbrainAmplify, SourceOutreach, SourcePaypalTransaction, SourcePaystack, SourcePendo, SourcePersistiq, SourcePexelsAPI, SourcePinterest, SourcePipedrive, SourcePocket, SourcePokeapi, SourcePolygonStockAPI, SourcePostgres, SourcePosthog, SourcePostmarkapp, SourcePrestashop, SourcePunkAPI, SourcePypi, SourceQualaroo, SourceQuickbooks, SourceRailz, SourceRecharge, SourceRecreation, SourceRecruitee, SourceRedshift, SourceRetently, SourceRkiCovid, SourceRss, SourceS3, SourceSalesforce, SourceSalesloft, SourceSapFieldglass, SourceSecoda, SourceSendgrid, SourceSendinblue, SourceSenseforce, SourceSentry, SourceSftp, SourceSftpBulk, SourceShopify, SourceShortio, SourceSlack, SourceSmaily, SourceSmartengage, SourceSmartsheets, SourceSnapchatMarketing, SourceSnowflake, SourceSonarCloud, SourceSpacexAPI, SourceSquare, SourceStrava, SourceStripe, SourceSurveySparrow, SourceSurveymonkey, SourceTempo, SourceTheGuardianAPI, SourceTiktokMarketing, SourceTrello, SourceTrustpilot, SourceTvmazeSchedule, SourceTwilio, SourceTwilioTaskrouter, SourceTwitter, SourceTypeform, SourceUsCensus, SourceVantage, SourceWebflow, SourceWhiskyHunter, SourceWikipediaPageviews, SourceWoocommerce, SourceXkcd, SourceYandexMetrica, SourceYotpo, SourceYoutubeAnalytics, SourceZendeskChat, SourceZendeskSell, SourceZendeskSunshine, SourceZendeskSupport, SourceZendeskTalk, SourceZenloop, SourceZohoCrm, SourceZoom] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('configuration') }})
r"""The values required to configure the source."""
name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('name') }})
-
-
diff --git a/src/airbyte/models/shared/sourceresponse.py b/src/airbyte_api/models/shared/sourceresponse.py
similarity index 99%
rename from src/airbyte/models/shared/sourceresponse.py
rename to src/airbyte_api/models/shared/sourceresponse.py
index 6abbc2bb..6b9cfaa3 100644
--- a/src/airbyte/models/shared/sourceresponse.py
+++ b/src/airbyte_api/models/shared/sourceresponse.py
@@ -195,7 +195,7 @@
from .source_zenloop import SourceZenloop
from .source_zoho_crm import SourceZohoCrm
from .source_zoom import SourceZoom
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Union
@@ -210,5 +210,3 @@ class SourceResponse:
source_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceId') }})
source_type: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceType') }})
workspace_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('workspaceId') }})
-
-
diff --git a/src/airbyte/models/shared/sourcesresponse.py b/src/airbyte_api/models/shared/sourcesresponse.py
similarity index 96%
rename from src/airbyte/models/shared/sourcesresponse.py
rename to src/airbyte_api/models/shared/sourcesresponse.py
index 610c4089..e60fc3b4 100644
--- a/src/airbyte/models/shared/sourcesresponse.py
+++ b/src/airbyte_api/models/shared/sourcesresponse.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .sourceresponse import SourceResponse
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import List, Optional
@@ -14,5 +14,3 @@ class SourcesResponse:
data: List[SourceResponse] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('data') }})
next: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('next'), 'exclude': lambda f: f is None }})
previous: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('previous'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/square.py b/src/airbyte_api/models/shared/square.py
similarity index 96%
rename from src/airbyte/models/shared/square.py
rename to src/airbyte_api/models/shared/square.py
index d03e2afb..86b3d9d8 100644
--- a/src/airbyte/models/shared/square.py
+++ b/src/airbyte_api/models/shared/square.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class SquareCredentials:
r"""The Square-issued ID of your application"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Square-issued application secret for your application"""
-
+
@@ -22,5 +22,3 @@ class SquareCredentials:
@dataclasses.dataclass
class Square:
credentials: Optional[SquareCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/strava.py b/src/airbyte_api/models/shared/strava.py
similarity index 96%
rename from src/airbyte/models/shared/strava.py
rename to src/airbyte_api/models/shared/strava.py
index b72052c6..fcf7d7ab 100644
--- a/src/airbyte/models/shared/strava.py
+++ b/src/airbyte_api/models/shared/strava.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,5 +14,3 @@ class Strava:
r"""The Client ID of your Strava developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your Strava developer application."""
-
-
diff --git a/src/airbyte/models/shared/streamconfiguration.py b/src/airbyte_api/models/shared/streamconfiguration.py
similarity index 97%
rename from src/airbyte/models/shared/streamconfiguration.py
rename to src/airbyte_api/models/shared/streamconfiguration.py
index 0fd9c9d4..21b56ced 100644
--- a/src/airbyte/models/shared/streamconfiguration.py
+++ b/src/airbyte_api/models/shared/streamconfiguration.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .connectionsyncmodeenum import ConnectionSyncModeEnum
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import List, Optional
@@ -18,5 +18,3 @@ class StreamConfiguration:
primary_key: Optional[List[List[str]]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('primaryKey'), 'exclude': lambda f: f is None }})
r"""Paths to the fields that will be used as primary key. This field is REQUIRED if `destination_sync_mode` is `*_dedup` unless it is already supplied by the source schema."""
sync_mode: Optional[ConnectionSyncModeEnum] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('syncMode'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/streamconfigurations.py b/src/airbyte_api/models/shared/streamconfigurations.py
similarity index 95%
rename from src/airbyte/models/shared/streamconfigurations.py
rename to src/airbyte_api/models/shared/streamconfigurations.py
index c2a389c2..6ef7324e 100644
--- a/src/airbyte/models/shared/streamconfigurations.py
+++ b/src/airbyte_api/models/shared/streamconfigurations.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .streamconfiguration import StreamConfiguration
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import List, Optional
@@ -13,5 +13,3 @@
class StreamConfigurations:
r"""A list of configured stream options for a connection."""
streams: Optional[List[StreamConfiguration]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('streams'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/streamproperties.py b/src/airbyte_api/models/shared/streamproperties.py
similarity index 98%
rename from src/airbyte/models/shared/streamproperties.py
rename to src/airbyte_api/models/shared/streamproperties.py
index 04ea479e..29c68a0d 100644
--- a/src/airbyte/models/shared/streamproperties.py
+++ b/src/airbyte_api/models/shared/streamproperties.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .connectionsyncmodeenum import ConnectionSyncModeEnum
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import List, Optional
@@ -18,5 +18,3 @@ class StreamProperties:
source_defined_primary_key: Optional[List[List[str]]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('sourceDefinedPrimaryKey'), 'exclude': lambda f: f is None }})
stream_name: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('streamName'), 'exclude': lambda f: f is None }})
sync_modes: Optional[List[ConnectionSyncModeEnum]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('syncModes'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/streampropertiesresponse.py b/src/airbyte_api/models/shared/streampropertiesresponse.py
similarity index 94%
rename from src/airbyte/models/shared/streampropertiesresponse.py
rename to src/airbyte_api/models/shared/streampropertiesresponse.py
index a658c976..c2ff14b2 100644
--- a/src/airbyte/models/shared/streampropertiesresponse.py
+++ b/src/airbyte_api/models/shared/streampropertiesresponse.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .streamproperties import StreamProperties
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import List, Optional
@@ -13,5 +13,3 @@
class StreamPropertiesResponse:
r"""A list of stream properties."""
streams: Optional[List[StreamProperties]] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('streams'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/surveymonkey.py b/src/airbyte_api/models/shared/surveymonkey.py
similarity index 96%
rename from src/airbyte/models/shared/surveymonkey.py
rename to src/airbyte_api/models/shared/surveymonkey.py
index 56125631..1028fa82 100644
--- a/src/airbyte/models/shared/surveymonkey.py
+++ b/src/airbyte_api/models/shared/surveymonkey.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class SurveymonkeyCredentials:
r"""The Client ID of the SurveyMonkey developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of the SurveyMonkey developer application."""
-
+
@@ -22,5 +22,3 @@ class SurveymonkeyCredentials:
@dataclasses.dataclass
class Surveymonkey:
credentials: Optional[SurveymonkeyCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/tiktok_marketing.py b/src/airbyte_api/models/shared/tiktok_marketing.py
similarity index 96%
rename from src/airbyte/models/shared/tiktok_marketing.py
rename to src/airbyte_api/models/shared/tiktok_marketing.py
index 5077f53a..58656eb7 100644
--- a/src/airbyte/models/shared/tiktok_marketing.py
+++ b/src/airbyte_api/models/shared/tiktok_marketing.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class TiktokMarketingCredentials:
r"""The Developer Application App ID."""
secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('secret'), 'exclude': lambda f: f is None }})
r"""The Developer Application Secret."""
-
+
@@ -22,5 +22,3 @@ class TiktokMarketingCredentials:
@dataclasses.dataclass
class TiktokMarketing:
credentials: Optional[TiktokMarketingCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/typeform.py b/src/airbyte_api/models/shared/typeform.py
similarity index 96%
rename from src/airbyte/models/shared/typeform.py
rename to src/airbyte_api/models/shared/typeform.py
index 27072aae..85248adf 100644
--- a/src/airbyte/models/shared/typeform.py
+++ b/src/airbyte_api/models/shared/typeform.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class TypeformCredentials:
r"""The Client ID of the Typeform developer application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret the Typeform developer application."""
-
+
@@ -22,5 +22,3 @@ class TypeformCredentials:
@dataclasses.dataclass
class Typeform:
credentials: Optional[TypeformCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/workspacecreaterequest.py b/src/airbyte_api/models/shared/workspacecreaterequest.py
similarity index 93%
rename from src/airbyte/models/shared/workspacecreaterequest.py
rename to src/airbyte_api/models/shared/workspacecreaterequest.py
index f72cc098..6842d39b 100644
--- a/src/airbyte/models/shared/workspacecreaterequest.py
+++ b/src/airbyte_api/models/shared/workspacecreaterequest.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
@@ -11,5 +11,3 @@
class WorkspaceCreateRequest:
name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('name') }})
r"""Name of the workspace"""
-
-
diff --git a/src/airbyte/models/shared/workspaceoauthcredentialsrequest.py b/src/airbyte_api/models/shared/workspaceoauthcredentialsrequest.py
similarity index 98%
rename from src/airbyte/models/shared/workspaceoauthcredentialsrequest.py
rename to src/airbyte_api/models/shared/workspaceoauthcredentialsrequest.py
index e11a54b8..4695541d 100644
--- a/src/airbyte/models/shared/workspaceoauthcredentialsrequest.py
+++ b/src/airbyte_api/models/shared/workspaceoauthcredentialsrequest.py
@@ -46,7 +46,7 @@
from .zendesk_sunshine import ZendeskSunshine
from .zendesk_support import ZendeskSupport
from .zendesk_talk import ZendeskTalk
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Any, Union
@@ -60,5 +60,3 @@ class WorkspaceOAuthCredentialsRequest:
configuration: Union[Airtable, AmazonAds, AmazonSellerPartner, Asana, BingAds, FacebookMarketing, Github, Gitlab, GoogleAds, GoogleAnalyticsDataAPI, GoogleDrive, GoogleSearchConsole, GoogleSheets, Harvest, Hubspot, Instagram, Intercom, LeverHiring, LinkedinAds, Mailchimp, MicrosoftSharepoint, MicrosoftTeams, Monday, Notion, Pinterest, Retently, Salesforce, Shopify, Slack, Smartsheets, SnapchatMarketing, Snowflake, Square, Strava, Surveymonkey, TiktokMarketing, Any, Typeform, YoutubeAnalytics, ZendeskChat, ZendeskSunshine, ZendeskSupport, ZendeskTalk] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('configuration') }})
r"""The values required to configure the source."""
name: OAuthActorNames = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('name') }})
-
-
diff --git a/src/airbyte/models/shared/workspaceresponse.py b/src/airbyte_api/models/shared/workspaceresponse.py
similarity index 96%
rename from src/airbyte/models/shared/workspaceresponse.py
rename to src/airbyte_api/models/shared/workspaceresponse.py
index 26d84770..ac67008c 100644
--- a/src/airbyte/models/shared/workspaceresponse.py
+++ b/src/airbyte_api/models/shared/workspaceresponse.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .geographyenum import GeographyEnum
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -15,5 +15,3 @@ class WorkspaceResponse:
name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('name') }})
workspace_id: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('workspaceId') }})
data_residency: Optional[GeographyEnum] = dataclasses.field(default=GeographyEnum.AUTO, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('dataResidency'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/workspacesresponse.py b/src/airbyte_api/models/shared/workspacesresponse.py
similarity index 96%
rename from src/airbyte/models/shared/workspacesresponse.py
rename to src/airbyte_api/models/shared/workspacesresponse.py
index 920e1457..bc690b6d 100644
--- a/src/airbyte/models/shared/workspacesresponse.py
+++ b/src/airbyte_api/models/shared/workspacesresponse.py
@@ -3,7 +3,7 @@
from __future__ import annotations
import dataclasses
from .workspaceresponse import WorkspaceResponse
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import List, Optional
@@ -14,5 +14,3 @@ class WorkspacesResponse:
data: List[WorkspaceResponse] = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('data') }})
next: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('next'), 'exclude': lambda f: f is None }})
previous: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('previous'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/workspaceupdaterequest.py b/src/airbyte_api/models/shared/workspaceupdaterequest.py
similarity index 93%
rename from src/airbyte/models/shared/workspaceupdaterequest.py
rename to src/airbyte_api/models/shared/workspaceupdaterequest.py
index 4a16bf0a..b525f3c2 100644
--- a/src/airbyte/models/shared/workspaceupdaterequest.py
+++ b/src/airbyte_api/models/shared/workspaceupdaterequest.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
@@ -11,5 +11,3 @@
class WorkspaceUpdateRequest:
name: str = dataclasses.field(metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('name') }})
r"""Name of the workspace"""
-
-
diff --git a/src/airbyte/models/shared/youtube_analytics.py b/src/airbyte_api/models/shared/youtube_analytics.py
similarity index 96%
rename from src/airbyte/models/shared/youtube_analytics.py
rename to src/airbyte_api/models/shared/youtube_analytics.py
index bb5c5777..6ccf1030 100644
--- a/src/airbyte/models/shared/youtube_analytics.py
+++ b/src/airbyte_api/models/shared/youtube_analytics.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class YoutubeAnalyticsCredentials:
r"""The Client ID of your developer application"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The client secret of your developer application"""
-
+
@@ -22,5 +22,3 @@ class YoutubeAnalyticsCredentials:
@dataclasses.dataclass
class YoutubeAnalytics:
credentials: Optional[YoutubeAnalyticsCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/zendesk_chat.py b/src/airbyte_api/models/shared/zendesk_chat.py
similarity index 96%
rename from src/airbyte/models/shared/zendesk_chat.py
rename to src/airbyte_api/models/shared/zendesk_chat.py
index cc2748c2..0f3e7c72 100644
--- a/src/airbyte/models/shared/zendesk_chat.py
+++ b/src/airbyte_api/models/shared/zendesk_chat.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class ZendeskChatCredentials:
r"""The Client ID of your OAuth application"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your OAuth application."""
-
+
@@ -22,5 +22,3 @@ class ZendeskChatCredentials:
@dataclasses.dataclass
class ZendeskChat:
credentials: Optional[ZendeskChatCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/zendesk_sunshine.py b/src/airbyte_api/models/shared/zendesk_sunshine.py
similarity index 96%
rename from src/airbyte/models/shared/zendesk_sunshine.py
rename to src/airbyte_api/models/shared/zendesk_sunshine.py
index c761dd58..fb2d0312 100644
--- a/src/airbyte/models/shared/zendesk_sunshine.py
+++ b/src/airbyte_api/models/shared/zendesk_sunshine.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class ZendeskSunshineCredentials:
r"""The Client ID of your OAuth application."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The Client Secret of your OAuth application."""
-
+
@@ -22,5 +22,3 @@ class ZendeskSunshineCredentials:
@dataclasses.dataclass
class ZendeskSunshine:
credentials: Optional[ZendeskSunshineCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/zendesk_support.py b/src/airbyte_api/models/shared/zendesk_support.py
similarity index 97%
rename from src/airbyte/models/shared/zendesk_support.py
rename to src/airbyte_api/models/shared/zendesk_support.py
index 364fe2b3..9f905867 100644
--- a/src/airbyte/models/shared/zendesk_support.py
+++ b/src/airbyte_api/models/shared/zendesk_support.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class ZendeskSupportCredentials:
r"""The OAuth client's ID. See this guide for more information."""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""The OAuth client secret. See this guide for more information."""
-
+
@@ -22,5 +22,3 @@ class ZendeskSupportCredentials:
@dataclasses.dataclass
class ZendeskSupport:
credentials: Optional[ZendeskSupportCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/models/shared/zendesk_talk.py b/src/airbyte_api/models/shared/zendesk_talk.py
similarity index 96%
rename from src/airbyte/models/shared/zendesk_talk.py
rename to src/airbyte_api/models/shared/zendesk_talk.py
index 5fc3ba63..29225fb6 100644
--- a/src/airbyte/models/shared/zendesk_talk.py
+++ b/src/airbyte_api/models/shared/zendesk_talk.py
@@ -2,7 +2,7 @@
from __future__ import annotations
import dataclasses
-from airbyte import utils
+from airbyte_api import utils
from dataclasses_json import Undefined, dataclass_json
from typing import Optional
@@ -14,7 +14,7 @@ class ZendeskTalkCredentials:
r"""Client ID"""
client_secret: Optional[str] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('client_secret'), 'exclude': lambda f: f is None }})
r"""Client Secret"""
-
+
@@ -22,5 +22,3 @@ class ZendeskTalkCredentials:
@dataclasses.dataclass
class ZendeskTalk:
credentials: Optional[ZendeskTalkCredentials] = dataclasses.field(default=None, metadata={'dataclasses_json': { 'letter_case': utils.get_field_name('credentials'), 'exclude': lambda f: f is None }})
-
-
diff --git a/src/airbyte/sdk.py b/src/airbyte_api/sdk.py
similarity index 96%
rename from src/airbyte/sdk.py
rename to src/airbyte_api/sdk.py
index 1c5233bc..057b358f 100644
--- a/src/airbyte/sdk.py
+++ b/src/airbyte_api/sdk.py
@@ -8,8 +8,8 @@
from .sources import Sources
from .streams import Streams
from .workspaces import Workspaces
-from airbyte import utils
-from airbyte.models import shared
+from airbyte_api import utils
+from airbyte_api.models import shared
from typing import Callable, Dict, Union
class Airbyte:
@@ -32,7 +32,7 @@ def __init__(self,
retry_config: utils.RetryConfig = None
) -> None:
"""Instantiates the SDK configuring it with the provided parameters.
-
+
:param security: The security details required for authentication
:type security: Union[shared.Security,Callable[[], shared.Security]]
:param server_idx: The index of the server to use for all operations
@@ -48,15 +48,15 @@ def __init__(self,
"""
if client is None:
client = requests_http.Session()
-
+
if server_url is not None:
if url_params is not None:
server_url = utils.template_url(server_url, url_params)
self.sdk_configuration = SDKConfiguration(client, security, server_url, server_idx, retry_config=retry_config)
-
+
self._init_sdks()
-
+
def _init_sdks(self):
self.connections = Connections(self.sdk_configuration)
self.destinations = Destinations(self.sdk_configuration)
@@ -64,4 +64,3 @@ def _init_sdks(self):
self.sources = Sources(self.sdk_configuration)
self.streams = Streams(self.sdk_configuration)
self.workspaces = Workspaces(self.sdk_configuration)
-
\ No newline at end of file
diff --git a/src/airbyte/sdkconfiguration.py b/src/airbyte_api/sdkconfiguration.py
similarity index 96%
rename from src/airbyte/sdkconfiguration.py
rename to src/airbyte_api/sdkconfiguration.py
index 788a3227..07063cfc 100644
--- a/src/airbyte/sdkconfiguration.py
+++ b/src/airbyte_api/sdkconfiguration.py
@@ -4,7 +4,7 @@
import requests as requests_http
from .utils import utils
from .utils.retries import RetryConfig
-from airbyte.models import shared
+from airbyte_api.models import shared
from dataclasses import dataclass
from typing import Callable, Dict, Tuple, Union
diff --git a/src/airbyte/sources.py b/src/airbyte_api/sources.py
similarity index 96%
rename from src/airbyte/sources.py
rename to src/airbyte_api/sources.py
index 1bdb06fd..b055c15c 100644
--- a/src/airbyte/sources.py
+++ b/src/airbyte_api/sources.py
@@ -1,8 +1,8 @@
"""Code generated by Speakeasy (https://speakeasyapi.dev). DO NOT EDIT."""
from .sdkconfiguration import SDKConfiguration
-from airbyte import utils
-from airbyte.models import errors, operations, shared
+from airbyte_api import utils
+from airbyte_api.models import errors, operations, shared
from typing import Optional
class Sources:
@@ -10,15 +10,13 @@ class Sources:
def __init__(self, sdk_config: SDKConfiguration) -> None:
self.sdk_configuration = sdk_config
-
-
-
+
def create_source(self, request: Optional[shared.SourceCreateRequest]) -> operations.CreateSourceResponse:
r"""Create a source
Creates a source given a name, workspace id, and a json blob containing the configuration for the source.
"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/sources'
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, Optional[shared.SourceCreateRequest], "request", False, True, 'json')
@@ -26,17 +24,17 @@ def create_source(self, request: Optional[shared.SourceCreateRequest]) -> operat
headers['content-type'] = req_content_type
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('POST', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.CreateSourceResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.SourceResponse])
@@ -48,27 +46,25 @@ def create_source(self, request: Optional[shared.SourceCreateRequest]) -> operat
return res
-
-
def delete_source(self, request: operations.DeleteSourceRequest) -> operations.DeleteSourceResponse:
r"""Delete a Source"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.DeleteSourceRequest, base_url, '/sources/{sourceId}', request)
headers = {}
headers['Accept'] = '*/*'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('DELETE', url, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.DeleteSourceResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 204:
pass
elif http_res.status_code == 403 or http_res.status_code == 404 or http_res.status_code >= 400 and http_res.status_code < 500 or http_res.status_code >= 500 and http_res.status_code < 600:
@@ -76,27 +72,25 @@ def delete_source(self, request: operations.DeleteSourceRequest) -> operations.D
return res
-
-
def get_source(self, request: operations.GetSourceRequest) -> operations.GetSourceResponse:
r"""Get Source details"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.GetSourceRequest, base_url, '/sources/{sourceId}', request)
headers = {}
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.GetSourceResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.SourceResponse])
@@ -108,8 +102,6 @@ def get_source(self, request: operations.GetSourceRequest) -> operations.GetSour
return res
-
-
def initiate_o_auth(self, request: shared.InitiateOauthRequest) -> operations.InitiateOAuthResponse:
r"""Initiate OAuth for a source
Given a source ID, workspace ID, and redirect URL, initiates OAuth for the source.
@@ -119,7 +111,7 @@ def initiate_o_auth(self, request: shared.InitiateOauthRequest) -> operations.In
That secret ID can be used to create a source with credentials in place of actual tokens.
"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/sources/initiateOAuth'
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, shared.InitiateOauthRequest, "request", False, False, 'json')
@@ -129,17 +121,17 @@ def initiate_o_auth(self, request: shared.InitiateOauthRequest) -> operations.In
raise Exception('request body is required')
headers['Accept'] = '*/*'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('POST', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.InitiateOAuthResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
pass
elif http_res.status_code == 400 or http_res.status_code == 403 or http_res.status_code >= 400 and http_res.status_code < 500 or http_res.status_code >= 500 and http_res.status_code < 600:
@@ -147,28 +139,26 @@ def initiate_o_auth(self, request: shared.InitiateOauthRequest) -> operations.In
return res
-
-
def list_sources(self, request: operations.ListSourcesRequest) -> operations.ListSourcesResponse:
r"""List sources"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/sources'
headers = {}
query_params = utils.get_query_params(operations.ListSourcesRequest, request)
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, params=query_params, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.ListSourcesResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.SourcesResponse])
@@ -180,12 +170,10 @@ def list_sources(self, request: operations.ListSourcesRequest) -> operations.Lis
return res
-
-
def patch_source(self, request: operations.PatchSourceRequest) -> operations.PatchSourceResponse:
r"""Update a Source"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.PatchSourceRequest, base_url, '/sources/{sourceId}', request)
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, operations.PatchSourceRequest, "source_patch_request", False, True, 'json')
@@ -193,17 +181,17 @@ def patch_source(self, request: operations.PatchSourceRequest) -> operations.Pat
headers['content-type'] = req_content_type
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('PATCH', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.PatchSourceResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.SourceResponse])
@@ -215,12 +203,10 @@ def patch_source(self, request: operations.PatchSourceRequest) -> operations.Pat
return res
-
-
def put_source(self, request: operations.PutSourceRequest) -> operations.PutSourceResponse:
r"""Update a Source and fully overwrite it"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.PutSourceRequest, base_url, '/sources/{sourceId}', request)
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, operations.PutSourceRequest, "source_put_request", False, True, 'json')
@@ -228,17 +214,17 @@ def put_source(self, request: operations.PutSourceRequest) -> operations.PutSour
headers['content-type'] = req_content_type
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('PUT', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.PutSourceResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.SourceResponse])
@@ -249,5 +235,3 @@ def put_source(self, request: operations.PutSourceRequest) -> operations.PutSour
raise errors.SDKError('API error occurred', http_res.status_code, http_res.text, http_res)
return res
-
-
\ No newline at end of file
diff --git a/src/airbyte/streams.py b/src/airbyte_api/streams.py
similarity index 93%
rename from src/airbyte/streams.py
rename to src/airbyte_api/streams.py
index 081e2a88..0a9006c3 100644
--- a/src/airbyte/streams.py
+++ b/src/airbyte_api/streams.py
@@ -1,8 +1,8 @@
"""Code generated by Speakeasy (https://speakeasyapi.dev). DO NOT EDIT."""
from .sdkconfiguration import SDKConfiguration
-from airbyte import utils
-from airbyte.models import errors, operations, shared
+from airbyte_api import utils
+from airbyte_api.models import errors, operations, shared
from typing import Optional
class Streams:
@@ -10,29 +10,27 @@ class Streams:
def __init__(self, sdk_config: SDKConfiguration) -> None:
self.sdk_configuration = sdk_config
-
-
-
+
def get_stream_properties(self, request: operations.GetStreamPropertiesRequest) -> operations.GetStreamPropertiesResponse:
r"""Get stream properties"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/streams'
headers = {}
query_params = utils.get_query_params(operations.GetStreamPropertiesRequest, request)
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, params=query_params, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.GetStreamPropertiesResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.StreamPropertiesResponse])
@@ -43,5 +41,3 @@ def get_stream_properties(self, request: operations.GetStreamPropertiesRequest)
raise errors.SDKError('API error occurred', http_res.status_code, http_res.text, http_res)
return res
-
-
\ No newline at end of file
diff --git a/src/airbyte/utils/__init__.py b/src/airbyte_api/utils/__init__.py
similarity index 100%
rename from src/airbyte/utils/__init__.py
rename to src/airbyte_api/utils/__init__.py
diff --git a/src/airbyte/utils/retries.py b/src/airbyte_api/utils/retries.py
similarity index 100%
rename from src/airbyte/utils/retries.py
rename to src/airbyte_api/utils/retries.py
diff --git a/src/airbyte/utils/utils.py b/src/airbyte_api/utils/utils.py
similarity index 100%
rename from src/airbyte/utils/utils.py
rename to src/airbyte_api/utils/utils.py
diff --git a/src/airbyte/workspaces.py b/src/airbyte_api/workspaces.py
similarity index 96%
rename from src/airbyte/workspaces.py
rename to src/airbyte_api/workspaces.py
index 222882e7..9344c869 100644
--- a/src/airbyte/workspaces.py
+++ b/src/airbyte_api/workspaces.py
@@ -1,8 +1,8 @@
"""Code generated by Speakeasy (https://speakeasyapi.dev). DO NOT EDIT."""
from .sdkconfiguration import SDKConfiguration
-from airbyte import utils
-from airbyte.models import errors, operations, shared
+from airbyte_api import utils
+from airbyte_api.models import errors, operations, shared
from typing import Optional
class Workspaces:
@@ -10,16 +10,14 @@ class Workspaces:
def __init__(self, sdk_config: SDKConfiguration) -> None:
self.sdk_configuration = sdk_config
-
-
-
+
def create_or_update_workspace_o_auth_credentials(self, request: operations.CreateOrUpdateWorkspaceOAuthCredentialsRequest) -> operations.CreateOrUpdateWorkspaceOAuthCredentialsResponse:
r"""Create OAuth override credentials for a workspace and source type.
Create/update a set of OAuth credentials to override the Airbyte-provided OAuth credentials used for source/destination OAuth.
In order to determine what the credential configuration needs to be, please see the connector specification of the relevant source/destination.
"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.CreateOrUpdateWorkspaceOAuthCredentialsRequest, base_url, '/workspaces/{workspaceId}/oauthCredentials', request)
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, operations.CreateOrUpdateWorkspaceOAuthCredentialsRequest, "workspace_o_auth_credentials_request", False, False, 'json')
@@ -29,17 +27,17 @@ def create_or_update_workspace_o_auth_credentials(self, request: operations.Crea
raise Exception('request body is required')
headers['Accept'] = '*/*'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('PUT', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.CreateOrUpdateWorkspaceOAuthCredentialsResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
pass
elif http_res.status_code == 400 or http_res.status_code == 403 or http_res.status_code >= 400 and http_res.status_code < 500 or http_res.status_code >= 500 and http_res.status_code < 600:
@@ -47,12 +45,10 @@ def create_or_update_workspace_o_auth_credentials(self, request: operations.Crea
return res
-
-
def create_workspace(self, request: shared.WorkspaceCreateRequest) -> operations.CreateWorkspaceResponse:
r"""Create a workspace"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/workspaces'
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, shared.WorkspaceCreateRequest, "request", False, False, 'json')
@@ -62,17 +58,17 @@ def create_workspace(self, request: shared.WorkspaceCreateRequest) -> operations
raise Exception('request body is required')
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('POST', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.CreateWorkspaceResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.WorkspaceResponse])
@@ -84,27 +80,25 @@ def create_workspace(self, request: shared.WorkspaceCreateRequest) -> operations
return res
-
-
def delete_workspace(self, request: operations.DeleteWorkspaceRequest) -> operations.DeleteWorkspaceResponse:
r"""Delete a Workspace"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.DeleteWorkspaceRequest, base_url, '/workspaces/{workspaceId}', request)
headers = {}
headers['Accept'] = '*/*'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('DELETE', url, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.DeleteWorkspaceResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 204:
pass
elif http_res.status_code == 403 or http_res.status_code == 404 or http_res.status_code >= 400 and http_res.status_code < 500 or http_res.status_code >= 500 and http_res.status_code < 600:
@@ -112,27 +106,25 @@ def delete_workspace(self, request: operations.DeleteWorkspaceRequest) -> operat
return res
-
-
def get_workspace(self, request: operations.GetWorkspaceRequest) -> operations.GetWorkspaceResponse:
r"""Get Workspace details"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.GetWorkspaceRequest, base_url, '/workspaces/{workspaceId}', request)
headers = {}
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.GetWorkspaceResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.WorkspaceResponse])
@@ -144,28 +136,26 @@ def get_workspace(self, request: operations.GetWorkspaceRequest) -> operations.G
return res
-
-
def list_workspaces(self, request: operations.ListWorkspacesRequest) -> operations.ListWorkspacesResponse:
r"""List workspaces"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = base_url + '/workspaces'
headers = {}
query_params = utils.get_query_params(operations.ListWorkspacesRequest, request)
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('GET', url, params=query_params, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.ListWorkspacesResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.WorkspacesResponse])
@@ -177,12 +167,10 @@ def list_workspaces(self, request: operations.ListWorkspacesRequest) -> operatio
return res
-
-
def update_workspace(self, request: operations.UpdateWorkspaceRequest) -> operations.UpdateWorkspaceResponse:
r"""Update a workspace"""
base_url = utils.template_url(*self.sdk_configuration.get_server_details())
-
+
url = utils.generate_url(operations.UpdateWorkspaceRequest, base_url, '/workspaces/{workspaceId}', request)
headers = {}
req_content_type, data, form = utils.serialize_request_body(request, operations.UpdateWorkspaceRequest, "workspace_update_request", False, False, 'json')
@@ -192,17 +180,17 @@ def update_workspace(self, request: operations.UpdateWorkspaceRequest) -> operat
raise Exception('request body is required')
headers['Accept'] = 'application/json'
headers['user-agent'] = self.sdk_configuration.user_agent
-
+
if callable(self.sdk_configuration.security):
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security())
else:
client = utils.configure_security_client(self.sdk_configuration.client, self.sdk_configuration.security)
-
+
http_res = client.request('PATCH', url, data=data, files=form, headers=headers)
content_type = http_res.headers.get('Content-Type')
-
+
res = operations.UpdateWorkspaceResponse(status_code=http_res.status_code, content_type=content_type, raw_response=http_res)
-
+
if http_res.status_code == 200:
if utils.match_content_type(content_type, 'application/json'):
out = utils.unmarshal_json(http_res.text, Optional[shared.WorkspaceResponse])
@@ -213,5 +201,3 @@ def update_workspace(self, request: operations.UpdateWorkspaceRequest) -> operat
raise errors.SDKError('API error occurred', http_res.status_code, http_res.text, http_res)
return res
-
-
\ No newline at end of file