2023-01-10 09:39:14 INFO i.a.v.j.JsonSchemaValidator(test):130 - JSON schema validation failed. errors: $.s3_bucket_name: is missing but it is required, $.s3_bucket_path: is missing but it is required, $.s3_bucket_region: is missing but it is required, $.s3_access_key_id: is missing but it is required, $.s3_secret_access_key: is missing but it is required, $.data_source_type: does not have a value in the enumeration [S3] 2023-01-10 09:39:14 INFO i.a.v.j.JsonSchemaValidator(test):130 - JSON schema validation failed. errors: $.azure_blob_storage_sas_token: object found, string expected 2023-01-10 09:39:14 INFO i.a.w.t.TemporalAttemptExecution(get):136 - Docker volume job log path: /tmp/workspace/15/0/logs.log 2023-01-10 09:39:14 INFO i.a.w.t.TemporalAttemptExecution(get):141 - Executing worker wrapper. Airbyte version: 0.40.26 2023-01-10 09:39:14 INFO i.a.a.c.AirbyteApiClient(retryWithJitter):172 - Attempt 0 to save workflow id for cancellation 2023-01-10 09:39:15 INFO i.a.c.i.LineGobbler(voidCall):114 - 2023-01-10 09:39:15 INFO i.a.c.i.LineGobbler(voidCall):114 - ----- START CHECK ----- 2023-01-10 09:39:15 INFO i.a.c.i.LineGobbler(voidCall):114 - 2023-01-10 09:39:15 INFO i.a.c.i.LineGobbler(voidCall):114 - Checking if airbyte/source-s3:0.1.27 exists... 2023-01-10 09:39:15 INFO i.a.c.i.LineGobbler(voidCall):114 - airbyte/source-s3:0.1.27 was found locally. 2023-01-10 09:39:15 INFO i.a.w.p.DockerProcessFactory(create):120 - Creating docker container = source-s3-check-15-0-hdlzw with resources io.airbyte.config.ResourceRequirements@781d6b2f[cpuRequest=,cpuLimit=,memoryRequest=,memoryLimit=] 2023-01-10 09:39:15 INFO i.a.w.p.DockerProcessFactory(create):164 - Preparing command: docker run --rm --init -i -w /data/15/0 --log-driver none --name source-s3-check-15-0-hdlzw --network host -v airbyte_workspace:/data -v /tmp/airbyte_local:/local -e DEPLOYMENT_MODE=OSS -e USE_STREAM_CAPABLE_STATE=true -e WORKER_ENVIRONMENT=DOCKER -e AIRBYTE_ROLE= -e WORKER_CONNECTOR_IMAGE=airbyte/source-s3:0.1.27 -e WORKER_JOB_ATTEMPT=0 -e AUTO_DETECT_SCHEMA=false -e AIRBYTE_VERSION=0.40.26 -e WORKER_JOB_ID=15 airbyte/source-s3:0.1.27 check --config source_config.json 2023-01-10 09:39:15 INFO i.a.w.i.VersionedAirbyteStreamFactory(create):97 - Reading messages from protocol version 0.2.0 2023-01-10 09:39:16 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - initialised stream with format: 2023-01-10 09:39:16 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - initialised stream with format: {'encoding': 'utf8', 'filetype': 'csv', 'delimiter': ',', 'block_size': 1000000, 'quote_char': '"', 'double_quote': False, 'infer_datatypes': False, 'advanced_options': '{}', 'newlines_in_values': False, 'additional_reader_options': '{}'} 2023-01-10 09:39:16 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Iterating S3 bucket 'datamesh-datalake-demo' with prefix: 'small/' 2023-01-10 09:39:16 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Check succeeded 2023-01-10 09:39:17 INFO i.a.w.t.TemporalAttemptExecution(get):163 - Stopping cancellation check scheduling... 2023-01-10 09:39:17 INFO i.a.c.i.LineGobbler(voidCall):114 - 2023-01-10 09:39:17 INFO i.a.c.i.LineGobbler(voidCall):114 - ----- END CHECK ----- 2023-01-10 09:39:17 INFO i.a.c.i.LineGobbler(voidCall):114 - 2023-01-10 09:39:17 INFO i.a.w.t.TemporalAttemptExecution(get):136 - Docker volume job log path: /tmp/workspace/15/0/logs.log 2023-01-10 09:39:17 INFO i.a.w.t.TemporalAttemptExecution(get):141 - Executing worker wrapper. Airbyte version: 0.40.26 2023-01-10 09:39:17 INFO i.a.a.c.AirbyteApiClient(retryWithJitter):172 - Attempt 0 to save workflow id for cancellation 2023-01-10 09:39:17 INFO i.a.c.i.LineGobbler(voidCall):114 - 2023-01-10 09:39:17 INFO i.a.c.i.LineGobbler(voidCall):114 - ----- START CHECK ----- 2023-01-10 09:39:17 INFO i.a.c.i.LineGobbler(voidCall):114 - 2023-01-10 09:39:17 INFO i.a.c.i.LineGobbler(voidCall):114 - Checking if airbyte/destination-databricks:0.3.1 exists... 2023-01-10 09:39:17 INFO i.a.c.i.LineGobbler(voidCall):114 - airbyte/destination-databricks:0.3.1 was found locally. 2023-01-10 09:39:17 INFO i.a.w.p.DockerProcessFactory(create):120 - Creating docker container = destination-databricks-check-15-0-eblaz with resources io.airbyte.config.ResourceRequirements@781d6b2f[cpuRequest=,cpuLimit=,memoryRequest=,memoryLimit=] 2023-01-10 09:39:17 INFO i.a.w.p.DockerProcessFactory(create):164 - Preparing command: docker run --rm --init -i -w /data/15/0 --log-driver none --name destination-databricks-check-15-0-eblaz --network host -v airbyte_workspace:/data -v /tmp/airbyte_local:/local -e DEPLOYMENT_MODE=OSS -e USE_STREAM_CAPABLE_STATE=true -e WORKER_ENVIRONMENT=DOCKER -e AIRBYTE_ROLE= -e WORKER_CONNECTOR_IMAGE=airbyte/destination-databricks:0.3.1 -e WORKER_JOB_ATTEMPT=0 -e AUTO_DETECT_SCHEMA=false -e AIRBYTE_VERSION=0.40.26 -e WORKER_JOB_ID=15 airbyte/destination-databricks:0.3.1 check --config source_config.json 2023-01-10 09:39:17 INFO i.a.w.i.VersionedAirbyteStreamFactory(create):97 - Reading messages from protocol version 0.2.0 2023-01-10 09:39:18 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - integration args: {check=null, config=source_config.json} 2023-01-10 09:39:18 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Running integration: io.airbyte.integrations.destination.databricks.DatabricksDestination 2023-01-10 09:39:18 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Command: CHECK 2023-01-10 09:39:18 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Integration config: IntegrationConfig{command=CHECK, configPath='source_config.json', catalogPath='null', statePath='null'} 2023-01-10 09:39:18 WARN i.a.w.i.DefaultAirbyteStreamFactory(internalLog):117 - Unknown keyword order - you should define your own Meta Schema. If the keyword is irrelevant for validation, just use a NonValidationKeyword 2023-01-10 09:39:18 WARN i.a.w.i.DefaultAirbyteStreamFactory(internalLog):117 - Unknown keyword airbyte_secret - you should define your own Meta Schema. If the keyword is irrelevant for validation, just use a NonValidationKeyword 2023-01-10 09:39:18 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Using destination type: AZURE_BLOB_STORAGE 2023-01-10 09:39:18 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Databricks storage type config: "Azure_Blob_Storage" 2023-01-10 09:39:18 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Package versions: jackson-annotations=2.13.3, jackson-core=2.13.3, jackson-databind=2.13.3, jackson-dataformat-xml=2.13.3, jackson-datatype-jsr310=2.13.3, azure-core=1.30.0, Troubleshooting version conflicts: https://aka.ms/azsdk/java/dependency/troubleshoot 2023-01-10 09:39:20 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - blobContainerClient created 2023-01-10 09:39:20 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Writing test data to Azure Blob storage: Some test data 2023-01-10 09:39:20 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - blobCommittedBlockCount: 1 2023-01-10 09:39:22 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Blob name: testConnectionBlob05981f48-0fb7-495f-9ebc-a253fd7dc2ecSnapshot: null 2023-01-10 09:39:22 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Deleting blob: testConnectionBlob05981f48-0fb7-495f-9ebc-a253fd7dc2ec 2023-01-10 09:39:22 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Databricks storage type config: "Azure_Blob_Storage" 2023-01-10 09:39:22 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - HikariPool-1 - Starting... 2023-01-10 09:39:22 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - HikariPool-1 - Start completed. 2023-01-10 09:39:23 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - HikariPool-1 - Driver does not support get/set network timeout for connections. ([Databricks][JDBC](10220) Driver does not support this optional feature.) 2023-01-10 09:39:26 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - HikariPool-1 - Shutdown initiated... 2023-01-10 09:39:26 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - HikariPool-1 - Shutdown completed. 2023-01-10 09:39:26 INFO i.a.w.i.DefaultAirbyteStreamFactory(internalLog):120 - Completed integration: io.airbyte.integrations.destination.databricks.DatabricksDestination 2023-01-10 09:39:27 INFO i.a.w.t.TemporalAttemptExecution(get):163 - Stopping cancellation check scheduling... 2023-01-10 09:39:27 INFO i.a.c.i.LineGobbler(voidCall):114 - 2023-01-10 09:39:27 INFO i.a.c.i.LineGobbler(voidCall):114 - ----- END CHECK ----- 2023-01-10 09:39:27 INFO i.a.c.i.LineGobbler(voidCall):114 - 2023-01-10 09:39:27 INFO i.a.w.t.TemporalAttemptExecution(get):136 - Docker volume job log path: /tmp/workspace/15/0/logs.log 2023-01-10 09:39:27 INFO i.a.w.t.TemporalAttemptExecution(get):141 - Executing worker wrapper. Airbyte version: 0.40.26 2023-01-10 09:39:27 INFO i.a.a.c.AirbyteApiClient(retryWithJitter):172 - Attempt 0 to save workflow id for cancellation 2023-01-10 09:39:27 INFO i.a.c.f.EnvVariableFeatureFlags(getEnvOrDefault):64 - Using default value for environment variable LOG_CONNECTOR_MESSAGES: 'false' 2023-01-10 09:39:27 INFO i.a.c.EnvConfigs(getEnvOrDefault):1161 - Using default value for environment variable METRIC_CLIENT: '' 2023-01-10 09:39:27 WARN i.a.m.l.MetricClientFactory(initialize):60 - Metric client is already initialized to 2023-01-10 09:39:27 INFO i.a.c.f.EnvVariableFeatureFlags(getEnvOrDefault):64 - Using default value for environment variable LOG_CONNECTOR_MESSAGES: 'false' 2023-01-10 09:39:27 INFO i.a.w.g.DefaultReplicationWorker(run):142 - start sync worker. job id: 15 attempt id: 0 2023-01-10 09:39:27 INFO i.a.w.g.DefaultReplicationWorker(run):157 - configured sync modes: {null.output=full_refresh - overwrite} 2023-01-10 09:39:27 INFO i.a.w.i.DefaultAirbyteDestination(start):72 - Running destination... 2023-01-10 09:39:27 INFO i.a.c.i.LineGobbler(voidCall):114 - 2023-01-10 09:39:27 INFO i.a.c.i.LineGobbler(voidCall):114 - ----- START REPLICATION ----- 2023-01-10 09:39:27 INFO i.a.c.i.LineGobbler(voidCall):114 - 2023-01-10 09:39:27 INFO i.a.c.i.LineGobbler(voidCall):114 - Checking if airbyte/destination-databricks:0.3.1 exists... 2023-01-10 09:39:27 INFO i.a.w.p.DockerProcessFactory(create):120 - Creating docker container = destination-databricks-write-15-0-fxyqt with resources io.airbyte.config.ResourceRequirements@45ed0eb7[cpuRequest=,cpuLimit=,memoryRequest=,memoryLimit=] 2023-01-10 09:39:27 INFO i.a.c.i.LineGobbler(voidCall):114 - airbyte/destination-databricks:0.3.1 was found locally. 2023-01-10 09:39:27 INFO i.a.w.p.DockerProcessFactory(create):164 - Preparing command: docker run --rm --init -i -w /data/15/0 --log-driver none --name destination-databricks-write-15-0-fxyqt --network host -v airbyte_workspace:/data -v /tmp/airbyte_local:/local -e DEPLOYMENT_MODE=OSS -e USE_STREAM_CAPABLE_STATE=true -e WORKER_ENVIRONMENT=DOCKER -e AIRBYTE_ROLE= -e WORKER_CONNECTOR_IMAGE=airbyte/destination-databricks:0.3.1 -e WORKER_JOB_ATTEMPT=0 -e AUTO_DETECT_SCHEMA=false -e AIRBYTE_VERSION=0.40.26 -e WORKER_JOB_ID=15 airbyte/destination-databricks:0.3.1 write --config destination_config.json --catalog destination_catalog.json 2023-01-10 09:39:27 INFO i.a.w.i.VersionedAirbyteMessageBufferedWriterFactory(createWriter):33 - Writing messages to protocol version 0.2.0 2023-01-10 09:39:27 INFO i.a.w.i.VersionedAirbyteStreamFactory(create):97 - Reading messages from protocol version 0.2.0 2023-01-10 09:39:27 INFO i.a.c.i.LineGobbler(voidCall):114 - Checking if airbyte/source-s3:0.1.27 exists... 2023-01-10 09:39:27 INFO i.a.c.i.LineGobbler(voidCall):114 - airbyte/source-s3:0.1.27 was found locally. 2023-01-10 09:39:27 INFO i.a.w.p.DockerProcessFactory(create):120 - Creating docker container = source-s3-read-15-0-fqwkx with resources io.airbyte.config.ResourceRequirements@37b5ab09[cpuRequest=,cpuLimit=,memoryRequest=,memoryLimit=] 2023-01-10 09:39:27 INFO i.a.w.p.DockerProcessFactory(create):164 - Preparing command: docker run --rm --init -i -w /data/15/0 --log-driver none --name source-s3-read-15-0-fqwkx --network host -v airbyte_workspace:/data -v /tmp/airbyte_local:/local -e DEPLOYMENT_MODE=OSS -e USE_STREAM_CAPABLE_STATE=true -e WORKER_ENVIRONMENT=DOCKER -e AIRBYTE_ROLE= -e WORKER_CONNECTOR_IMAGE=airbyte/source-s3:0.1.27 -e WORKER_JOB_ATTEMPT=0 -e AUTO_DETECT_SCHEMA=false -e AIRBYTE_VERSION=0.40.26 -e WORKER_JOB_ID=15 airbyte/source-s3:0.1.27 read --config source_config.json --catalog source_catalog.json 2023-01-10 09:39:27 INFO i.a.w.i.VersionedAirbyteStreamFactory(create):97 - Reading messages from protocol version 0.2.0 2023-01-10 09:39:27 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromDstRunnable$4):260 - Destination output thread started. 2023-01-10 09:39:27 INFO i.a.w.g.DefaultReplicationWorker(replicate):234 - Waiting for source and destination threads to complete. 2023-01-10 09:39:27 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):310 - Replication thread started. 2023-01-10 09:39:29 source > initialised stream with format: 2023-01-10 09:39:29 source > Starting syncing SourceS3 2023-01-10 09:39:29 source > initialised stream with format: {'encoding': 'utf8', 'filetype': 'csv', 'delimiter': ',', 'block_size': 1000000, 'quote_char': '"', 'double_quote': False, 'infer_datatypes': False, 'advanced_options': '{}', 'newlines_in_values': False, 'additional_reader_options': '{}'} 2023-01-10 09:39:29 source > Syncing stream: output 2023-01-10 09:39:29 source > Iterating S3 bucket 'datamesh-datalake-demo' with prefix: 'small/' 2023-01-10 09:39:29 destination > integration args: {catalog=destination_catalog.json, write=null, config=destination_config.json} 2023-01-10 09:39:29 destination > Running integration: io.airbyte.integrations.destination.databricks.DatabricksDestination 2023-01-10 09:39:29 destination > Command: WRITE 2023-01-10 09:39:30 destination > Integration config: IntegrationConfig{command=WRITE, configPath='destination_config.json', catalogPath='destination_catalog.json', statePath='null'} 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'Cancelled', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'boolean', but found 'string'. Airbyte will attempt to coerce this to boolean on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'Diverted', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'boolean', but found 'string'. Airbyte will attempt to coerce this to boolean on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'CRSDepTime', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DepTime', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DepDelayMinutes', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DepDelay', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'ArrTime', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'ArrDelayMinutes', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'AirTime', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'CRSElapsedTime', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'ActualElapsedTime', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'Distance', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'Year', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'Quarter', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'Month', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DayofMonth', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DayOfWeek', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DOT_ID_Marketing_Airline', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'Flight_Number_Marketing_Airline', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DOT_ID_Operating_Airline', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'Flight_Number_Operating_Airline', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'OriginAirportID', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'OriginAirportSeqID', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'OriginCityMarketID', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'OriginStateFips', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 destination > Unknown keyword order - you should define your own Meta Schema. If the keyword is irrelevant for validation, just use a NonValidationKeyword 2023-01-10 09:39:30 destination > Unknown keyword airbyte_secret - you should define your own Meta Schema. If the keyword is irrelevant for validation, just use a NonValidationKeyword 2023-01-10 09:39:30 destination > Using destination type: AZURE_BLOB_STORAGE 2023-01-10 09:39:30 destination > Databricks storage type config: "Azure_Blob_Storage" 2023-01-10 09:39:30 destination > Databricks storage type config: "Azure_Blob_Storage" 2023-01-10 09:39:30 destination > HikariPool-1 - Starting... 2023-01-10 09:39:30 destination > HikariPool-1 - Start completed. 2023-01-10 09:39:30 destination > Package versions: jackson-annotations=2.13.3, jackson-core=2.13.3, jackson-databind=2.13.3, jackson-dataformat-xml=2.13.3, jackson-datatype-jsr310=2.13.3, azure-core=1.30.0, Troubleshooting version conflicts: https://aka.ms/azsdk/java/dependency/troubleshoot 2023-01-10 09:39:30 destination > [Stream output] Database schema: public 2023-01-10 09:39:30 destination > [Stream output] Tmp table _airbyte_tmp_zpw_output location: abfss://dbtsparkstoragecontainer@dbtsparkstorageaccount.blob.core.windows.net/71270d3d-06af-4f7d-bb70-8dad60c0660d/public/_airbyte_tmp_zpw_output/ 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'OriginWac', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DestAirportID', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DestAirportSeqID', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DestCityMarketID', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DestStateFips', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DestWac', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DepDel15', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DepartureDelayGroups', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'TaxiOut', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'WheelsOff', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'WheelsOn', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'TaxiIn', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'CRSArrTime', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'ArrDelay', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'ArrDel15', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'ArrivalDelayGroups', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DistanceGroup', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column 'DivAirportLandings', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'number', but found 'string'. Airbyte will attempt to coerce this to number on read. 2023-01-10 09:39:30 source > Detected mismatched datatype on column '__index_level_0__', in file 'Key: small/flights_small.csv, LastModified: 2023-01-10T08:03:55+00:00, Size: 71.3623Mb'. Should be 'integer', but found 'string'. Airbyte will attempt to coerce this to integer on read. 2023-01-10 09:39:30 source > determined master schema: {'FlightDate': 'string', 'Airline': 'string', 'Origin': 'string', 'Dest': 'string', 'Cancelled': 'boolean', 'Diverted': 'boolean', 'CRSDepTime': 'integer', 'DepTime': 'number', 'DepDelayMinutes': 'number', 'DepDelay': 'number', 'ArrTime': 'number', 'ArrDelayMinutes': 'number', 'AirTime': 'number', 'CRSElapsedTime': 'number', 'ActualElapsedTime': 'number', 'Distance': 'number', 'Year': 'integer', 'Quarter': 'integer', 'Month': 'integer', 'DayofMonth': 'integer', 'DayOfWeek': 'integer', 'Marketing_Airline_Network': 'string', 'Operated_or_Branded_Code_Share_Partners': 'string', 'DOT_ID_Marketing_Airline': 'integer', 'IATA_Code_Marketing_Airline': 'string', 'Flight_Number_Marketing_Airline': 'integer', 'Operating_Airline': 'string', 'DOT_ID_Operating_Airline': 'integer', 'IATA_Code_Operating_Airline': 'string', 'Tail_Number': 'string', 'Flight_Number_Operating_Airline': 'integer', 'OriginAirportID': 'integer', 'OriginAirportSeqID': 'integer', 'OriginCityMarketID': 'integer', 'OriginCityName': 'string', 'OriginState': 'string', 'OriginStateFips': 'integer', 'OriginStateName': 'string', 'OriginWac': 'integer', 'DestAirportID': 'integer', 'DestAirportSeqID': 'integer', 'DestCityMarketID': 'integer', 'DestCityName': 'string', 'DestState': 'string', 'DestStateFips': 'integer', 'DestStateName': 'string', 'DestWac': 'integer', 'DepDel15': 'number', 'DepartureDelayGroups': 'number', 'DepTimeBlk': 'string', 'TaxiOut': 'number', 'WheelsOff': 'number', 'WheelsOn': 'number', 'TaxiIn': 'number', 'CRSArrTime': 'integer', 'ArrDelay': 'number', 'ArrDel15': 'number', 'ArrivalDelayGroups': 'number', 'ArrTimeBlk': 'string', 'DistanceGroup': 'integer', 'DivAirportLandings': 'number', '__index_level_0__': 'integer'} 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 destination > [Stream output] Data table output location: abfss://dbtsparkstoragecontainer@dbtsparkstorageaccount.dfs.core.windows.net/public/output/ 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 ERROR c.n.s.DateTimeValidator(tryParse):82 - Invalid date-time: Invalid timezone offset: +0000 2023-01-10 09:39:30 destination > class io.airbyte.integrations.destination.buffered_stream_consumer.BufferedStreamConsumer started. 2023-01-10 09:39:31 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 1000 (1 MB) 2023-01-10 09:39:32 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 2000 (2 MB) 2023-01-10 09:39:33 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 3000 (4 MB) 2023-01-10 09:39:33 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 4000 (5 MB) 2023-01-10 09:39:33 destination > Flushing output: 4190 records (24 MB) 2023-01-10 09:39:36 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 5000 (7 MB) 2023-01-10 09:39:36 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 6000 (8 MB) 2023-01-10 09:39:37 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 7000 (10 MB) 2023-01-10 09:39:38 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 8000 (11 MB) 2023-01-10 09:39:38 destination > Flushing output: 4208 records (24 MB) 2023-01-10 09:39:38 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 9000 (13 MB) 2023-01-10 09:39:39 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 10000 (14 MB) 2023-01-10 09:39:40 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 11000 (16 MB) 2023-01-10 09:39:41 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 12000 (17 MB) 2023-01-10 09:39:41 destination > Flushing output: 4203 records (24 MB) 2023-01-10 09:39:42 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 13000 (19 MB) 2023-01-10 09:39:42 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 14000 (20 MB) 2023-01-10 09:39:43 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 15000 (22 MB) 2023-01-10 09:39:44 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 16000 (23 MB) 2023-01-10 09:39:44 destination > Flushing output: 4172 records (24 MB) 2023-01-10 09:39:44 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 17000 (25 MB) 2023-01-10 09:39:45 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 18000 (26 MB) 2023-01-10 09:39:46 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 19000 (28 MB) 2023-01-10 09:39:46 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 20000 (29 MB) 2023-01-10 09:39:47 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 21000 (31 MB) 2023-01-10 09:39:47 destination > Flushing output: 4227 records (24 MB) 2023-01-10 09:39:47 destination > Begin closing non current file writers 2023-01-10 09:39:47 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:39:47 destination > Closing csvPrinter when succeed 2023-01-10 09:39:48 destination > Upload completed for stream 'output'. 2023-01-10 09:39:48 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 22000 (32 MB) 2023-01-10 09:39:49 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 23000 (34 MB) 2023-01-10 09:39:49 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 24000 (35 MB) 2023-01-10 09:39:50 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 25000 (37 MB) 2023-01-10 09:39:50 destination > Flushing output: 4219 records (24 MB) 2023-01-10 09:39:51 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 26000 (38 MB) 2023-01-10 09:39:51 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 27000 (40 MB) 2023-01-10 09:39:52 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 28000 (41 MB) 2023-01-10 09:39:53 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 29000 (43 MB) 2023-01-10 09:39:53 destination > Flushing output: 4182 records (24 MB) 2023-01-10 09:39:54 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 30000 (44 MB) 2023-01-10 09:39:54 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 31000 (46 MB) 2023-01-10 09:39:55 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 32000 (47 MB) 2023-01-10 09:39:55 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 33000 (49 MB) 2023-01-10 09:39:56 destination > Flushing output: 4213 records (24 MB) 2023-01-10 09:39:56 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 34000 (50 MB) 2023-01-10 09:39:57 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 35000 (52 MB) 2023-01-10 09:39:57 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 36000 (53 MB) 2023-01-10 09:39:58 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 37000 (55 MB) 2023-01-10 09:39:59 destination > Flushing output: 4229 records (24 MB) 2023-01-10 09:39:59 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 38000 (56 MB) 2023-01-10 09:39:59 destination > Begin closing non current file writers 2023-01-10 09:39:59 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:39:59 destination > Closing csvPrinter when succeed 2023-01-10 09:40:00 destination > Upload completed for stream 'output'. 2023-01-10 09:40:00 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 39000 (57 MB) 2023-01-10 09:40:00 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 40000 (59 MB) 2023-01-10 09:40:01 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 41000 (60 MB) 2023-01-10 09:40:02 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 42000 (62 MB) 2023-01-10 09:40:02 destination > Flushing output: 4214 records (24 MB) 2023-01-10 09:40:02 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 43000 (63 MB) 2023-01-10 09:40:03 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 44000 (65 MB) 2023-01-10 09:40:04 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 45000 (66 MB) 2023-01-10 09:40:05 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 46000 (68 MB) 2023-01-10 09:40:05 destination > Flushing output: 4178 records (24 MB) 2023-01-10 09:40:05 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 47000 (69 MB) 2023-01-10 09:40:06 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 48000 (71 MB) 2023-01-10 09:40:07 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 49000 (72 MB) 2023-01-10 09:40:07 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 50000 (74 MB) 2023-01-10 09:40:07 destination > Flushing output: 4175 records (24 MB) 2023-01-10 09:40:08 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 51000 (75 MB) 2023-01-10 09:40:09 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 52000 (77 MB) 2023-01-10 09:40:09 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 53000 (78 MB) 2023-01-10 09:40:10 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 54000 (80 MB) 2023-01-10 09:40:10 destination > Flushing output: 4177 records (24 MB) 2023-01-10 09:40:11 destination > Begin closing non current file writers 2023-01-10 09:40:11 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:40:11 destination > Closing csvPrinter when succeed 2023-01-10 09:40:11 destination > Upload completed for stream 'output'. 2023-01-10 09:40:11 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 55000 (81 MB) 2023-01-10 09:40:11 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 56000 (83 MB) 2023-01-10 09:40:12 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 57000 (84 MB) 2023-01-10 09:40:13 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 58000 (86 MB) 2023-01-10 09:40:13 destination > Flushing output: 4202 records (24 MB) 2023-01-10 09:40:13 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 59000 (87 MB) 2023-01-10 09:40:14 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 60000 (89 MB) 2023-01-10 09:40:15 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 61000 (90 MB) 2023-01-10 09:40:15 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 62000 (92 MB) 2023-01-10 09:40:16 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 63000 (93 MB) 2023-01-10 09:40:16 destination > Flushing output: 4212 records (24 MB) 2023-01-10 09:40:17 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 64000 (95 MB) 2023-01-10 09:40:17 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 65000 (96 MB) 2023-01-10 09:40:18 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 66000 (98 MB) 2023-01-10 09:40:19 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 67000 (99 MB) 2023-01-10 09:40:19 destination > Flushing output: 4216 records (24 MB) 2023-01-10 09:40:19 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 68000 (101 MB) 2023-01-10 09:40:20 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 69000 (102 MB) 2023-01-10 09:40:20 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 70000 (104 MB) 2023-01-10 09:40:21 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 71000 (105 MB) 2023-01-10 09:40:21 destination > Flushing output: 4211 records (24 MB) 2023-01-10 09:40:22 destination > Begin closing non current file writers 2023-01-10 09:40:22 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:40:22 destination > Closing csvPrinter when succeed 2023-01-10 09:40:22 destination > Upload completed for stream 'output'. 2023-01-10 09:40:22 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 72000 (107 MB) 2023-01-10 09:40:22 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 73000 (108 MB) 2023-01-10 09:40:23 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 74000 (110 MB) 2023-01-10 09:40:24 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 75000 (111 MB) 2023-01-10 09:40:24 destination > Flushing output: 4212 records (24 MB) 2023-01-10 09:40:24 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 76000 (113 MB) 2023-01-10 09:40:25 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 77000 (114 MB) 2023-01-10 09:40:26 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 78000 (115 MB) 2023-01-10 09:40:26 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 79000 (117 MB) 2023-01-10 09:40:27 destination > Flushing output: 4212 records (24 MB) 2023-01-10 09:40:27 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 80000 (118 MB) 2023-01-10 09:40:28 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 81000 (120 MB) 2023-01-10 09:40:28 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 82000 (121 MB) 2023-01-10 09:40:29 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 83000 (123 MB) 2023-01-10 09:40:30 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 84000 (124 MB) 2023-01-10 09:40:30 destination > Flushing output: 4207 records (24 MB) 2023-01-10 09:40:30 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 85000 (126 MB) 2023-01-10 09:40:31 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 86000 (127 MB) 2023-01-10 09:40:32 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 87000 (129 MB) 2023-01-10 09:40:32 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 88000 (130 MB) 2023-01-10 09:40:33 destination > Flushing output: 4188 records (24 MB) 2023-01-10 09:40:33 destination > Begin closing non current file writers 2023-01-10 09:40:33 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:40:33 destination > Closing csvPrinter when succeed 2023-01-10 09:40:33 destination > Upload completed for stream 'output'. 2023-01-10 09:40:33 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 89000 (132 MB) 2023-01-10 09:40:34 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 90000 (133 MB) 2023-01-10 09:40:34 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 91000 (135 MB) 2023-01-10 09:40:35 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 92000 (136 MB) 2023-01-10 09:40:36 destination > Flushing output: 4207 records (24 MB) 2023-01-10 09:40:36 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 93000 (138 MB) 2023-01-10 09:40:37 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 94000 (139 MB) 2023-01-10 09:40:37 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 95000 (141 MB) 2023-01-10 09:40:38 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 96000 (142 MB) 2023-01-10 09:40:38 destination > Flushing output: 4211 records (24 MB) 2023-01-10 09:40:39 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 97000 (144 MB) 2023-01-10 09:40:39 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 98000 (145 MB) 2023-01-10 09:40:40 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 99000 (147 MB) 2023-01-10 09:40:41 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 100000 (148 MB) 2023-01-10 09:40:41 destination > Flushing output: 4209 records (24 MB) 2023-01-10 09:40:41 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 101000 (150 MB) 2023-01-10 09:40:42 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 102000 (151 MB) 2023-01-10 09:40:43 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 103000 (153 MB) 2023-01-10 09:40:43 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 104000 (154 MB) 2023-01-10 09:40:44 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 105000 (156 MB) 2023-01-10 09:40:44 destination > Flushing output: 4206 records (24 MB) 2023-01-10 09:40:44 destination > Begin closing non current file writers 2023-01-10 09:40:44 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:40:44 destination > Closing csvPrinter when succeed 2023-01-10 09:40:45 destination > Upload completed for stream 'output'. 2023-01-10 09:40:45 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 106000 (157 MB) 2023-01-10 09:40:46 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 107000 (159 MB) 2023-01-10 09:40:46 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 108000 (160 MB) 2023-01-10 09:40:47 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 109000 (162 MB) 2023-01-10 09:40:47 destination > Flushing output: 4211 records (24 MB) 2023-01-10 09:40:48 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 110000 (163 MB) 2023-01-10 09:40:48 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 111000 (165 MB) 2023-01-10 09:40:49 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 112000 (166 MB) 2023-01-10 09:40:50 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 113000 (167 MB) 2023-01-10 09:40:50 destination > Flushing output: 4207 records (24 MB) 2023-01-10 09:40:51 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 114000 (169 MB) 2023-01-10 09:40:51 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 115000 (170 MB) 2023-01-10 09:40:52 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 116000 (172 MB) 2023-01-10 09:40:52 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 117000 (173 MB) 2023-01-10 09:40:53 destination > Flushing output: 4209 records (24 MB) 2023-01-10 09:40:53 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 118000 (175 MB) 2023-01-10 09:40:54 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 119000 (176 MB) 2023-01-10 09:40:55 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 120000 (178 MB) 2023-01-10 09:40:55 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 121000 (179 MB) 2023-01-10 09:40:56 destination > Flushing output: 4216 records (24 MB) 2023-01-10 09:40:56 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 122000 (181 MB) 2023-01-10 09:40:56 destination > Begin closing non current file writers 2023-01-10 09:40:56 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:40:56 destination > Closing csvPrinter when succeed 2023-01-10 09:40:57 destination > Upload completed for stream 'output'. 2023-01-10 09:40:57 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 123000 (182 MB) 2023-01-10 09:40:58 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 124000 (184 MB) 2023-01-10 09:40:58 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 125000 (185 MB) 2023-01-10 09:40:59 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 126000 (187 MB) 2023-01-10 09:40:59 destination > Flushing output: 4214 records (24 MB) 2023-01-10 09:41:00 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 127000 (188 MB) 2023-01-10 09:41:00 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 128000 (190 MB) 2023-01-10 09:41:01 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 129000 (191 MB) 2023-01-10 09:41:02 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 130000 (193 MB) 2023-01-10 09:41:02 destination > Flushing output: 4205 records (24 MB) 2023-01-10 09:41:03 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 131000 (194 MB) 2023-01-10 09:41:03 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 132000 (196 MB) 2023-01-10 09:41:04 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 133000 (197 MB) 2023-01-10 09:41:05 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 134000 (199 MB) 2023-01-10 09:41:05 destination > Flushing output: 4206 records (24 MB) 2023-01-10 09:41:05 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 135000 (200 MB) 2023-01-10 09:41:06 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 136000 (202 MB) 2023-01-10 09:41:07 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 137000 (203 MB) 2023-01-10 09:41:08 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 138000 (205 MB) 2023-01-10 09:41:08 destination > Flushing output: 4190 records (24 MB) 2023-01-10 09:41:08 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 139000 (206 MB) 2023-01-10 09:41:08 destination > Begin closing non current file writers 2023-01-10 09:41:08 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:41:08 destination > Closing csvPrinter when succeed 2023-01-10 09:41:09 destination > Upload completed for stream 'output'. 2023-01-10 09:41:09 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 140000 (208 MB) 2023-01-10 09:41:10 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 141000 (209 MB) 2023-01-10 09:41:11 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 142000 (211 MB) 2023-01-10 09:41:11 destination > Flushing output: 4205 records (24 MB) 2023-01-10 09:41:11 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 143000 (212 MB) 2023-01-10 09:41:12 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 144000 (214 MB) 2023-01-10 09:41:12 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 145000 (215 MB) 2023-01-10 09:41:13 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 146000 (217 MB) 2023-01-10 09:41:14 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 147000 (218 MB) 2023-01-10 09:41:14 destination > Flushing output: 4210 records (24 MB) 2023-01-10 09:41:15 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 148000 (219 MB) 2023-01-10 09:41:15 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 149000 (221 MB) 2023-01-10 09:41:16 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 150000 (222 MB) 2023-01-10 09:41:17 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 151000 (224 MB) 2023-01-10 09:41:17 destination > Flushing output: 4206 records (24 MB) 2023-01-10 09:41:17 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 152000 (225 MB) 2023-01-10 09:41:18 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 153000 (227 MB) 2023-01-10 09:41:19 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 154000 (228 MB) 2023-01-10 09:41:19 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 155000 (230 MB) 2023-01-10 09:41:20 destination > Flushing output: 4155 records (24 MB) 2023-01-10 09:41:20 destination > Begin closing non current file writers 2023-01-10 09:41:20 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:41:20 destination > Closing csvPrinter when succeed 2023-01-10 09:41:20 destination > Upload completed for stream 'output'. 2023-01-10 09:41:20 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 156000 (231 MB) 2023-01-10 09:41:21 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 157000 (233 MB) 2023-01-10 09:41:22 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 158000 (234 MB) 2023-01-10 09:41:22 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 159000 (236 MB) 2023-01-10 09:41:23 destination > Flushing output: 4152 records (24 MB) 2023-01-10 09:41:23 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 160000 (237 MB) 2023-01-10 09:41:24 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 161000 (239 MB) 2023-01-10 09:41:24 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 162000 (240 MB) 2023-01-10 09:41:25 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 163000 (242 MB) 2023-01-10 09:41:25 destination > Flushing output: 4184 records (24 MB) 2023-01-10 09:41:25 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 164000 (243 MB) 2023-01-10 09:41:26 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 165000 (245 MB) 2023-01-10 09:41:27 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 166000 (246 MB) 2023-01-10 09:41:27 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 167000 (248 MB) 2023-01-10 09:41:28 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 168000 (249 MB) 2023-01-10 09:41:28 destination > Flushing output: 4206 records (24 MB) 2023-01-10 09:41:29 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 169000 (251 MB) 2023-01-10 09:41:29 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 170000 (252 MB) 2023-01-10 09:41:30 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 171000 (254 MB) 2023-01-10 09:41:31 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 172000 (255 MB) 2023-01-10 09:41:31 destination > Flushing output: 4176 records (24 MB) 2023-01-10 09:41:31 destination > Begin closing non current file writers 2023-01-10 09:41:31 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:41:31 destination > Closing csvPrinter when succeed 2023-01-10 09:41:32 destination > Upload completed for stream 'output'. 2023-01-10 09:41:32 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 173000 (257 MB) 2023-01-10 09:41:33 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 174000 (258 MB) 2023-01-10 09:41:33 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 175000 (260 MB) 2023-01-10 09:41:34 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 176000 (261 MB) 2023-01-10 09:41:34 destination > Flushing output: 4189 records (24 MB) 2023-01-10 09:41:35 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 177000 (263 MB) 2023-01-10 09:41:35 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 178000 (264 MB) 2023-01-10 09:41:36 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 179000 (266 MB) 2023-01-10 09:41:37 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 180000 (267 MB) 2023-01-10 09:41:37 destination > Flushing output: 4211 records (24 MB) 2023-01-10 09:41:38 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 181000 (269 MB) 2023-01-10 09:41:38 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 182000 (270 MB) 2023-01-10 09:41:39 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 183000 (272 MB) 2023-01-10 09:41:40 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 184000 (273 MB) 2023-01-10 09:41:40 destination > Flushing output: 4211 records (24 MB) 2023-01-10 09:41:40 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 185000 (275 MB) 2023-01-10 09:41:41 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 186000 (276 MB) 2023-01-10 09:41:42 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 187000 (278 MB) 2023-01-10 09:41:42 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 188000 (279 MB) 2023-01-10 09:41:43 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 189000 (281 MB) 2023-01-10 09:41:43 destination > Flushing output: 4215 records (24 MB) 2023-01-10 09:41:43 destination > Begin closing non current file writers 2023-01-10 09:41:43 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:41:43 destination > Closing csvPrinter when succeed 2023-01-10 09:41:44 destination > Upload completed for stream 'output'. 2023-01-10 09:41:44 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 190000 (282 MB) 2023-01-10 09:41:45 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 191000 (284 MB) 2023-01-10 09:41:45 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 192000 (285 MB) 2023-01-10 09:41:46 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 193000 (287 MB) 2023-01-10 09:41:46 destination > Flushing output: 4211 records (24 MB) 2023-01-10 09:41:47 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 194000 (288 MB) 2023-01-10 09:41:47 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 195000 (290 MB) 2023-01-10 09:41:48 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 196000 (291 MB) 2023-01-10 09:41:49 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 197000 (293 MB) 2023-01-10 09:41:49 destination > Flushing output: 4210 records (24 MB) 2023-01-10 09:41:50 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 198000 (294 MB) 2023-01-10 09:41:51 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 199000 (295 MB) 2023-01-10 09:41:51 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):347 - Records read: 200000 (297 MB) 2023-01-10 09:41:51 source > finished reading a stream slice 2023-01-10 09:41:51 source > Read 200000 records from output stream 2023-01-10 09:41:51 source > Finished syncing output 2023-01-10 09:41:51 source > SourceS3 runtimes: Syncing stream output 0:02:22.376064 2023-01-10 09:41:51 source > Finished syncing SourceS3 2023-01-10 09:41:52 INFO i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$6):359 - Total records read: 200000 (297 MB) 2023-01-10 09:41:52 WARN i.a.w.g.DefaultReplicationWorker(lambda$readFromSrcAndWriteToDstRunnable$5):362 - Schema validation errors found for stream _output. Error messages: [$._ab_source_file_last_modified is of an incorrect type. Expected it to be date-time] 2023-01-10 09:41:52 INFO i.a.w.g.DefaultReplicationWorker(replicate):239 - One of source or destination thread complete. Waiting on the other. 2023-01-10 09:41:52 destination > Airbyte message consumer: succeeded. 2023-01-10 09:41:52 destination > executing on success close procedure. 2023-01-10 09:41:52 destination > Flushing output: 2521 records (14 MB) 2023-01-10 09:41:52 destination > Uploading remaining data for output stream. 2023-01-10 09:41:52 destination > Uploading remaining data for stream 'output'. 2023-01-10 09:41:52 destination > Closing csvPrinter when succeed 2023-01-10 09:41:52 destination > Upload completed for stream 'output'. 2023-01-10 09:41:52 destination > All data for output stream uploaded. 2023-01-10 09:41:52 destination > [Stream output] Creating database schema if it does not exist: public 2023-01-10 09:41:53 destination > HikariPool-1 - Driver does not support get/set network timeout for connections. ([Databricks][JDBC](10220) Driver does not support this optional feature.) 2023-01-10 09:41:53 destination > [Stream output] Creating tmp table _airbyte_tmp_zpw_output from staging file: abfss://dbtsparkstoragecontainer@dbtsparkstorageaccount.blob.core.windows.net/71270d3d-06af-4f7d-bb70-8dad60c0660d/public/_airbyte_tmp_zpw_output/