Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

iceberg.catalog.warehouse can override from connector config? #274

Open
chethan-bs opened this issue Jul 4, 2024 · 1 comment
Open

iceberg.catalog.warehouse can override from connector config? #274

chethan-bs opened this issue Jul 4, 2024 · 1 comment

Comments

@chethan-bs
Copy link

chethan-bs commented Jul 4, 2024

i am trying to use reset URL for ADLS and multiple storage container from connector config but i am unable to override the CATALOG_WAREHOUSE env from iceberg-rest.

"iceberg.catalog.warehouse": "abfss://[email protected]/warehouse", is not taking any action change from connector config and aways going with default from the rest.

could you please help on this configuration ?

"iceberg.catalog.io-impl": "org.apache.iceberg.azure.adlsv2.ADLSFileIO" works fine from connector config

@kevingomez93
Copy link

kevingomez93 commented Sep 30, 2024

Hello, I am experiencing a similar issue, but with the S3 implementation instead of Azure. My connector configuration includes the following:

connector.class=io.tabular.iceberg.connect.IcebergSinkConnector
transforms.DeriveEventDate.format=yyyy-MM-dd
iceberg.tables.evolve-schema-enabled=true
tasks.max=1
iceberg.tables.schema-override.event_date.type=Date
transforms=ExtractField,CopyId,DeriveEventDate
transforms.DeriveEventDate.type=org.apache.kafka.connect.transforms.TimestampConverter$Value
iceberg.catalog.client.region=us-east-1
transforms.CopyId.type=io.tabular.iceberg.connect.transforms.CopyValue
iceberg.tables.auto-create-enabled=true
iceberg.control.commit.timeout-ms=1480000
transforms.CopyId.source.field=datetime
iceberg.tables=schema.some_topic
transforms.DeriveEventDate.target.type=Date
transforms.DeriveEventDate.input.format=UNIX_MS
value.converter=org.apache.kafka.connect.json.JsonConverter
key.converter=org.apache.kafka.connect.storage.StringConverter
transforms.DeriveEventDate.field=event_date
iceberg.tables.default-partition-by=event_date
topics=some_topic
transforms.ExtractField.field=event
iceberg.catalog.io-impl=org.apache.iceberg.aws.s3.S3FileIO
iceberg.control.commit.interval-ms=120000
iceberg.catalog.uri=thrift://some-hive-connection:9083
value.converter.schemas.enable=false
transforms.CopyId.target.field=event_date
iceberg.catalog.warehouse=s3a://some-bucket/
iceberg.catalog.type=hive
iceberg.catalog.s3.path-style-access=true
transforms.ExtractField.type=org.apache.kafka.connect.transforms.ExtractField$Value

Despite specifying iceberg.catalog.warehouse=s3a://some-bucket/, it doesn't seem to override the default S3 bucket, and the configuration is still taking the default value from the Hive service.

Has anyone found a solution to this problem, or is there something we might be missing?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants