diff --git a/getting-started/eclipselink/docker-compose.yml b/getting-started/eclipselink/docker-compose.yml index 985a6a9e54..58e89390d3 100644 --- a/getting-started/eclipselink/docker-compose.yml +++ b/getting-started/eclipselink/docker-compose.yml @@ -75,13 +75,13 @@ services: /opt/spark/bin/spark-sql, --packages, "org.apache.iceberg:iceberg-spark-runtime-3.5_2.12:1.7.0,software.amazon.awssdk:bundle:2.28.17,software.amazon.awssdk:url-connection-client:2.28.17,org.apache.iceberg:iceberg-gcp-bundle:1.7.0,org.apache.iceberg:iceberg-azure-bundle:1.7.0", --conf, "spark.sql.extensions=org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions", - --conf, "spark.sql.catalog.polaris=org.apache.iceberg.spark.SparkCatalog", - --conf, "spark.sql.catalog.polaris.type=rest", - --conf, "spark.sql.catalog.polaris.warehouse=quickstart_catalog", - --conf, "spark.sql.catalog.polaris.uri=http://polaris:8181/api/catalog", - --conf, "spark.sql.catalog.polaris.credential=root:s3cr3t", - --conf, "spark.sql.catalog.polaris.scope=PRINCIPAL_ROLE:ALL", - --conf, "spark.sql.defaultCatalog=polaris", + --conf, "spark.sql.catalog.quickstart_catalog=org.apache.iceberg.spark.SparkCatalog", + --conf, "spark.sql.catalog.quickstart_catalog.type=rest", + --conf, "spark.sql.catalog.quickstart_catalog.warehouse=quickstart_catalog", + --conf, "spark.sql.catalog.quickstart_catalog.uri=http://polaris:8181/api/catalog", + --conf, "spark.sql.catalog.quickstart_catalog.credential=root:s3cr3t", + --conf, "spark.sql.catalog.quickstart_catalog.scope=PRINCIPAL_ROLE:ALL", + --conf, "spark.sql.defaultCatalog=quickstart_catalog", --conf, "spark.sql.catalogImplementation=in-memory", --conf, "spark.driver.extraJavaOptions=-Divy.cache.dir=/tmp -Divy.home=/tmp" ] diff --git a/site/content/in-dev/unreleased/getting-started/using-polaris.md b/site/content/in-dev/unreleased/getting-started/using-polaris.md index 14524e495a..3203122752 100644 --- a/site/content/in-dev/unreleased/getting-started/using-polaris.md +++ b/site/content/in-dev/unreleased/getting-started/using-polaris.md @@ -174,7 +174,7 @@ Replace the credentials used in the Docker container using the following code: ```shell USER_CLIENT_ID="XXXX" USER_CLIENT_SECRET="YYYY" -sed -i "s/^\(.*spark\.sql\.catalog\.polaris\.credential=\).*/\1${USER_CLIENT_ID}:${USER_CLIENT_SECRET}\",/" getting-started/eclipselink/docker-compose.yml +sed -i "s/^\(.*spark\.sql\.catalog\.quickstart_catalog\.credential=\).*/\1${USER_CLIENT_ID}:${USER_CLIENT_SECRET}\",/" getting-started/eclipselink/docker-compose.yml docker compose -f getting-started/eclipselink/docker-compose.yml up -d ``` @@ -189,7 +189,7 @@ docker attach $(docker ps -q --filter name=spark-sql) Once the Spark session starts, we can create a namespace and table within the catalog: ```sql -USE polaris; +USE quickstart_catalog; CREATE NAMESPACE IF NOT EXISTS quickstart_namespace; CREATE NAMESPACE IF NOT EXISTS quickstart_namespace.schema; USE NAMESPACE quickstart_namespace.schema;