Skip to content

Commit

Permalink
Update docs for Spark client v0.14.0 (#7901)
Browse files Browse the repository at this point in the history
  • Loading branch information
itaigilo authored Jun 20, 2024
1 parent 41a937e commit 7e801f5
Show file tree
Hide file tree
Showing 4 changed files with 10 additions and 10 deletions.
4 changes: 2 additions & 2 deletions clients/spark/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ io.lakefs:lakefs-spark-client_2.12:${CLIENT_VERSION}
Replace `<version>` below with the latest version available. See [available versions](https://mvnrepository.com/artifact/io.lakefs/lakefs-spark-client_2.12).

```
CLIENT_VERSION=0.11.0
CLIENT_VERSION=0.14.0
spark-submit --conf spark.hadoop.lakefs.api.url=https://lakefs.example.com/api/v1 \
--conf spark.hadoop.fs.s3a.access.key=<S3_ACCESS_KEY> \
--conf spark.hadoop.fs.s3a.secret.key=<S3_SECRET_KEY> \
Expand All @@ -42,7 +42,7 @@ spark-submit --conf spark.hadoop.lakefs.api.url=https://lakefs.example.com/api/v
Replace `<version>` below with the latest version available. See [available versions](https://mvnrepository.com/artifact/io.lakefs/lakefs-spark-client_2.12).

```
CLIENT_VERSION=0.11.0
CLIENT_VERSION=0.14.0
spark-submit --conf spark.hadoop.lakefs.api.url=https://lakefs.example.com/api/v1 \
--conf spark.hadoop.fs.s3a.access.key=<S3_ACCESS_KEY> \
--conf spark.hadoop.fs.s3a.secret.key=<S3_SECRET_KEY> \
Expand Down
2 changes: 1 addition & 1 deletion docs/howto/export.md
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,7 @@ The complete `spark-submit` command would look as follows:
spark-submit --conf spark.hadoop.lakefs.api.url=https://<LAKEFS_ENDPOINT>/api/v1 \
--conf spark.hadoop.lakefs.api.access_key=<LAKEFS_ACCESS_KEY_ID> \
--conf spark.hadoop.lakefs.api.secret_key=<LAKEFS_SECRET_ACCESS_KEY> \
--packages io.lakefs:lakefs-spark-client_2.12:0.11.0 \
--packages io.lakefs:lakefs-spark-client_2.12:0.14.0 \
--class io.treeverse.clients.Main export-app example-repo s3://example-bucket/prefix \
--branch=example-branch
```
Expand Down
8 changes: 4 additions & 4 deletions docs/howto/garbage-collection/gc.md
Original file line number Diff line number Diff line change
Expand Up @@ -121,7 +121,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \
-c spark.hadoop.lakefs.api.secret_key=<LAKEFS_SECRET_KEY> \
-c spark.hadoop.fs.s3a.access.key=<S3_ACCESS_KEY> \
-c spark.hadoop.fs.s3a.secret.key=<S3_SECRET_KEY> \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.11.0/lakefs-spark-client-assembly-0.11.0.jar \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar \
example-repo us-east-1
```
</div>
Expand All @@ -136,7 +136,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \
-c spark.hadoop.lakefs.api.access_key=<LAKEFS_ACCESS_KEY> \
-c spark.hadoop.lakefs.api.secret_key=<LAKEFS_SECRET_KEY> \
-c spark.hadoop.fs.azure.account.key.<AZURE_STORAGE_ACCOUNT>.dfs.core.windows.net=<AZURE_STORAGE_ACCESS_KEY> \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.11.0/lakefs-spark-client-assembly-0.11.0.jar \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar \
example-repo
```

Expand All @@ -153,7 +153,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \
-c spark.hadoop.fs.azure.account.oauth2.client.id.<AZURE_STORAGE_ACCOUNT>.dfs.core.windows.net=<application-id> \
-c spark.hadoop.fs.azure.account.oauth2.client.secret.<AZURE_STORAGE_ACCOUNT>.dfs.core.windows.net=<service-credential-key> \
-c spark.hadoop.fs.azure.account.oauth2.client.endpoint.<AZURE_STORAGE_ACCOUNT>.dfs.core.windows.net=https://login.microsoftonline.com/<directory-id>/oauth2/token \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.11.0/lakefs-spark-client-assembly-0.11.0.jar \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar \
example-repo
```

Expand Down Expand Up @@ -181,7 +181,7 @@ spark-submit --class io.treeverse.gc.GarbageCollection \
-c spark.hadoop.fs.gs.impl=com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystem \
-c spark.hadoop.fs.AbstractFileSystem.gs.impl=com.google.cloud.hadoop.fs.gcs.GoogleHadoopFS \
-c spark.hadoop.lakefs.gc.do_sweep=false \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.11.0/lakefs-spark-client-assembly-0.11.0.jar \
http://treeverse-clients-us-east.s3-website-us-east-1.amazonaws.com/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar \
example-repo
```
Expand Down
6 changes: 3 additions & 3 deletions docs/reference/spark-client.md
Original file line number Diff line number Diff line change
Expand Up @@ -31,18 +31,18 @@ can work for other Spark versions and higher Hadoop versions.
Start Spark Shell / PySpark with the `--packages` flag, for instance:

```bash
spark-shell --packages io.lakefs:lakefs-spark-client_2.12:0.13.0
spark-shell --packages io.lakefs:lakefs-spark-client_2.12:0.14.0
```

Alternatively use the assembled jar (an "Überjar") on S3, from
`s3://treeverse-clients-us-east/lakefs-spark-client/0.13.0/lakefs-spark-client-assembly-0.13.0.jar`
`s3://treeverse-clients-us-east/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar`
by passing its path to `--jars`.
The assembled jar is larger but shades several common libraries. Use it if Spark
complains about bad classes or missing methods.
</div>
<div markdown="1" id="databricks">
Include this assembled jar (an "Überjar") from S3, from
`s3://treeverse-clients-us-east/lakefs-spark-client/0.13.0/lakefs-spark-client-assembly-0.13.0.jar`.
`s3://treeverse-clients-us-east/lakefs-spark-client/0.14.0/lakefs-spark-client-assembly-0.14.0.jar`.
</div>
</div>

Expand Down

0 comments on commit 7e801f5

Please sign in to comment.