-
Notifications
You must be signed in to change notification settings - Fork 672
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Signed-off-by: Kevin Su <[email protected]>
- Loading branch information
Showing
5 changed files
with
134 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -107,6 +107,7 @@ helm install gateway bitnami/contour -n flyte | |
| configmap.task_logs.plugins.logs.cloudwatch-enabled | bool | `false` | One option is to enable cloudwatch logging for EKS, update the region and log group accordingly | | ||
| configmap.task_resource_defaults | object | `{"task_resources":{"defaults":{"cpu":"100m","memory":"500Mi","storage":"500Mi"},"limits":{"cpu":2,"gpu":1,"memory":"1Gi","storage":"20Mi"}}}` | Task default resources configuration Refer to the full [structure](https://pkg.go.dev/github.com/lyft/[email protected]/pkg/runtime/interfaces#TaskResourceConfiguration). | | ||
| configmap.task_resource_defaults.task_resources | object | `{"defaults":{"cpu":"100m","memory":"500Mi","storage":"500Mi"},"limits":{"cpu":2,"gpu":1,"memory":"1Gi","storage":"20Mi"}}` | Task default resources parameters | | ||
| databricks | object | `{"enabled":false,"plugin_config":{"plugins":{"databricks":{"databricksInstance":"dbc-a53b7a3c-614c","entrypointFile":"dbfs:///FileStore/tables/entrypoint.py"}}}}` | Optional: Databricks Plugin allows us to run the spark job on the Databricks platform. | | ||
| datacatalog.affinity | object | `{}` | affinity for Datacatalog deployment | | ||
| datacatalog.configPath | string | `"/etc/datacatalog/config/*.yaml"` | Default regex string for searching configuration files | | ||
| datacatalog.enabled | bool | `true` | | | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,103 @@ | ||
.. _deployment-plugin-setup-webapi-databricks: | ||
|
||
Databricks Plugin Setup | ||
----------------------- | ||
|
||
This guide gives an overview of how to set up Databricks in your Flyte deployment. | ||
|
||
1. Add Flyte chart repo to Helm | ||
|
||
.. code-block:: | ||
helm repo add flyteorg https://flyteorg.github.io/flyte | ||
2. Setup the cluster | ||
|
||
.. tabbed:: Sandbox | ||
|
||
* Start the sandbox cluster | ||
|
||
.. code-block:: bash | ||
flytectl sandbox start | ||
* Generate Flytectl sandbox config | ||
|
||
.. code-block:: bash | ||
flytectl config init | ||
.. tabbed:: AWS/GCP | ||
|
||
* Make sure you have up and running flyte cluster in `AWS <https://docs.flyte.org/en/latest/deployment/aws/index.html#deployment-aws>`__ / `GCP <https://docs.flyte.org/en/latest/deployment/gcp/index.html#deployment-gcp>`__ | ||
* Make sure you have correct kubeconfig and selected the correct kubernetes context | ||
* make sure you have the correct flytectl config at ~/.flyte/config.yaml | ||
|
||
3. Upload an `entrypoint.py <https://gist.github.com/pingsutw/482e7f0134414dac437500344bac5134>`__ to dbfs or s3. Spark driver node run this file to override the default command in the dbx job. | ||
|
||
4. Create a file named ``values-override.yaml`` and add the following config to it: | ||
|
||
.. code-block:: yaml | ||
configmap: | ||
enabled_plugins: | ||
# -- Tasks specific configuration [structure](https://pkg.go.dev/github.com/flyteorg/flytepropeller/pkg/controller/nodes/task/config#GetConfig) | ||
tasks: | ||
# -- Plugins configuration, [structure](https://pkg.go.dev/github.com/flyteorg/flytepropeller/pkg/controller/nodes/task/config#TaskPluginConfig) | ||
task-plugins: | ||
# -- [Enabled Plugins](https://pkg.go.dev/github.com/flyteorg/flyteplugins/go/tasks/config#Config). Enable sagemaker*, athena if you install the backend | ||
# plugins | ||
enabled-plugins: | ||
- container | ||
- sidecar | ||
- k8s-array | ||
- databricks | ||
default-for-task-types: | ||
container: container | ||
sidecar: sidecar | ||
container_array: k8s-array | ||
spark: databricks | ||
databricks: | ||
enabled: True | ||
plugin_config: | ||
plugins: | ||
databricks: | ||
entrypointFile: dbfs:///FileStore/tables/entrypoint.py | ||
databricksInstance: dbc-a53b7a3c-614c | ||
5. Create a Databricks account and follow the docs for creating an Access token. | ||
|
||
6. Create a `Instance Profile <https://docs.databricks.com/administration-guide/cloud-configurations/aws/instance-profiles.html>`_ for the Spark cluster, it allows the spark job to access your data in the s3 bucket. | ||
|
||
7. Add Databricks access token to FlytePropeller. | ||
|
||
.. note:: | ||
Refer to the `access token <https://docs.databricks.com/dev-tools/auth.html#databricks-personal-access-tokens>`__ to understand setting up the Databricks access token. | ||
|
||
.. code-block:: bash | ||
kubectl edit secret -n flyte flyte-secret-auth | ||
The configuration will look as follows: | ||
|
||
.. code-block:: yaml | ||
apiVersion: v1 | ||
data: | ||
FLYTE_DATABRICKS_API_TOKEN: <ACCESS_TOKEN> | ||
client_secret: Zm9vYmFy | ||
kind: Secret | ||
metadata: | ||
annotations: | ||
meta.helm.sh/release-name: flyte | ||
meta.helm.sh/release-namespace: flyte | ||
... | ||
Replace ``<ACCESS_TOKEN>`` with your access token. | ||
|
||
8. Upgrade the Flyte Helm release. | ||
|
||
.. code-block:: bash | ||
helm upgrade -n flyte -f https://raw.githubusercontent.com/flyteorg/flyte/master/charts/flyte-core/values-sandbox.yaml -f values-override.yaml flyteorg/flyte-core |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters