-
Notifications
You must be signed in to change notification settings - Fork 63
Remove DatabricksLibrary
external help file: azure.databricks.cicd.tools-help.xml Module Name: azure.databricks.cicd.tools online version: schema: 2.0.0
Remove a library from a cluster - note the cluster must be restarted to complete the uninstall (this command will NOT restart the cluster for you)
Remove-DatabricksLibrary [[-BearerToken] <String>] [[-Region] <String>] [-ClusterId] <String>
[-LibraryType] <String> [-LibrarySettings] <String> [<CommonParameters>]
Remove a library from a cluster - note the cluster must be restarted to complete the uninstall (this command will NOT restart the cluster for you)
Remove-DatabricksLibrary -BearerToken $BearerToken -Region $Region -ClusterId 'Bob-1234'
Your Databricks Bearer token to authenticate to your workspace (see User Settings in Datatbricks WebUI)
Type: String
Parameter Sets: (All)
Aliases:
Required: False
Position: 1
Default value: None
Accept pipeline input: False
Accept wildcard characters: False
Azure Region - must match the URL of your Databricks workspace, example northeurope
Type: String
Parameter Sets: (All)
Aliases:
Required: False
Position: 2
Default value: None
Accept pipeline input: False
Accept wildcard characters: False
ClusterId for existing Databricks cluster. Does not need to be running. See Get-DatabricksClusters.
Type: String
Parameter Sets: (All)
Aliases:
Required: True
Position: 3
Default value: None
Accept pipeline input: False
Accept wildcard characters: False
egg, jar, pypi, whl, cran, maven
Type: String
Parameter Sets: (All)
Aliases:
Required: True
Position: 4
Default value: None
Accept pipeline input: False
Accept wildcard characters: False
Settings can by path to jar (starting dbfs), pypi name (optionally with repo), or egg
If jar, URI of the jar to be installed. DBFS and S3 URIs are supported. For example: { "jar": "dbfs:/mnt/databricks/library.jar" } or { "jar": "s3://my-bucket/library.jar" }. If S3 is used, make sure the cluster has read access on the library. You may need to launch the cluster with an IAM role to access the S3 URI.
If egg, URI of the egg to be installed. DBFS and S3 URIs are supported. For example: { "egg": "dbfs:/my/egg" } or { "egg": "s3://my-bucket/egg" }. If S3 is used, make sure the cluster has read access on the library. You may need to launch the cluster with an IAM role to access the S3 URI.
If whl, URI of the wheel or zipped wheels to be installed. DBFS and S3 URIs are supported. For example: { "whl": "dbfs:/my/whl" } or { "whl": "s3://my-bucket/whl" }. If S3 is used, make sure the cluster has read access on the library. You may need to launch the cluster with an IAM role to access the S3 URI. Also the wheel file name needs to use the correct convention. If zipped wheels are to be installed, the file name suffix should be .wheelhouse.zip.
If pypi, specification of a PyPI library to be installed. For example: { "package": "simplejson" }
If maven, specification of a Maven library to be installed. For example: { "coordinates": "org.jsoup:jsoup:1.7.2" }
If cran, specification of a CRAN library to be installed.
Type: String
Parameter Sets: (All)
Aliases:
Required: True
Position: 5
Default value: None
Accept pipeline input: False
Accept wildcard characters: False
This cmdlet supports the common parameters: -Debug, -ErrorAction, -ErrorVariable, -InformationAction, -InformationVariable, -OutVariable, -OutBuffer, -PipelineVariable, -Verbose, -WarningAction, and -WarningVariable. For more information, see about_CommonParameters.
Author: Simon D'Morias / Data Thirst Ltd