Big Data News Hubb
Advertisement
  • Home
  • Big Data
  • News
  • Contact us
No Result
View All Result
  • Home
  • Big Data
  • News
  • Contact us
No Result
View All Result
Big Data News Hubb
No Result
View All Result
Home Big Data

Announcing Terraform Databricks modules – The Databricks Blog

admin by admin
May 8, 2023
in Big Data


The Databricks Terraform provider reached more than 10 million installations, significantly increasing adoption since it became generally available less than one year ago.

This significant milestone showcases that Terraform and the Databricks provider are widely used by many customers to automate infrastructure deployment and management of their Lakehouse Platform.

Figure: Monthly downloads of Databricks Terraform provider

To easily maintain, manage and scale their infrastructure, DevOps teams build their infrastructure using modular and reusable components called Terraform modules. Terraform modules allow you to easily reuse the same components across multiple use cases and environments. It also helps enforce a standardized approach of defining resources and adopting best practices across your organization. Not only does consistency ensure best practices are followed, it also helps to enforce compliant deployment and avoid accidental misconfigurations, which could lead to costly errors.

Introducing Terraform Databricks modules

To help customers test and deploy their Lakehouse environments, we’re releasing the experimental Terraform Registry modules for Databricks, a set of more than 30 reusable Terraform modules and examples to provision your Databricks Lakehouse platform on Azure, AWS, and GCP using Databricks Terraform provider. It is using the content in terraform-databricks-examples github repository.

Databricks Terraform modules available on Terraform Registry
Figure: Databricks Terraform modules available on Terraform Registry

There are two ways to use these modules:

  1. Use examples as a reference for your own Terraform code.
  2. Directly reference the submodule in your Terraform configuration.

The full set of available modules and examples can be found in Terraform Registry modules for Databricks.

Getting started with Terraform Databricks examples

To use one of the different available Terraform Databricks modules, you should follow these steps:

  1. Reference this module using one of the different module source types
  2. Add a variables.tf file with the required inputs for the module
  3. Add a terraform.tfvars file and provide values to each defined variable
  4. Add an output.tf file with the module outputs
  5. (Strongly recommended) Configure your remote backend
  6. Run terraform init to initialize terraform and to download the needed providers.
  7. Run terraform validate to validate the configuration files in your directory.
  8. Run terraform plan to preview the resources that Terraform plans to create.
  9. Run terraform apply to create the resources.

Example walkthrough

In this section we demonstrate how to use the examples provided in the Databricks Terraform registry modules page. Each example is independent of each other and has a dedicated README.md. We now use this Azure Databricks example adb-vnet-injection to deploy a VNet-injected Databricks Workspace with an auto scaling cluster.

Step 1: Authenticate to the providers.

Navigate to providers.tf to check for providers used in the example, here we need to configure authentication to Azure provider and Databricks providers. Read the following docs for extensive information on how to configure authentication to providers:

Step 2: Read through the readme of the example, prepare input values.

The identity that you use in az login to deploy this template should have a contributor role in your azure subscription, or the minimum required permissions to deploy resources in this template.

Then do following steps:

  1. Run terraform init to initialize terraform and download required providers.
  2. Run terraform plan to see what resources will be deployed.
  3. Run terraform apply to deploy resources to your Azure environment.
    Since we used az login method to authenticate to providers, you will be prompted to login to Azure via browser. Enter yes when prompted to deploy resources.

Step 3: Verify that resources were deployed successfully in your environment.

Navigate to Azure Portal and verify that all resources were deployed successfully. You should now have a vnet-injected workspace with one cluster deployed.

Azure Portal
Compute

You can find more examples and their requirements, provider authentications in this repository under /examples.

How to contribute

Terraform-databricks-examples content will be continuously updated with new modules covering different architectures and also more features of the Databricks platform.

Note that it is a community project, developed by Databricks Field Engineering and is provided as-is. Databricks does not offer official support. In order to add new examples or new modules, you can contribute to the project by opening a new pull-request. For any issue, please open an issue under the terraform-databricks-examples repository.

Please look into the current list of available modules and examples and try them to deploy your Databricks infrastructure!



Source link

Previous Post

Improve reliability and reduce costs of your Apache Spark workloads with vertical autoscaling on Amazon EMR on EKS

Next Post

State of Java Ecosystem Report

Next Post

State of Java Ecosystem Report

Recommended

Data Quality Is Getting Worse, Monte Carlo Says

May 7, 2023

HPE Brings Analytics Together on its Data Fabric

May 20, 2023

Retail Business Intelligence Uses Predictive Analytics To Better Serve Customers

March 15, 2023

Don't miss it

News

BWH Hotels scales enterprise business intelligence adoption while reducing costs with Amazon QuickSight

June 3, 2023
News

Snowflake Bolsters Data Cloud Search Capabilities with Neeva Acquisition

June 3, 2023
News

From Small To Big: Tips On Growing Your Business Successfully

June 2, 2023
Big Data

AI Empowers Microfinance: Revolutionizing Fraud Detection

June 2, 2023
Big Data

Handling “Right to be Forgotten” in GDPR and CCPA using Delta Live Tables (DLT)

June 2, 2023
News

Real-time inference using deep learning within Amazon Kinesis Data Analytics for Apache Flink

June 2, 2023
big-data-footer-white

© Big Data News Hubb All rights reserved.

Use of these names, logos, and brands does not imply endorsement unless specified. By using this site, you agree to the Privacy Policy and Terms & Conditions.

Navigate Site

  • Home
  • Big Data
  • News
  • Contact us

Newsletter Sign Up

No Result
View All Result
  • Home
  • Big Data
  • News
  • Contact us

© 2022 Big Data News Hubb All rights reserved.