You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/big-data-cluster/deployment-guidance.md
+24-23Lines changed: 24 additions & 23 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,9 +2,9 @@
2
2
title: Deployment guidance
3
3
titleSuffix: SQL Server Big Data Clusters
4
4
description: Learn how to deploy SQL Server Big Data Clusters on Kubernetes.
5
-
author: MikeRayMSFT
6
-
ms.author: mikeray
7
-
ms.reviewer: mihaelab
5
+
author: WilliamDAssafMSFT
6
+
ms.author: wiassaf
7
+
ms.reviewer:
8
8
ms.date: 06/22/2020
9
9
ms.topic: conceptual
10
10
ms.prod: sql
@@ -15,7 +15,7 @@ ms.technology: big-data-cluster
15
15
16
16
[!INCLUDE[SQL Server 2019](../includes/applies-to-version/sqlserver2019.md)]
17
17
18
-
A SQL Server big data cluster is deployed as docker containers on a Kubernetes cluster. This is an overview of the setup and configuration steps:
18
+
SQL Server Big Data Cluster is deployed as docker containers on a Kubernetes cluster. This is an overview of the setup and configuration steps:
19
19
20
20
- Set up a Kubernetes cluster on a single VM, cluster of VMs, in Azure Kubernetes Service (AKS), Red Hat OpenShift or in Azure Red Hat OpenShift (ARO).
21
21
- Install the cluster configuration tool [!INCLUDE [azure-data-cli-azdata](../includes/azure-data-cli-azdata.md)] on your client machine.
@@ -25,7 +25,7 @@ A SQL Server big data cluster is deployed as docker containers on a Kubernetes c
25
25
26
26
See [Supported platforms](release-notes-big-data-cluster.md#supported-platforms) for a complete list of the various Kubernetes platforms validated for deploying SQL Server Big Data Clusters.
27
27
28
-
### SQL Server Editions
28
+
### SQL Server editions
29
29
30
30
|Edition|Notes|
31
31
|---------|---------|
@@ -58,15 +58,16 @@ kubectl config view
58
58
```
59
59
60
60
> [!Important]
61
-
> If you are deploying on a multi node Kuberntes cluster that you bootstrapped using `kubeadm`, before starting the big data cluster deployment, ensure the clocks are synchronized across all the Kubernetes nodes the deployment is targeting. The big data cluster has built-in health properties for various services that are time sensitive and clock skews can result in incorrect status.
61
+
> If you are deploying on a multi node Kubernetes cluster that you bootstrapped using `kubeadm`, before starting the big data cluster deployment, ensure the clocks are synchronized across all the Kubernetes nodes the deployment is targeting. The big data cluster has built-in health properties for various services that are time sensitive and clock skews can result in incorrect status.
62
62
63
63
After you have configured your Kubernetes cluster, you can proceed with the deployment of a new SQL Server big data cluster. If you are upgrading from a previous release, please see [How to upgrade [!INCLUDE[big-data-clusters-2019](../includes/ssbigdataclusters-ss-nover.md)]](deployment-upgrade.md).
64
64
65
65
## Ensure you have storage configured
66
66
67
-
Most big data cluster deployments should have persistent storage. At this time, you need to make sure you have a plan for how you're going to provide persistent storage on the Kubernetes cluster before you deploy the BDC.
67
+
Most big data cluster deployments should have persistent storage. At this time, you need to make sure you have a plan for how you're going to provide persistent storage on the Kubernetes cluster before you deploy.
68
68
69
-
If you deploy in AKS, no storage setup is necessary. AKS provides built-in storage classes with dynamic provisioning. You can customize the storage class (`default` or `managed-premium`) in the deployment configuration file. The built-in profiles use a `default` storage class. If you are deploying on a Kubernetes cluster you deployed using `kubeadm`, you'll need to ensure you have sufficient storage for a cluster of your desired scale available and configured for use. If you wish to customize how your storage is used, you should do this before proceeding. See [Data persistence with SQL Server big data cluster on Kubernetes](concept-data-persistence.md).
69
+
- If you deploy in AKS, no storage setup is necessary. AKS provides built-in storage classes with dynamic provisioning. You can customize the storage class (`default` or `managed-premium`) in the deployment configuration file. The built-in profiles use a `default` storage class.
70
+
- If you are deploying on a Kubernetes cluster you deployed using `kubeadm`, you'll need to ensure you have sufficient storage for a cluster of your desired scale available and configured for use. If you wish to customize how your storage is used, you should do this before proceeding. See [Data persistence with SQL Server big data cluster on Kubernetes](concept-data-persistence.md).
70
71
71
72
## Install SQL Server 2019 Big Data tools
72
73
@@ -92,24 +93,24 @@ Big data cluster deployment options are defined in JSON configuration files. You
92
93
> [!NOTE]
93
94
> The container images required for the big data cluster deployment are hosted on Microsoft Container Registry (`mcr.microsoft.com`), in the `mssql/bdc` repository. By default, these settings are already included in the `control.json` configuration file in each of the deployment profiles included with [!INCLUDE [azure-data-cli-azdata](../includes/azure-data-cli-azdata.md)]. In addition, the container image tag for each release is also pre-populated in the same configuration file. If you need to pull the container images into your own private container registry and or modify the container registry/repository settings, follow the instructions in the [Offline installation article](deploy-offline.md)
94
95
95
-
Run this command to find what are the templates available:
96
+
Run this command to find the templates available:
96
97
97
-
```
98
+
```bash
98
99
azdata bdc config list -o table
99
100
```
100
101
101
102
The following templates are available as of SQL Server 2019 CU5:
102
103
103
104
| Deployment profile | Kubernetes environment |
104
105
|---|---|
105
-
|`aks-dev-test`| Deploy SQL Server big data cluster on Azure Kubernetes Service (AKS)|
106
-
|`aks-dev-test-ha`| Deploy SQL Server big data cluster on Azure Kubernetes Service (AKS). Mission critical services like SQL Server master and HDFS name node are configured for high availability.|
107
-
|`aro-dev-test`|Deploy SQL Server big data cluster on Azure Red Hat OpenShift for development and testing. <br/><br/>Introduced in SQL Server 2019 CU 5.|
108
-
|`aro-dev-test-ha`|Deploy SQL Server big data cluster with high availability on a Red Hat OpenShift cluster for development and testing. <br/><br/>Introduced in SQL Server 2019 CU 5.|
109
-
|`kubeadm-dev-test`| Deploy SQL Server big data cluster on a Kubernetes cluster created with kubeadm using a single or multiple physical or virtual machines.|
110
-
|`kubeadm-prod`| Deploy SQL Server big data cluster on a Kubernetes cluster created with kubeadm using a single or multiple physical or virtual machines. Use this template to enable big data cluster services to integrate with Active Directory. Mission critical services like SQL Server master instance and HDFS name node are deployed in a highly available configuration. |
111
-
|`openshift-dev-test`|Deploy SQL Server big data cluster on a Red Hat OpenShift cluster for development and testing. <br/><br/>Introduced in SQL Server 2019 CU 5.|
112
-
|`openshift-prod`|Deploy SQL Server big data cluster with high availability on a Red Hat OpenShift cluster. <br/><br/>Introduced in SQL Server 2019 CU 5.|
106
+
|`aks-dev-test`| Deploy SQL Server Big Data Cluster on Azure Kubernetes Service (AKS)|
107
+
|`aks-dev-test-ha`| Deploy SQL Server Big Data Cluster on Azure Kubernetes Service (AKS). Mission critical services like SQL Server master and HDFS name node are configured for high availability.|
108
+
|`aro-dev-test`|Deploy SQL Server Big Data Cluster on Azure Red Hat OpenShift for development and testing. <br/><br/>Introduced in SQL Server 2019 CU 5.|
109
+
|`aro-dev-test-ha`|Deploy SQL Server Big Data Cluster with high availability on a Red Hat OpenShift cluster for development and testing. <br/><br/>Introduced in SQL Server 2019 CU 5.|
110
+
|`kubeadm-dev-test`| Deploy SQL Server Big Data Cluster on a Kubernetes cluster created with kubeadm using a single or multiple physical or virtual machines.|
111
+
|`kubeadm-prod`| Deploy SQL Server Big Data Cluster on a Kubernetes cluster created with kubeadm using a single or multiple physical or virtual machines. Use this template to enable big data cluster services to integrate with Active Directory. Mission critical services like SQL Server master instance and HDFS name node are deployed in a highly available configuration. |
112
+
|`openshift-dev-test`|Deploy SQL Server Big Data Cluster on a Red Hat OpenShift cluster for development and testing. <br/><br/>Introduced in SQL Server 2019 CU 5.|
113
+
|`openshift-prod`|Deploy SQL Server Big Data Cluster with high availability on a Red Hat OpenShift cluster. <br/><br/>Introduced in SQL Server 2019 CU 5.|
113
114
114
115
You can deploy a big data cluster by running `azdata bdc create`. This prompts you to choose one of the default configurations and then guides you through the deployment.
115
116
@@ -150,27 +151,27 @@ It is also possible to customize your deployment to accommodate the workloads yo
150
151
```
151
152
152
153
> [!TIP]
153
-
> You can also pass in the cluster name at deployment time using the *--name* parameter for *azdata create bdc* command. The parameters in the command have precedence over the values in the configuration files.
154
+
> You can also pass in the cluster name at deployment time using the *--name* parameter for `azdata create bdc` command. The parameters in the command have precedence over the values in the configuration files.
154
155
>
155
156
> A useful tool for finding JSON paths is the [JSONPath Online Evaluator](https://jsonpath.com/).
156
157
>
157
-
In addition to passing key-value pairs, you can also provide inline JSON values or pass JSON patch files. For more information, see [Configure deployment settings for big data clusters](deployment-custom-configuration.md).
158
+
In addition to passing key-value pairs, you can also provide inline JSON values or pass JSON patch files. For more information, see [Configure deployment settings for Big Data Cluster resources and services](deployment-custom-configuration.md).
158
159
159
160
1. Pass the custom configuration file to `azdata bdc create`. Note that you must set the required [environment variables](#env), otherwise the terminal prompts for the values:
> For more information on the structure of a deployment configuration file, see the [Deployment configuration file reference](reference-deployment-config.md). For more configuration examples, see [Configure deployment settings for big data clusters](deployment-custom-configuration.md).
166
+
> For more information on the structure of a deployment configuration file, see the [Deployment configuration file reference](reference-deployment-config.md). For more configuration examples, see [Configure deployment settings for Big Data Clusters](deployment-custom-configuration.md).
166
167
167
168
## <aid="env"></a> Environment variables
168
169
169
170
The following environment variables are used for security settings that are not stored in a deployment configuration file. Note that Docker settings except credentials can be set in the configuration file.
|`AZDATA_USERNAME`| Required |The username for SQL Server big data cluster administrator. A sysadmin login with the same name is created in SQL Server master instance. As a security best practice, `sa` account is disabled. <br/><br/>[!INCLUDE [big-data-cluster-root-user](../includes/big-data-cluster-root-user.md)]|
174
+
|`AZDATA_USERNAME`| Required |The username for SQL Server Big Data Cluster administrator. A sysadmin login with the same name is created in SQL Server master instance. As a security best practice, `sa` account is disabled. <br/><br/>[!INCLUDE [big-data-cluster-root-user](../includes/big-data-cluster-root-user.md)]|
174
175
|`AZDATA_PASSWORD`| Required |The password for the user accounts created above. On clusters deployed prior to SQL Server 2019 CU5, the same password is used for the `root` user, to secure Knox gateway and HDFS. |
175
176
|`ACCEPT_EULA`| Required for first use of [!INCLUDE [azure-data-cli-azdata](../includes/azure-data-cli-azdata.md)]| Set to "yes". When set as an environment variable, it applies EULA to both SQL Server and [!INCLUDE [azure-data-cli-azdata](../includes/azure-data-cli-azdata.md)]. If not set as environment variable, you can include `--accept-eula=yes` in the first use of [!INCLUDE [azure-data-cli-azdata](../includes/azure-data-cli-azdata.md)] command.|
176
177
|`DOCKER_USERNAME`| Optional | The username to access the container images in case they are stored in a private repository. See the [Offline deployments](deploy-offline.md) topic for more details on how to use a private Docker repository for big data cluster deployment.|
@@ -428,7 +429,7 @@ For more information on how to connect to the big data cluster, see [Connect to
428
429
429
430
## Next steps
430
431
431
-
To learn more about big data cluster deployment, see the following resources:
432
+
To learn more about SQL Server Big Data Cluster deployment, see the following resources:
432
433
433
434
-[Configure deployment settings for big data clusters](deployment-custom-configuration.md)
434
435
-[Perform an offline deployment of a SQL Server big data cluster](deploy-offline.md)
0 commit comments