diff --git a/docs/data-sources/dms_kafka_flavors.md b/docs/data-sources/dms_kafka_flavors.md
new file mode 100644
index 000000000..89a60a9c1
--- /dev/null
+++ b/docs/data-sources/dms_kafka_flavors.md
@@ -0,0 +1,143 @@
+---
+subcategory: "Distributed Message Service (DMS)"
+---
+
+# flexibleengine_dms_kafka_flavors
+
+Use this data source to get the list of available flavor details within FlexibleEngine.
+
+## Example Usage
+
+### Query the list of kafka flavors for cluster type
+
+```hcl
+data "flexibleengine_dms_kafka_flavors" "test" {
+ type = "cluster"
+}
+```
+
+### Query the kafka flavor details of the specified ID
+
+```hcl
+data "flexibleengine_dms_kafka_flavors" "test" {
+ flavor_id = "c6.2u4g.cluster"
+}
+```
+
+### Query list of kafka flavors that available in the availability zone list
+
+```hcl
+variable "az1" {}
+variable "az2" {}
+
+data "flexibleengine_dms_kafka_flavors" "test" {
+ availability_zones = [
+ var.az1,
+ var.az2,
+ ]
+}
+```
+
+## Argument Reference
+
+* `region` - (Optional, String) Specifies the region in which to obtain the dms kafka flavors.
+ If omitted, the provider-level region will be used.
+
+* `flavor_id` - (Optional, String) Specifies the DMS flvaor ID, e.g. **c6.2u4g.cluster**.
+
+* `storage_spec_code` - (Optional, String) Specifies the disk IO encoding.
+ + **dms.physical.storage.high.v2**: Type of the disk that uses high I/O.
+ + **dms.physical.storage.ultra.v2**: Type of the disk that uses ultra-high I/O.
+
+* `type` - (Optional, String) Specifies flavor type. The valid values are **single** and **cluster**.
+
+* `arch_type` - (Optional, String) Specifies the type of CPU architecture, e.g. **X86**.
+
+* `availability_zones` - (Optional, List) Specifies the list of availability zones with available resources.
+
+* `charging_mode` - (Optional, String) Specifies the flavor billing mode.
+ The valid valus are **prePaid** and **postPaid**.
+
+## Attribute Reference
+
+In addition to all arguments above, the following attributes are exported:
+
+* `id` - The data source ID.
+
+* `versions` - The supported flavor versions.
+
+* `flavors` - The list of flavor details.
+ The [flavors](#dms_kafka_flavors) structure is documented below.
+
+
+The `flavors` block supports:
+
+* `id` - The flavor ID.
+
+* `type` - The flavor type.
+
+* `vm_specification` - The underlying VM specification.
+
+* `arch_types` - The list of supported CPU architectures.
+
+* `charging_modes` - The list of supported billing modes.
+
+* `ios` - The list of supported disk IO types.
+ The [object](#dms_kafka_flavor_ios) structure is documented below.
+
+* `support_features` - The list of features supported by the current specification.
+ The [object](#dms_kafka_flavor_support_features) structure is documented below.
+
+* `properties` - The properties of the current specification.
+ The [object](#dms_kafka_flavor_properties) structure is documented below.
+
+
+The `ios` block supports:
+
+* `storage_spec_code` - The disk IO encoding.
+
+* `type` - The disk type.
+
+* `availability_zones` - The list of availability zones with available resources.
+
+* `unavailability_zones` - The list of unavailability zones with available resources.
+
+
+The `support_features` block supports:
+
+* `name` - The function name, e.g. **connector_obs**.
+
+* `properties` - The function property details.
+ The [object](#dms_kafka_flavor_support_feature_properties) structure is documented below.
+
+
+The `properties` block supports:
+
+* `max_task` - The maximum number of tasks for the dump function.
+
+* `min_task` - The minimum number of tasks for the dump function.
+
+* `max_node` - The maximum number of nodes for the dump function.
+
+* `min_node` - The minimum number of nodes for the dump function.
+
+
+The `properties` block supports:
+
+* `max_broker` - The maximum number of brokers.
+
+* `min_broker` - The minimum number of brokers.
+
+* `max_bandwidth_per_broker` - The maximum bandwidth per broker.
+
+* `max_consumer_per_broker` - The maximum number of consumers per broker.
+
+* `max_partition_per_broker` - The maximum number of partitions per broker.
+
+* `max_tps_per_broker` - The maximum TPS per broker.
+
+* `max_storage_per_node` - The maximum storage per node. The unit is GB.
+
+* `min_storage_per_node` - The minimum storage per node. The unit is GB.
+
+* `flavor_alias` - The flavor ID alias.
diff --git a/docs/data-sources/dms_maintainwindow.md b/docs/data-sources/dms_maintainwindow.md
new file mode 100644
index 000000000..2d37cbe97
--- /dev/null
+++ b/docs/data-sources/dms_maintainwindow.md
@@ -0,0 +1,34 @@
+---
+subcategory: "Distributed Message Service (DMS)"
+---
+
+# flexibleengine_dms_maintainwindow
+
+Use this data source to get the ID of an available FlexibleEngine dms maintainwindow.
+
+## Example Usage
+
+```hcl
+data "flexibleengine_dms_maintainwindow" "maintainwindow1" {
+ seq = 1
+}
+```
+
+## Argument Reference
+
+* `region` - (Optional, String) The region in which to obtain the dms maintainwindows. If omitted, the provider-level
+ region will be used.
+
+* `seq` - (Optional, Int) Indicates the sequential number of a maintenance time window.
+
+* `begin` - (Optional, String) Indicates the time at which a maintenance time window starts.
+
+* `end` - (Optional, String) Indicates the time at which a maintenance time window ends.
+
+* `default` - (Optional, Bool) Indicates whether a maintenance time window is set to the default time segment.
+
+## Attribute Reference
+
+In addition to all arguments above, the following attributes are exported:
+
+* `id` - Specifies a data source ID in UUID format.
diff --git a/flexibleengine/acceptance/acceptance.go b/flexibleengine/acceptance/acceptance.go
index 265460d83..0ba0d8792 100644
--- a/flexibleengine/acceptance/acceptance.go
+++ b/flexibleengine/acceptance/acceptance.go
@@ -15,12 +15,13 @@ import (
)
var (
- OS_DEPRECATED_ENVIRONMENT = os.Getenv("OS_DEPRECATED_ENVIRONMENT")
- OS_AVAILABILITY_ZONE = os.Getenv("OS_AVAILABILITY_ZONE")
- OS_REGION_NAME = os.Getenv("OS_REGION_NAME")
- OS_ACCESS_KEY = os.Getenv("OS_ACCESS_KEY")
- OS_SECRET_KEY = os.Getenv("OS_SECRET_KEY")
- OS_PROJECT_ID = os.Getenv("OS_PROJECT_ID")
+ OS_DEPRECATED_ENVIRONMENT = os.Getenv("OS_DEPRECATED_ENVIRONMENT")
+ OS_AVAILABILITY_ZONE = os.Getenv("OS_AVAILABILITY_ZONE")
+ OS_REGION_NAME = os.Getenv("OS_REGION_NAME")
+ OS_ACCESS_KEY = os.Getenv("OS_ACCESS_KEY")
+ OS_SECRET_KEY = os.Getenv("OS_SECRET_KEY")
+ OS_PROJECT_ID = os.Getenv("OS_PROJECT_ID")
+ OS_ENTERPRISE_PROJECT_ID_TEST = os.Getenv("OS_ENTERPRISE_PROJECT_ID_TEST")
OS_VPC_ID = os.Getenv("OS_VPC_ID")
OS_NETWORK_ID = os.Getenv("OS_NETWORK_ID")
diff --git a/flexibleengine/acceptance/data_source_flexibleengine_dms_kafka_flavors_test.go b/flexibleengine/acceptance/data_source_flexibleengine_dms_kafka_flavors_test.go
new file mode 100644
index 000000000..70fc7c6c4
--- /dev/null
+++ b/flexibleengine/acceptance/data_source_flexibleengine_dms_kafka_flavors_test.go
@@ -0,0 +1,74 @@
+package acceptance
+
+import (
+ "regexp"
+ "testing"
+
+ "github.com/hashicorp/terraform-plugin-sdk/v2/helper/resource"
+
+ "github.com/huaweicloud/terraform-provider-huaweicloud/huaweicloud/services/acceptance"
+)
+
+func TestAccKafkaFlavorsDataSource_basic(t *testing.T) {
+ dataSourceName := "data.flexibleengine_dms_kafka_flavors.test"
+ dc := acceptance.InitDataSourceCheck(dataSourceName)
+
+ resource.ParallelTest(t, resource.TestCase{
+ PreCheck: func() { testAccPreCheck(t) },
+ ProviderFactories: TestAccProviderFactories,
+ Steps: []resource.TestStep{
+ {
+ Config: testAccKafkaFlavorsDataSource_basic,
+ Check: resource.ComposeTestCheckFunc(
+ dc.CheckResourceExists(),
+ resource.TestMatchResourceAttr(dataSourceName, "versions.#", regexp.MustCompile(`[1-9]\d*`)),
+ resource.TestMatchResourceAttr(dataSourceName, "flavors.#", regexp.MustCompile(`[1-9]\d*`)),
+ resource.TestCheckOutput("type_validation", "true"),
+ resource.TestCheckOutput("arch_types_validation", "true"),
+ resource.TestCheckOutput("charging_modes_validation", "true"),
+ resource.TestCheckOutput("storage_spec_code_validation", "true"),
+ resource.TestCheckOutput("availability_zones_validation", "true"),
+ ),
+ },
+ },
+ })
+}
+
+const testAccKafkaFlavorsDataSource_basic = `
+data "flexibleengine_dms_kafka_flavors" "baisc" {
+ type = "cluster"
+}
+
+data "flexibleengine_dms_kafka_flavors" "test" {
+ type = local.test_refer.type
+ arch_type = local.test_refer.arch_types[0]
+ charging_mode = local.test_refer.charging_modes[0]
+ storage_spec_code = local.test_refer.ios[0].storage_spec_code
+ availability_zones = local.test_refer.ios[0].availability_zones
+}
+
+locals {
+ test_refer = data.flexibleengine_dms_kafka_flavors.baisc.flavors[0]
+ test_results = data.flexibleengine_dms_kafka_flavors.test
+}
+
+output "type_validation" {
+ value = contains(local.test_results.flavors[*].type, local.test_refer.type)
+}
+
+output "arch_types_validation" {
+ value = !contains([for a in local.test_results.flavors[*].arch_types : contains(a, local.test_refer.arch_types[0])], false)
+}
+
+output "charging_modes_validation" {
+ value = !contains([for c in local.test_results.flavors[*].charging_modes : contains(c, local.test_refer.charging_modes[0])], false)
+}
+
+output "storage_spec_code_validation" {
+ value = !contains([for ios in local.test_results.flavors[*].ios : !contains([for io in ios : io.storage_spec_code == local.test_refer.ios[0].storage_spec_code], false)], false)
+}
+
+output "availability_zones_validation" {
+ value = !contains([for ios in local.test_results.flavors[*].ios : !contains([for io in ios : length(setintersection(io.availability_zones, local.test_refer.ios[0].availability_zones)) == length(local.test_refer.ios[0].availability_zones)], false)], false)
+}
+`
diff --git a/flexibleengine/acceptance/data_source_flexibleengine_dms_maintainwindow_test.go b/flexibleengine/acceptance/data_source_flexibleengine_dms_maintainwindow_test.go
new file mode 100644
index 000000000..bd5e37d95
--- /dev/null
+++ b/flexibleengine/acceptance/data_source_flexibleengine_dms_maintainwindow_test.go
@@ -0,0 +1,35 @@
+package acceptance
+
+import (
+ "testing"
+
+ "github.com/hashicorp/terraform-plugin-sdk/v2/helper/resource"
+
+ "github.com/huaweicloud/terraform-provider-huaweicloud/huaweicloud/services/acceptance"
+)
+
+func TestAccDmsMaintainWindowDataSource_basic(t *testing.T) {
+ dataSourceName := "data.flexibleengine_dms_maintainwindow.maintainwindow1"
+ dc := acceptance.InitDataSourceCheck(dataSourceName)
+
+ resource.ParallelTest(t, resource.TestCase{
+ PreCheck: func() { testAccPreCheck(t) },
+ ProviderFactories: TestAccProviderFactories,
+ Steps: []resource.TestStep{
+ {
+ Config: testAccDmsMaintainWindowDataSource_basic,
+ Check: resource.ComposeTestCheckFunc(
+ dc.CheckResourceExists(),
+ resource.TestCheckResourceAttr(dataSourceName, "seq", "1"),
+ resource.TestCheckResourceAttr(dataSourceName, "begin", "22:00:00"),
+ ),
+ },
+ },
+ })
+}
+
+var testAccDmsMaintainWindowDataSource_basic = `
+data "flexibleengine_dms_maintainwindow" "maintainwindow1" {
+ seq = 1
+}
+`
diff --git a/flexibleengine/provider.go b/flexibleengine/provider.go
index 37f75ca24..6975ff8c8 100644
--- a/flexibleengine/provider.go
+++ b/flexibleengine/provider.go
@@ -299,8 +299,10 @@ func Provider() *schema.Provider {
"flexibleengine_ddm_accounts": ddm.DataSourceDdmAccounts(),
"flexibleengine_dms_kafka_instances": dms.DataSourceDmsKafkaInstances(),
+ "flexibleengine_dms_kafka_flavors": dms.DataSourceKafkaFlavors(),
"flexibleengine_dms_rocketmq_broker": dms.DataSourceDmsRocketMQBroker(),
"flexibleengine_dms_rocketmq_instances": dms.DataSourceDmsRocketMQInstances(),
+ "flexibleengine_dms_maintainwindow": dms.DataSourceDmsMaintainWindow(),
"flexibleengine_dws_flavors": dws.DataSourceDwsFlavors(),