diff --git a/eng/jacoco-test-coverage/pom.xml b/eng/jacoco-test-coverage/pom.xml
index 2d71ee05fad60..936013fbc564a 100644
--- a/eng/jacoco-test-coverage/pom.xml
+++ b/eng/jacoco-test-coverage/pom.xml
@@ -64,22 +64,22 @@
com.azure
azure-communication-common
- 1.0.2
+ 1.1.0-beta.1
com.azure
azure-communication-identity
- 1.1.1
+ 1.2.0-beta.1
com.azure
azure-communication-sms
- 1.0.3
+ 1.1.0-beta.1
com.azure
azure-communication-phonenumbers
- 1.0.3
+ 1.1.0-beta.1
com.azure
@@ -194,7 +194,7 @@
com.azure
azure-identity
- 1.3.1
+ 1.4.0-beta.1
com.azure
diff --git a/eng/versioning/version_client.txt b/eng/versioning/version_client.txt
index e4149a025d605..0bf8475ede39b 100644
--- a/eng/versioning/version_client.txt
+++ b/eng/versioning/version_client.txt
@@ -51,12 +51,11 @@ com.azure:azure-analytics-synapse-artifacts;1.0.0-beta.4;1.0.0-beta.5
com.azure:azure-analytics-synapse-spark;1.0.0-beta.3;1.0.0-beta.4
com.azure:azure-analytics-synapse-managedprivateendpoints;1.0.0-beta.3;1.0.0-beta.4
com.azure:azure-analytics-synapse-monitoring;1.0.0-beta.3;1.0.0-beta.4
-com.azure:azure-communication-callingserver;1.0.0;1.0.0-beta.1
com.azure:azure-communication-chat;1.0.0;1.1.0-beta.1
-com.azure:azure-communication-common;1.0.2;1.0.2
-com.azure:azure-communication-sms;1.0.1;1.0.3
-com.azure:azure-communication-identity;1.1.0;1.1.1
-com.azure:azure-communication-phonenumbers;1.0.1;1.0.3
+com.azure:azure-communication-common;1.0.2;1.1.0-beta.1
+com.azure:azure-communication-sms;1.0.3;1.1.0-beta.1
+com.azure:azure-communication-identity;1.1.1;1.2.0-beta.1
+com.azure:azure-communication-phonenumbers;1.0.3;1.1.0-beta.1
com.azure:azure-containers-containerregistry;1.0.0-beta.2;1.0.0-beta.3
com.azure:azure-core;1.17.0;1.18.0-beta.1
com.azure:azure-core-amqp;2.0.6;2.1.0-beta.1
diff --git a/sdk/cosmos/azure-cosmos-spark_3-1_2-12/src/test/scala/com/azure/cosmos/spark/SparkE2EQueryITest.scala b/sdk/cosmos/azure-cosmos-spark_3-1_2-12/src/test/scala/com/azure/cosmos/spark/SparkE2EQueryITest.scala
index 8b18c09306eea..a814e5e0e4f80 100644
--- a/sdk/cosmos/azure-cosmos-spark_3-1_2-12/src/test/scala/com/azure/cosmos/spark/SparkE2EQueryITest.scala
+++ b/sdk/cosmos/azure-cosmos-spark_3-1_2-12/src/test/scala/com/azure/cosmos/spark/SparkE2EQueryITest.scala
@@ -524,57 +524,6 @@ class SparkE2EQueryITest
}
}
- "spark query" can "when forceNullableProperties is false and rows have different schema" in {
- val cosmosEndpoint = TestConfigurations.HOST
- val cosmosMasterKey = TestConfigurations.MASTER_KEY
- val samplingSize = 100
- val expectedResults = samplingSize * 2
- val container = cosmosClient.getDatabase(cosmosDatabase).getContainer(cosmosContainer)
-
- // Inserting documents with slightly different schema
- for( _ <- 1 to expectedResults) {
- val objectNode = Utils.getSimpleObjectMapper.createObjectNode()
- val arr = objectNode.putArray("object_array")
- val nested = Utils.getSimpleObjectMapper.createObjectNode()
- nested.put("A", "test")
- nested.put("B", "test")
- arr.add(nested)
- objectNode.put("id", UUID.randomUUID().toString)
- container.createItem(objectNode).block()
- }
-
- for( _ <- 1 to samplingSize) {
- val objectNode2 = Utils.getSimpleObjectMapper.createObjectNode()
- val arr = objectNode2.putArray("object_array")
- val nested = Utils.getSimpleObjectMapper.createObjectNode()
- nested.put("A", "test")
- arr.add(nested)
- objectNode2.put("id", UUID.randomUUID().toString)
- container.createItem(objectNode2).block()
- }
-
- val cfgWithInference = Map("spark.cosmos.accountEndpoint" -> cosmosEndpoint,
- "spark.cosmos.accountKey" -> cosmosMasterKey,
- "spark.cosmos.database" -> cosmosDatabase,
- "spark.cosmos.container" -> cosmosContainer,
- "spark.cosmos.read.inferSchema.enabled" -> "true",
- "spark.cosmos.read.inferSchema.forceNullableProperties" -> "false",
- "spark.cosmos.read.inferSchema.samplingSize" -> samplingSize.toString,
- "spark.cosmos.read.inferSchema.query" -> "SELECT * FROM c ORDER BY c._ts",
- "spark.cosmos.read.partitioning.strategy" -> "Restrictive"
- )
-
- val dfWithInference = spark.read.format("cosmos.oltp").options(cfgWithInference).load()
- try {
- dfWithInference.collect()
- fail("Should have thrown an exception")
- }
- catch {
- case inner: Exception =>
- inner.toString.contains("The 1th field 'B' of input row cannot be null") shouldBe true
- }
- }
-
"spark query" can "use custom sampling size" in {
val cosmosEndpoint = TestConfigurations.HOST
val cosmosMasterKey = TestConfigurations.MASTER_KEY