diff --git a/eng/jacoco-test-coverage/pom.xml b/eng/jacoco-test-coverage/pom.xml index 2d71ee05fad60..936013fbc564a 100644 --- a/eng/jacoco-test-coverage/pom.xml +++ b/eng/jacoco-test-coverage/pom.xml @@ -64,22 +64,22 @@ com.azure azure-communication-common - 1.0.2 + 1.1.0-beta.1 com.azure azure-communication-identity - 1.1.1 + 1.2.0-beta.1 com.azure azure-communication-sms - 1.0.3 + 1.1.0-beta.1 com.azure azure-communication-phonenumbers - 1.0.3 + 1.1.0-beta.1 com.azure @@ -194,7 +194,7 @@ com.azure azure-identity - 1.3.1 + 1.4.0-beta.1 com.azure diff --git a/eng/versioning/version_client.txt b/eng/versioning/version_client.txt index e4149a025d605..0bf8475ede39b 100644 --- a/eng/versioning/version_client.txt +++ b/eng/versioning/version_client.txt @@ -51,12 +51,11 @@ com.azure:azure-analytics-synapse-artifacts;1.0.0-beta.4;1.0.0-beta.5 com.azure:azure-analytics-synapse-spark;1.0.0-beta.3;1.0.0-beta.4 com.azure:azure-analytics-synapse-managedprivateendpoints;1.0.0-beta.3;1.0.0-beta.4 com.azure:azure-analytics-synapse-monitoring;1.0.0-beta.3;1.0.0-beta.4 -com.azure:azure-communication-callingserver;1.0.0;1.0.0-beta.1 com.azure:azure-communication-chat;1.0.0;1.1.0-beta.1 -com.azure:azure-communication-common;1.0.2;1.0.2 -com.azure:azure-communication-sms;1.0.1;1.0.3 -com.azure:azure-communication-identity;1.1.0;1.1.1 -com.azure:azure-communication-phonenumbers;1.0.1;1.0.3 +com.azure:azure-communication-common;1.0.2;1.1.0-beta.1 +com.azure:azure-communication-sms;1.0.3;1.1.0-beta.1 +com.azure:azure-communication-identity;1.1.1;1.2.0-beta.1 +com.azure:azure-communication-phonenumbers;1.0.3;1.1.0-beta.1 com.azure:azure-containers-containerregistry;1.0.0-beta.2;1.0.0-beta.3 com.azure:azure-core;1.17.0;1.18.0-beta.1 com.azure:azure-core-amqp;2.0.6;2.1.0-beta.1 diff --git a/sdk/cosmos/azure-cosmos-spark_3-1_2-12/src/test/scala/com/azure/cosmos/spark/SparkE2EQueryITest.scala b/sdk/cosmos/azure-cosmos-spark_3-1_2-12/src/test/scala/com/azure/cosmos/spark/SparkE2EQueryITest.scala index 8b18c09306eea..a814e5e0e4f80 100644 --- a/sdk/cosmos/azure-cosmos-spark_3-1_2-12/src/test/scala/com/azure/cosmos/spark/SparkE2EQueryITest.scala +++ b/sdk/cosmos/azure-cosmos-spark_3-1_2-12/src/test/scala/com/azure/cosmos/spark/SparkE2EQueryITest.scala @@ -524,57 +524,6 @@ class SparkE2EQueryITest } } - "spark query" can "when forceNullableProperties is false and rows have different schema" in { - val cosmosEndpoint = TestConfigurations.HOST - val cosmosMasterKey = TestConfigurations.MASTER_KEY - val samplingSize = 100 - val expectedResults = samplingSize * 2 - val container = cosmosClient.getDatabase(cosmosDatabase).getContainer(cosmosContainer) - - // Inserting documents with slightly different schema - for( _ <- 1 to expectedResults) { - val objectNode = Utils.getSimpleObjectMapper.createObjectNode() - val arr = objectNode.putArray("object_array") - val nested = Utils.getSimpleObjectMapper.createObjectNode() - nested.put("A", "test") - nested.put("B", "test") - arr.add(nested) - objectNode.put("id", UUID.randomUUID().toString) - container.createItem(objectNode).block() - } - - for( _ <- 1 to samplingSize) { - val objectNode2 = Utils.getSimpleObjectMapper.createObjectNode() - val arr = objectNode2.putArray("object_array") - val nested = Utils.getSimpleObjectMapper.createObjectNode() - nested.put("A", "test") - arr.add(nested) - objectNode2.put("id", UUID.randomUUID().toString) - container.createItem(objectNode2).block() - } - - val cfgWithInference = Map("spark.cosmos.accountEndpoint" -> cosmosEndpoint, - "spark.cosmos.accountKey" -> cosmosMasterKey, - "spark.cosmos.database" -> cosmosDatabase, - "spark.cosmos.container" -> cosmosContainer, - "spark.cosmos.read.inferSchema.enabled" -> "true", - "spark.cosmos.read.inferSchema.forceNullableProperties" -> "false", - "spark.cosmos.read.inferSchema.samplingSize" -> samplingSize.toString, - "spark.cosmos.read.inferSchema.query" -> "SELECT * FROM c ORDER BY c._ts", - "spark.cosmos.read.partitioning.strategy" -> "Restrictive" - ) - - val dfWithInference = spark.read.format("cosmos.oltp").options(cfgWithInference).load() - try { - dfWithInference.collect() - fail("Should have thrown an exception") - } - catch { - case inner: Exception => - inner.toString.contains("The 1th field 'B' of input row cannot be null") shouldBe true - } - } - "spark query" can "use custom sampling size" in { val cosmosEndpoint = TestConfigurations.HOST val cosmosMasterKey = TestConfigurations.MASTER_KEY