diff --git a/.github/workflows/build_main.yml b/.github/workflows/build_main.yml index 1ac6c87b7d041..9ef52f326375b 100644 --- a/.github/workflows/build_main.yml +++ b/.github/workflows/build_main.yml @@ -17,27 +17,11 @@ # under the License. # -<<<<<<<< HEAD:.github/workflows/build_and_test_ansi.yml -name: "Build and test (ANSI)" -======== name: "Build" ->>>>>>>> 17a8e67a6a03fd5a33f4ed078f8325665a0635aa:.github/workflows/build_main.yml on: push: branches: -<<<<<<<< HEAD:.github/workflows/build_and_test_ansi.yml - - branch-3.3 - -jobs: - call-build-and-test: - name: Call main build - uses: ./.github/workflows/build_and_test.yml - if: github.repository == 'apache/spark' - with: - ansi_enabled: true - -======== - '**' jobs: @@ -46,4 +30,3 @@ jobs: packages: write name: Run uses: ./.github/workflows/build_and_test.yml ->>>>>>>> 17a8e67a6a03fd5a33f4ed078f8325665a0635aa:.github/workflows/build_main.yml diff --git a/core/src/main/scala/org/apache/spark/status/protobuf/Utils.scala b/core/src/main/scala/org/apache/spark/status/protobuf/Utils.scala index 0f45ca3599758..bfd6eb9f2e0fc 100644 --- a/core/src/main/scala/org/apache/spark/status/protobuf/Utils.scala +++ b/core/src/main/scala/org/apache/spark/status/protobuf/Utils.scala @@ -14,23 +14,7 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -package org.apache.spark.deploy.k8s.integrationtest -<<<<<<<< HEAD:resource-managers/kubernetes/integration-tests/src/test/scala/org/apache/spark/deploy/k8s/integrationtest/YuniKornSuite.scala -@YuniKornTag -class YuniKornSuite extends KubernetesSuite { - - override protected def setUpTest(): Unit = { - super.setUpTest() - val namespace = sparkAppConf.get("spark.kubernetes.namespace") - sparkAppConf - .set("spark.kubernetes.scheduler.name", "yunikorn") - .set("spark.kubernetes.driver.label.queue", "root." + namespace) - .set("spark.kubernetes.executor.label.queue", "root." + namespace) - .set("spark.kubernetes.driver.annotation.yunikorn.apache.org/app-id", "{{APP_ID}}") - .set("spark.kubernetes.executor.annotation.yunikorn.apache.org/app-id", "{{APP_ID}}") - } -======== package org.apache.spark.status.protobuf import java.util.{Map => JMap} @@ -59,5 +43,4 @@ private[protobuf] object Utils { putAllFunc(input) } } ->>>>>>>> 17a8e67a6a03fd5a33f4ed078f8325665a0635aa:core/src/main/scala/org/apache/spark/status/protobuf/Utils.scala } diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/SupportsPushDownOffset.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/SupportsPushDownOffset.java index 8888b9ae4fc84..ffa2cad371594 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/SupportsPushDownOffset.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/SupportsPushDownOffset.java @@ -15,14 +15,6 @@ * limitations under the License. */ -<<<<<<<< HEAD:sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/RowLevelOperationBuilder.java -package org.apache.spark.sql.connector.write; - -import org.apache.spark.annotation.Experimental; - -/** - * An interface for building a {@link RowLevelOperation}. -======== package org.apache.spark.sql.connector.read; import org.apache.spark.annotation.Evolving; @@ -31,19 +23,9 @@ * A mix-in interface for {@link ScanBuilder}. Data sources can implement this interface to * push down OFFSET. We can push down OFFSET with many other operations if they follow the * operator order we defined in {@link ScanBuilder}'s class doc. ->>>>>>>> 17a8e67a6a03fd5a33f4ed078f8325665a0635aa:sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/SupportsPushDownOffset.java * * @since 3.4.0 */ -<<<<<<<< HEAD:sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/RowLevelOperationBuilder.java -@Experimental -public interface RowLevelOperationBuilder { - /** - * Returns a {@link RowLevelOperation} that controls how Spark rewrites data - * for DELETE, UPDATE, MERGE commands. - */ - RowLevelOperation build(); -======== @Evolving public interface SupportsPushDownOffset extends ScanBuilder { @@ -51,5 +33,4 @@ public interface SupportsPushDownOffset extends ScanBuilder { * Pushes down OFFSET to the data source. */ boolean pushOffset(int offset); ->>>>>>>> 17a8e67a6a03fd5a33f4ed078f8325665a0635aa:sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/SupportsPushDownOffset.java } diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/colstats/HistogramBin.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/colstats/HistogramBin.java index b5a2e9cf34cc8..50bb00bd3cb49 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/colstats/HistogramBin.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/colstats/HistogramBin.java @@ -15,39 +15,15 @@ * limitations under the License. */ -<<<<<<<< HEAD:sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/RowLevelOperationInfo.java -package org.apache.spark.sql.connector.write; - -import org.apache.spark.annotation.Experimental; -import org.apache.spark.sql.connector.write.RowLevelOperation.Command; -import org.apache.spark.sql.util.CaseInsensitiveStringMap; - -/** - * An interface with logical information for a row-level operation such as DELETE, UPDATE, MERGE. -======== package org.apache.spark.sql.connector.read.colstats; import org.apache.spark.annotation.Evolving; /** * An interface to represent a bin in an equi-height histogram. ->>>>>>>> 17a8e67a6a03fd5a33f4ed078f8325665a0635aa:sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/colstats/HistogramBin.java * * @since 3.4.0 */ -<<<<<<<< HEAD:sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/RowLevelOperationInfo.java -@Experimental -public interface RowLevelOperationInfo { - /** - * Returns options that the user specified when performing the row-level operation. - */ - CaseInsensitiveStringMap options(); - - /** - * Returns the row-level SQL command (e.g. DELETE, UPDATE, MERGE). - */ - Command command(); -======== @Evolving public interface HistogramBin { /** @@ -64,5 +40,4 @@ public interface HistogramBin { * @return approximate number of distinct values in this bin */ long ndv(); ->>>>>>>> 17a8e67a6a03fd5a33f4ed078f8325665a0635aa:sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/colstats/HistogramBin.java } diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/SupportsDelta.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/SupportsDelta.java index 613fbb82c0487..6315b65f610f4 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/SupportsDelta.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/SupportsDelta.java @@ -15,17 +15,6 @@ * limitations under the License. */ -<<<<<<<< HEAD:sql/catalyst/src/main/java/org/apache/spark/sql/connector/expressions/Cast.java -package org.apache.spark.sql.connector.expressions; - -import java.io.Serializable; - -import org.apache.spark.annotation.Evolving; -import org.apache.spark.sql.types.DataType; - -/** - * Represents a cast expression in the public logical expression API. -======== package org.apache.spark.sql.connector.write; import org.apache.spark.annotation.Experimental; @@ -34,27 +23,9 @@ /** * A mix-in interface for {@link RowLevelOperation}. Data sources can implement this interface * to indicate they support handling deltas of rows. ->>>>>>>> 17a8e67a6a03fd5a33f4ed078f8325665a0635aa:sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/SupportsDelta.java * * @since 3.4.0 */ -<<<<<<<< HEAD:sql/catalyst/src/main/java/org/apache/spark/sql/connector/expressions/Cast.java -@Evolving -public class Cast implements Expression, Serializable { - private Expression expression; - private DataType dataType; - - public Cast(Expression expression, DataType dataType) { - this.expression = expression; - this.dataType = dataType; - } - - public Expression expression() { return expression; } - public DataType dataType() { return dataType; } - - @Override - public Expression[] children() { return new Expression[]{ expression() }; } -======== @Experimental public interface SupportsDelta extends RowLevelOperation { @Override @@ -64,5 +35,4 @@ public interface SupportsDelta extends RowLevelOperation { * Returns the row ID column references that should be used for row equality. */ NamedReference[] rowId(); ->>>>>>>> 17a8e67a6a03fd5a33f4ed078f8325665a0635aa:sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/SupportsDelta.java }