Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@
/**
* Serializable ByteBuffer.
*
* <p>By default, ByteBuffer instances are not serializable, tihs class implemented
* <p>By default, ByteBuffer instances are not serializable, this class implemented
* serializable wrapper for byte buffer to communicate serialize buffer content.
*/
public class SerializableByteBuffer implements Serializable {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -360,7 +360,7 @@ class SparkRowCodec(sliceSchemas: Array[StructType]) {
case BooleanType =>
nm.MakeConstNode(row.getBoolean(idx))
case StringType => {
// generally safe, native ConstNode copyed the string
// generally safe, native ConstNode copied the string
val str = row.getString(idx)
nm.MakeConstNode(str)
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ object GroupByAggregationPlan {
// Check if we should keep the index column
val keepIndexColumn = SparkInstance.keepIndexColumn(ctx, node.GetNodeId())

// Get parition keys
// Get partition keys
val groupByExprs = node.getGroup_.keys()
val groupByCols = mutable.ArrayBuffer[Column]()
val groupIdxs = mutable.ArrayBuffer[Int]()
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -84,7 +84,7 @@ object GraphvizUtil {
case _ => opType.toString
}

// No need to use cache because graphviz will merge ndoe with the same name
// No need to use cache because graphviz will merge node with the same name
val nodeNameWithId = "[%s]%s".format(node.GetNodeId().toString, readableNodeName)
var graphNode = mutNode(nodeNameWithId)

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ object SparkUtil {
logger.info("Has order by column and do not support native last join")
false
} else if (joinType != JoinType.kJoinTypeLast && joinType != JoinType.kJoinTypeConcat) {
logger.info("Join type is neighter last join or concat join and do not support native last join")
logger.info("Join type is neither last join or concat join and do not support native last join")
false
} else {
try {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@ class WindowComputer(config: WindowAggConfig, jit: HybridSeJitWrapper, keepIndex
// append slices cnt = needAppendInput ? inputSchemaSlices.size : 0
private val appendSlices = if (config.needAppendInput) config.inputSchemaSlices.length else 0

// group key comparation
// group key comparison
private var groupKeyComparator = HybridseUtil.createGroupKeyComparator(config.groupIdxs)

private var unsafeGroupKeyComparator = HybridseUtil.createUnsafeGroupKeyComparator(
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ import org.apache.spark.sql.types.{DoubleType, IntegerType, StringType, StructFi

class TestWindowSkewOptAndParallelization extends SparkTestSuite {

test("Test end2end window skew optimization and window paralleization with union") {
test("Test end2end window skew optimization and window parallelization with union") {

getSparkSession.conf.set("spark.openmldb.window.parallelization", true)
getSparkSession.conf.set("spark.openmldb.window.skew.opt", true)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ public Connection connect(String url, Properties info) throws SQLException {
if (info == null) {
info = new Properties();
}
// just url missmatch, don't throw exception
// just url mismatch, don't throw exception
if (!parseAndMergeClusterProps(url, info)) {
return null;
}
Expand All @@ -62,7 +62,7 @@ public Connection connect(String url, Properties info) throws SQLException {
}

/**
* parse and verification of URL. If url is just not acceptted, return false.
* parse and verification of URL. If url is just not accepted, return false.
*
* <p>basic syntax :<br>
* {@code
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,7 @@ public java.sql.ResultSet get() throws InterruptedException, ExecutionException

/**
*
* @param l current timeout set by executeQeuryAsyn, so the param is invalid
* @param l current timeout set by executeQueryAsyn, so the param is invalid
* @param timeUnit the time unit of the timeout, which is also invalid.
* @return the result of the query from the database
* @throws InterruptedException throws when a thread is waiting, sleeping, or otherwise occupied, and the thread is interrupted, either before or during the activity.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -242,7 +242,7 @@ public synchronized void closeTunnel(int tid) {
// miniCluster.stopWithSavepoint(id, null, true); // can't?
miniCluster.cancelJob(id); // checkpoint is retain on cancellation, but it's better to wait a while?

// task record count? == source count or wahtever?
// task record count? == source count or whatever?
}

/** Use first field for buckets. */
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -101,7 +101,7 @@ protected void initZnode() {
}

/**
* Override this mothod to deal with events for leader election.
* Override this method to deal with events for leader election.
*
* @param event the ZooKeeper event
*/
Expand Down
Loading