Skip to content

Commit 0af1249

Browse files
authored
Fix typos shoulde should (#71350)
1 parent 971869a commit 0af1249

File tree

11 files changed

+16
-15
lines changed

11 files changed

+16
-15
lines changed

paddle/cinn/hlir/framework/pir/trivial_op_util.cc

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1072,7 +1072,7 @@ void CheckLoopAlignment(const std::vector<ir::Expr>& roots) {
10721072
}
10731073
PADDLE_ENFORCE(fusion::VectorEqual(base_loop_vars, loop_vars, var_equal),
10741074
::common::errors::PreconditionNotMet(
1075-
"CheckLoopAlignment Failed, The loop vars are not euqal "
1075+
"CheckLoopAlignment Failed, The loop vars are not equal "
10761076
"between FusionOps: \n%s\n%s",
10771077
roots[base_loop_idx],
10781078
roots[i]));
@@ -1087,7 +1087,7 @@ void CheckLoopAlignment(const std::vector<ir::Expr>& roots) {
10871087
PADDLE_ENFORCE(
10881088
fusion::VectorEqual(base_reduce_vars, reduce_vars, var_equal),
10891089
::common::errors::PreconditionNotMet(
1090-
"CheckLoopAlignment Failed, The reduce vars are not euqal "
1090+
"CheckLoopAlignment Failed, The reduce vars are not equal "
10911091
"between FusionOps: \n%s\n%s",
10921092
roots[base_reduce_idx],
10931093
roots[i]));

paddle/phi/core/distributed/auto_parallel/placement_types.cc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -87,7 +87,7 @@ phi::distributed::Placements cvt_dim_map_to_placements(
8787
auto& p = placements[mesh_id];
8888
if (p->is_shard()) {
8989
PADDLE_THROW(common::errors::PreconditionNotMet(
90-
"ProcessMesh dimension cann't be mapped to two dimension of the "
90+
"ProcessMesh dimension can't be mapped to two dimension of the "
9191
"same tensor: {%d} and {%d}",
9292
i,
9393
dynamic_cast<phi::distributed::Shard&>(*p).get_dim()));

paddle/phi/kernels/funcs/blas/blaslt_gemm_search.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -168,7 +168,7 @@ class CublasLtAlgoCache {
168168
cudaDataType_t bias_type,
169169
cudaDataType_t c_type,
170170
cudaStream_t stream) {
171-
// If we don't have config file and we donot search, here return nullptr
171+
// If we don't have config file and we do not search, here return nullptr
172172
if (!has_config_file_ && search_times_ <= 0) {
173173
return nullptr;
174174
}

paddle/phi/kernels/funcs/distribution_helper.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -311,7 +311,7 @@ void distribution_and_transform(const GPUContext &ctx,
311311
size_t total_thread = block_size * grid_size;
312312
size_t curand4_loop_times =
313313
(size + 4 * total_thread - 1) / (4 * total_thread);
314-
// 'increment' shoulde be multiple of 4
314+
// 'increment' should be multiple of 4
315315
uint64_t increment = curand4_loop_times * 4;
316316

317317
auto seed_offset = gen_cuda->IncrementOffset(increment);

paddle/phi/kernels/funcs/jit/gen/seqpool.h

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -131,8 +131,8 @@ class SeqPoolJitCode : public JitCode {
131131
rest_used_num_regs,
132132
common::errors::InvalidArgument(
133133
"All heights of SeqPool should use the same number of registers."
134-
"It equals to the numbr of rest registers. But use %d registers "
135-
"and the numbr of rest registers is %d.",
134+
"It equals to the number of rest registers. But use %d registers "
135+
"and the number of rest registers is %d.",
136136
reg_idx,
137137
rest_used_num_regs));
138138
for (int i = 0; i < reg_idx; ++i) {

paddle/phi/kernels/funcs/stack_and_unstack.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -255,7 +255,7 @@ void UnStackRawKernel(const Context& ctx,
255255
std::vector<DenseTensor*>* outs) {
256256
auto x_dims = x.dims();
257257

258-
// Input tensor is splited to split_dim tensors along split_dim dimension.
258+
// Input tensor is split to split_dim tensors along split_dim dimension.
259259
int64_t split_dim = x_dims[axis];
260260

261261
// zero sized tensor case

paddle/phi/kernels/fusion/cutlass/memory_efficient_attention/iterators/epilogue_predicated_tile_iterator.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,7 @@ namespace threadblock {
8686
/// Satisfies: ReadableTileIterator | PredicatedTileIterator |
8787
/// ForwardTileIterator
8888
///
89-
template <typename ThreadMap_, ///< Thread map (conept: OutputTileThreadMap)
89+
template <typename ThreadMap_, ///< Thread map (concept: OutputTileThreadMap)
9090
typename Element_, ///< Element data type
9191
bool ScatterD = false, ///< Scatter D operand or not
9292
bool UseCUDAStore = false>

paddle/phi/kernels/fusion/gpu/masked_multihead_attention_kernel.cu

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1086,7 +1086,7 @@ void DispatchWithDtype(const Context &dev_ctx,
10861086
params.inv_sqrt_dh = inv_sqrt_dh;
10871087
params.rotary_emb_dims = rotary_emb_dims;
10881088

1089-
params.steps_per_block = timestep; // if not SPLIT, this is unuseful.
1089+
params.steps_per_block = timestep; // if not SPLIT, this is useless.
10901090
params.split_seq = 1; // if not SPLIT, grid.x==1
10911091

10921092
bool SPLIT = false;

paddle/phi/kernels/gpu/lars_momentum_kernel.cu

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -154,7 +154,8 @@ __device__ inline void VectorizeLarsUpdate(const T* __restrict__ grad,
154154
--rdc=true compile flag, then L2_norm kernel can be set with __device__ and
155155
cooperative_groups::grid_group also can be involved. Otherwise, adding this
156156
flag may affect much, L2_norm kernel shall be set with __global__.*/
157-
// TODO(limingshu): declaration of cooperative_groups wapper is invalid in host.
157+
// TODO(limingshu): declaration of cooperative_groups wrapper is invalid in
158+
// host.
158159
template <typename T, typename MT>
159160
__forceinline__ __device__ void L2NormKernel(
160161
const cooperative_groups::grid_group* cg,
@@ -193,7 +194,7 @@ __global__ void L2NormKernel(
193194
g_buffer[blockIdx.x] = g_tmp;
194195
}
195196
#if CUDA_VERSION >= 11000
196-
cg->sync(); // Grid sync for writring partial result to global memory
197+
cg->sync(); // Grid sync for writing partial result to global memory
197198
MT p_part_sum = threadIdx.x < gridDim.x ? p_buffer[threadIdx.x] : 0;
198199
MT g_part_sum = threadIdx.x < gridDim.x ? g_buffer[threadIdx.x] : 0;
199200
MT tmp0 = phi::funcs::BlockReduceSum<MT>(p_part_sum, FINAL_MASK);

paddle/phi/kernels/gpu/multinomial_kernel.cu

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -244,7 +244,7 @@ void MultinomialKernel(const Context& dev_ctx,
244244
auto gen_cuda = dev_ctx.GetGenerator();
245245
size_t curand4_loop_times =
246246
(num_distributions + 4 * grid_y - 1) / (4 * grid_y);
247-
// 'increment' shoulde be multiple of 4
247+
// 'increment' should be multiple of 4
248248
uint64_t increment = curand4_loop_times * 4;
249249
auto seed_offset = gen_cuda->IncrementOffset(increment);
250250

0 commit comments

Comments
 (0)