Skip to content

Commit aa29521

Browse files
authored
Fix typo "tranpose" (#124929)
1 parent 27e01d1 commit aa29521

File tree

16 files changed

+27
-27
lines changed

16 files changed

+27
-27
lines changed

clang/lib/Headers/amxtf32transposeintrin.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@
88
*/
99
#ifndef __IMMINTRIN_H
1010
#error \
11-
"Never use <amxtf32tranposeintrin.h> directly; include <immintrin.h> instead."
11+
"Never use <amxtf32transposeintrin.h> directly; include <immintrin.h> instead."
1212
#endif // __IMMINTRIN_H
1313

1414
#ifndef __AMX_TF32TRANSPOSEINTRIN_H

llvm/lib/Target/Hexagon/HexagonISelDAGToDAGHVX.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2269,7 +2269,7 @@ OpRef HvxSelector::perfect(ShuffleMask SM, OpRef Va, ResultStack &Results) {
22692269
// For example, with the inputs as above, the result will be:
22702270
// 0 8 2 A 4 C 6 E
22712271
// 1 9 3 B 5 D 7 F
2272-
// Now, this result can be tranposed again, but with the group size of 2:
2272+
// Now, this result can be transposed again, but with the group size of 2:
22732273
// 08 19 4C 5D
22742274
// 2A 3B 6E 7F
22752275
// If we then transpose that result, but with the group size of 4, we get:

mlir/docs/Canonicalization.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@ For example, a pattern that transform
7171
outs(%init1 : tensor<2x1x3xf32>)
7272
dimensions = [1, 0, 2]
7373
%out = linalg.transpose
74-
ins(%tranpose: tensor<2x1x3xf32>)
74+
ins(%transpose: tensor<2x1x3xf32>)
7575
outs(%init2 : tensor<3x1x2xf32>)
7676
permutation = [2, 1, 0]
7777
```

mlir/include/mlir/Dialect/Linalg/TransformOps/LinalgTransformOps.td

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1007,7 +1007,7 @@ def PackTransposeOp : Op<Transform_Dialect, "structured.pack_transpose", [
10071007

10081008
This operation may produce a silenceableFailure if the transpose spec is
10091009
ill-formed (i.e. `outer_perm` or `inner_perm` are not permutations of the
1010-
proper rank) or if the tranposition of all involved operations fails for any
1010+
proper rank) or if the transposition of all involved operations fails for any
10111011
reason.
10121012

10131013
This operation returns 3 handles, one to the transformed LinalgOp, one to

mlir/include/mlir/Dialect/Vector/IR/VectorOps.td

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2779,7 +2779,7 @@ def Vector_MatmulOp : Vector_Op<"matrix_multiply", [Pure,
27792779
"`:` `(` type($lhs) `,` type($rhs) `)` `->` type($res)";
27802780
}
27812781

2782-
/// Vector dialect matrix tranposition op that operates on flattened 1-D
2782+
/// Vector dialect matrix transposition op that operates on flattened 1-D
27832783
/// MLIR vectors. This is the counterpart of llvm.matrix.transpose in MLIR.
27842784
/// This may seem redundant with vector.transpose but it serves the purposes of
27852785
/// more progressive lowering and localized type conversion on the path:
@@ -2799,7 +2799,7 @@ def Vector_FlatTransposeOp : Vector_Op<"flat_transpose", [Pure,
27992799
let description = [{
28002800
This is the counterpart of llvm.matrix.transpose in MLIR. It serves
28012801
the purposes of more progressive lowering and localized type conversion.
2802-
Higher levels typically lower matrix tranpositions into 'vector.transpose'
2802+
Higher levels typically lower matrix transpositions into 'vector.transpose'
28032803
operations. Subsequent rewriting rule progressively lower these operations
28042804
into 'vector.flat_transpose' operations to bring the operations closer
28052805
to the hardware ISA.

mlir/lib/Conversion/VectorToArmSME/VectorToArmSME.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -332,7 +332,7 @@ struct SplatOpToArmSMELowering : public OpRewritePattern<vector::SplatOp> {
332332
/// %transposed_src = arm_sme.tile_load %alloca[%c0, %c0]
333333
/// layout<vertical> : memref<?x?xi32>, vector<[4]x[4]xi32>
334334
///
335-
/// NOTE: Tranposing via memory is obviously expensive, the current intention
335+
/// NOTE: Transposing via memory is obviously expensive, the current intention
336336
/// is to avoid the transpose if possible, this is therefore intended as a
337337
/// fallback and to provide base support for Vector ops. If it turns out
338338
/// transposes can't be avoided then this should be replaced with a more optimal

mlir/lib/Conversion/VectorToXeGPU/VectorToXeGPU.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -179,7 +179,7 @@ struct TransferReadLowering : public OpRewritePattern<vector::TransferReadOp> {
179179
if (isTransposeLoad &&
180180
elementType.getIntOrFloatBitWidth() < minTransposeBitWidth)
181181
return rewriter.notifyMatchFailure(
182-
readOp, "Unsupported data type for tranposition");
182+
readOp, "Unsupported data type for transposition");
183183

184184
// If load is transposed, get the base shape for the tensor descriptor.
185185
SmallVector<int64_t> descShape(vecTy.getShape());

mlir/lib/Dialect/ArmSME/Transforms/VectorLegalization.cpp

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -304,7 +304,7 @@ struct LegalizeTransferReadOpsByDecomposition
304304
kMatchFailureNonPermutationMap);
305305

306306
// Note: For 2D vector types the only non-identity permutation is a simple
307-
// tranpose [1, 0].
307+
// transpose [1, 0].
308308
bool transposed = !permutationMap.isIdentity();
309309

310310
auto loc = readOp.getLoc();
@@ -352,7 +352,7 @@ struct LegalizeTransferWriteOpsByDecomposition
352352
kMatchFailureNonPermutationMap);
353353

354354
// Note: For 2D vector types the only non-identity permutation is a simple
355-
// tranpose [1, 0].
355+
// transpose [1, 0].
356356
bool transposed = !permutationMap.isIdentity();
357357

358358
auto loc = writeOp.getLoc();

mlir/lib/Dialect/LLVMIR/IR/NVVMDialect.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -935,7 +935,7 @@ LogicalResult NVVM::WgmmaMmaAsyncOp::verify() {
935935
// Check transpose (only available for f16/bf16)
936936
// Matrices A should be stored in row-major and B in column-major.
937937
// Only f16/bf16 matrices can be stored in either column-major or row-major
938-
// by setting the tranpose value(imm-trans-a,imm-trans-b) in PTX code.
938+
// by setting the transpose value(imm-trans-a,imm-trans-b) in PTX code.
939939
if ((typeA != WGMMATypes::f16 && typeA != WGMMATypes::bf16) &&
940940
(getLayoutA() == mlir::NVVM::MMALayout::col ||
941941
getLayoutB() == mlir::NVVM::MMALayout::row)) {

mlir/lib/Dialect/Linalg/IR/LinalgInterfaces.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -147,7 +147,7 @@ linalg::isaBroadcastOpInterface(GenericOp op) {
147147
}
148148

149149
//===----------------------------------------------------------------------===//
150-
// TranposeOpInterface implementation
150+
// TransposeOpInterface implementation
151151
//===----------------------------------------------------------------------===//
152152
std::optional<SmallVector<int64_t>>
153153
linalg::isaTransposeOpInterface(GenericOp op) {

mlir/lib/Dialect/Linalg/Transforms/TransposeConv2D.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -138,7 +138,7 @@ FailureOr<Operation *> transposeConv2D(RewriterBase &rewriter,
138138
linalg::Conv2DNhwcHwcfQOp>(rewriter, op);
139139
}
140140

141-
void populateTranposeConv2DPatterns(RewritePatternSet &patterns) {
141+
void populateTransposeConv2DPatterns(RewritePatternSet &patterns) {
142142
MLIRContext *context = patterns.getContext();
143143
patterns.insert<
144144
ConvConverter<linalg::Conv2DNhwcFhwcOp, linalg::Conv2DNhwcHwcfOp>,

mlir/lib/Dialect/SparseTensor/Transforms/SparseGPUCodegen.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1269,7 +1269,7 @@ struct LinalgOpRewriter : public OpRewritePattern<linalg::GenericOp> {
12691269
AffineExpr i, j, k;
12701270
bindDims(getContext(), i, j, k);
12711271

1272-
// TODO: more robust patterns, tranposed versions, more kernels,
1272+
// TODO: more robust patterns, transposed versions, more kernels,
12731273
// identify alpha and beta and pass them to the CUDA calls.
12741274

12751275
// Recognize a SpMV kernel.

mlir/lib/Dialect/Vector/IR/VectorOps.cpp

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1488,7 +1488,7 @@ class ExtractFromInsertTransposeChainState {
14881488

14891489
/// Try to fold in place to extract(source, extractPosition) and return the
14901490
/// folded result. Return null if folding is not possible (e.g. due to an
1491-
/// internal tranposition in the result).
1491+
/// internal transposition in the result).
14921492
Value tryToFoldExtractOpInPlace(Value source);
14931493

14941494
ExtractOp extractOp;
@@ -1582,7 +1582,7 @@ ExtractFromInsertTransposeChainState::handleInsertOpWithPrefixPos(Value &res) {
15821582

15831583
/// Try to fold in place to extract(source, extractPosition) and return the
15841584
/// folded result. Return null if folding is not possible (e.g. due to an
1585-
/// internal tranposition in the result).
1585+
/// internal transposition in the result).
15861586
Value ExtractFromInsertTransposeChainState::tryToFoldExtractOpInPlace(
15871587
Value source) {
15881588
// TODO: Canonicalization for dynamic position not implemented yet.

mlir/lib/Dialect/Vector/Transforms/VectorDropLeadUnitDim.cpp

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -377,18 +377,18 @@ mlir::vector::castAwayContractionLeadingOneDim(vector::ContractionOp contractOp,
377377
int64_t orginalZeroDim = it.value().getDimPosition(0);
378378
if (orginalZeroDim != dimToDrop) {
379379
// There are two reasons to be in this path, 1. We need to
380-
// tranpose the operand to make the dim to be dropped
380+
// transpose the operand to make the dim to be dropped
381381
// leading. 2. The dim to be dropped does not exist and in
382-
// that case we dont want to add a unit tranpose but we must
382+
// that case we dont want to add a unit transpose but we must
383383
// check all the indices to make sure this is the case.
384-
bool tranposeNeeded = false;
384+
bool transposeNeeded = false;
385385
SmallVector<int64_t> perm;
386386
SmallVector<AffineExpr> transposeResults;
387387

388388
for (int64_t i = 0, e = map.getNumResults(); i < e; ++i) {
389389
int64_t currDim = map.getDimPosition(i);
390390
if (currDim == dimToDrop) {
391-
tranposeNeeded = true;
391+
transposeNeeded = true;
392392
perm.insert(perm.begin(), i);
393393
auto targetExpr = rewriter.getAffineDimExpr(currDim);
394394
transposeResults.insert(transposeResults.begin(), targetExpr);
@@ -413,9 +413,9 @@ mlir::vector::castAwayContractionLeadingOneDim(vector::ContractionOp contractOp,
413413
}
414414
}
415415

416-
// Do the tranpose now if needed so that we can drop the
416+
// Do the transpose now if needed so that we can drop the
417417
// correct dim using extract later.
418-
if (tranposeNeeded) {
418+
if (transposeNeeded) {
419419
map = AffineMap::get(map.getNumDims(), 0, transposeResults,
420420
contractOp.getContext());
421421
if (transposeNonOuterUnitDims) {
@@ -474,7 +474,7 @@ namespace {
474474

475475
/// Turns vector.contract on vector with leading 1 dimensions into
476476
/// vector.extract followed by vector.contract on vector without leading
477-
/// 1 dimensions. Also performs tranpose of lhs and rhs operands if required
477+
/// 1 dimensions. Also performs transpose of lhs and rhs operands if required
478478
/// prior to extract.
479479
struct CastAwayContractionLeadingOneDim
480480
: public MaskableOpRewritePattern<vector::ContractionOp> {

mlir/lib/Dialect/Vector/Transforms/VectorTransforms.cpp

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1792,11 +1792,11 @@ struct DropUnitDimsFromTransposeOp final
17921792
auto dropDimsShapeCast = rewriter.create<vector::ShapeCastOp>(
17931793
loc, sourceTypeWithoutUnitDims, op.getVector());
17941794
// Create the new transpose.
1795-
auto tranposeWithoutUnitDims =
1795+
auto transposeWithoutUnitDims =
17961796
rewriter.create<vector::TransposeOp>(loc, dropDimsShapeCast, newPerm);
17971797
// Restore the unit dims via shape cast.
17981798
rewriter.replaceOpWithNewOp<vector::ShapeCastOp>(
1799-
op, op.getResultVectorType(), tranposeWithoutUnitDims);
1799+
op, op.getResultVectorType(), transposeWithoutUnitDims);
18001800

18011801
return success();
18021802
}

mlir/test/Dialect/Vector/vector-unroll-options.mlir

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -232,11 +232,11 @@ func.func @vector_reduction(%v : vector<8xf32>) -> f32 {
232232
// CHECK: %[[add3:.*]] = arith.addf %[[add2]], %[[r3]]
233233
// CHECK: return %[[add3]]
234234

235-
func.func @vector_tranpose(%v : vector<2x4x3x8xf32>) -> vector<2x3x8x4xf32> {
235+
func.func @vector_transpose(%v : vector<2x4x3x8xf32>) -> vector<2x3x8x4xf32> {
236236
%t = vector.transpose %v, [0, 2, 3, 1] : vector<2x4x3x8xf32> to vector<2x3x8x4xf32>
237237
return %t : vector<2x3x8x4xf32>
238238
}
239-
// CHECK-LABEL: func @vector_tranpose
239+
// CHECK-LABEL: func @vector_transpose
240240
// CHECK: %[[VI:.*]] = arith.constant dense<0.000000e+00> : vector<2x3x8x4xf32>
241241
// CHECK: %[[E0:.*]] = vector.extract_strided_slice %{{.*}} {offsets = [0, 0, 0, 0], sizes = [1, 2, 3, 4], strides = [1, 1, 1, 1]} : vector<2x4x3x8xf32> to vector<1x2x3x4xf32>
242242
// CHECK: %[[T0:.*]] = vector.transpose %[[E0]], [0, 2, 3, 1] : vector<1x2x3x4xf32> to vector<1x3x4x2xf32>

0 commit comments

Comments
 (0)