@@ -222,8 +222,8 @@ struct CollapseShapeOpInterface
222
222
MemRefType::get (collapseShapeOp.getSrcType ().getShape (),
223
223
collapseShapeOp.getSrcType ().getElementType (),
224
224
AffineMap (), bufferType.getMemorySpace ());
225
- buffer = rewriter. create < bufferization::ToBufferOp>(
226
- op-> getLoc (), memrefType, *tensorAlloc);
225
+ buffer = bufferization::ToBufferOp::create (rewriter, op-> getLoc (),
226
+ memrefType, *tensorAlloc);
227
227
}
228
228
229
229
// Result type is inferred by the builder.
@@ -349,8 +349,8 @@ struct ExpandShapeOpInterface
349
349
if (failed (buffer))
350
350
return failure ();
351
351
352
- auto memrefExpandShape = rewriter. create < memref::ExpandShapeOp> (
353
- op->getLoc (), tensorResultType.getShape (), *buffer,
352
+ auto memrefExpandShape = memref::ExpandShapeOp::create (
353
+ rewriter, op->getLoc (), tensorResultType.getShape (), *buffer,
354
354
expandShapeOp.getReassociationIndices (),
355
355
expandShapeOp.getMixedOutputShape ());
356
356
replaceOpWithBufferizedValues (rewriter, op,
@@ -398,8 +398,8 @@ struct ExtractSliceOpInterface
398
398
extractSliceOp.getResult (), options, state);
399
399
if (failed (resultMemrefType))
400
400
return failure ();
401
- Value subView = rewriter. create < memref::SubViewOp> (
402
- loc, llvm::cast<MemRefType>(*resultMemrefType), *srcMemref,
401
+ Value subView = memref::SubViewOp::create (
402
+ rewriter, loc, llvm::cast<MemRefType>(*resultMemrefType), *srcMemref,
403
403
mixedOffsets, mixedSizes, mixedStrides);
404
404
405
405
replaceOpWithBufferizedValues (rewriter, op, subView);
@@ -469,7 +469,7 @@ static void createStores(RewriterBase &rewriter, Location loc, int dim,
469
469
if (dim == static_cast <int >(shape.size ()) - 1 ) {
470
470
for (int i = 0 ; i < shape.back (); ++i) {
471
471
indices.back () = constants[i];
472
- rewriter. create < memref::StoreOp>( loc, *elementIt, buffer, indices);
472
+ memref::StoreOp::create (rewriter, loc, *elementIt, buffer, indices);
473
473
++elementIt;
474
474
}
475
475
return ;
@@ -507,8 +507,8 @@ struct FromElementsOpInterface
507
507
bufferization::getBufferType (*tensorAlloc, options, state);
508
508
if (failed (memrefType))
509
509
return failure ();
510
- Value buffer = rewriter. create < bufferization::ToBufferOp>(
511
- op-> getLoc (), *memrefType, *tensorAlloc);
510
+ Value buffer = bufferization::ToBufferOp::create (rewriter, op-> getLoc (),
511
+ *memrefType, *tensorAlloc);
512
512
513
513
// Case: tensor<0xelem_type>.
514
514
if (fromElementsOp.getElements ().empty ()) {
@@ -518,8 +518,8 @@ struct FromElementsOpInterface
518
518
519
519
// Case: tensor<elem_type>.
520
520
if (shape.empty ()) {
521
- rewriter. create < memref::StoreOp>(
522
- loc, fromElementsOp.getElements ().front (), buffer);
521
+ memref::StoreOp::create (rewriter, loc,
522
+ fromElementsOp.getElements ().front (), buffer);
523
523
replaceOpWithBufferizedValues (rewriter, op, buffer);
524
524
return success ();
525
525
}
@@ -529,7 +529,7 @@ struct FromElementsOpInterface
529
529
SmallVector<Value, 2 > constants;
530
530
constants.reserve (maxDim);
531
531
for (int i = 0 ; i < maxDim; ++i)
532
- constants.push_back (rewriter. create < arith::ConstantIndexOp>( loc, i));
532
+ constants.push_back (arith::ConstantIndexOp::create (rewriter, loc, i));
533
533
534
534
// Traverse all `elements` and create `memref.store` ops.
535
535
auto elementIt = fromElementsOp.getElements ().begin ();
@@ -576,15 +576,15 @@ static Value lowerGenerateLikeOpBody(RewriterBase &rewriter, Location loc,
576
576
// Create linalg::MapOp.
577
577
OpBuilder::InsertionGuard g (rewriter);
578
578
auto linalgOp =
579
- rewriter. create < linalg::MapOp>( loc, tensorType, /* inputs=*/ ValueRange (),
580
- /* init=*/ tensorDestination);
579
+ linalg::MapOp::create (rewriter, loc, tensorType, /* inputs=*/ ValueRange (),
580
+ /* init=*/ tensorDestination);
581
581
Block &linalgBody = linalgOp.getMapper ().emplaceBlock ();
582
582
583
583
// Create linalg::IndexOps.
584
584
rewriter.setInsertionPointToStart (&linalgBody);
585
585
SmallVector<Value> indices;
586
586
for (int64_t dim = 0 ; dim < tensorType.getRank (); ++dim)
587
- indices.push_back (rewriter. create < linalg::IndexOp>( loc, dim));
587
+ indices.push_back (linalg::IndexOp::create (rewriter, loc, dim));
588
588
589
589
// Move over body.
590
590
rewriter.mergeBlocks (&generateBody.front (), &linalgBody, indices);
@@ -644,8 +644,8 @@ struct InsertOpInterface
644
644
getBuffer (rewriter, insertOp.getDest (), options, state);
645
645
if (failed (destMemref))
646
646
return failure ();
647
- rewriter. create < memref::StoreOp>( insertOp.getLoc (), insertOp.getScalar (),
648
- *destMemref, insertOp.getIndices ());
647
+ memref::StoreOp::create (rewriter, insertOp.getLoc (), insertOp.getScalar (),
648
+ *destMemref, insertOp.getIndices ());
649
649
replaceOpWithBufferizedValues (rewriter, op, *destMemref);
650
650
return success ();
651
651
}
@@ -713,9 +713,9 @@ struct InsertSliceOpInterface
713
713
memref::SubViewOp::inferRankReducedResultType (
714
714
insertSliceOp.getSourceType ().getShape (), dstMemrefType,
715
715
mixedOffsets, mixedSizes, mixedStrides);
716
- Value subView = rewriter. create <memref::SubViewOp>(
717
- loc, subviewMemRefType, *dstMemref, mixedOffsets, mixedSizes ,
718
- mixedStrides);
716
+ Value subView =
717
+ memref::SubViewOp::create (rewriter, loc, subviewMemRefType, *dstMemref,
718
+ mixedOffsets, mixedSizes, mixedStrides);
719
719
720
720
// Copy tensor. If this tensor.insert_slice has a matching
721
721
// tensor.extract_slice, the copy operation will eventually fold away.
@@ -796,14 +796,14 @@ struct PadOpInterface
796
796
for (int64_t i = 0 ; i < resultType.getRank (); ++i) {
797
797
if (!resultType.isDynamicDim (i))
798
798
continue ;
799
- Value srcDim = rewriter. create < tensor::DimOp>( loc, padOp.getSource (), i);
799
+ Value srcDim = tensor::DimOp::create (rewriter, loc, padOp.getSource (), i);
800
800
Value lowPad = toValue (mixedLowPad[i]);
801
801
Value highPad = toValue (mixedHighPad[i]);
802
802
AffineExpr s0, s1, s2;
803
803
bindSymbols (op->getContext (), s0, s1, s2);
804
804
AffineExpr sumExpr = s0 + s1 + s2;
805
- Value sum = rewriter. create < affine::AffineApplyOp> (
806
- loc, sumExpr, ValueRange{srcDim, lowPad, highPad});
805
+ Value sum = affine::AffineApplyOp::create (
806
+ rewriter, loc, sumExpr, ValueRange{srcDim, lowPad, highPad});
807
807
dynamicSizes.push_back (sum);
808
808
}
809
809
@@ -995,9 +995,9 @@ struct ParallelInsertSliceOpInterface
995
995
parallelInsertSliceOp.getMixedOffsets (),
996
996
parallelInsertSliceOp.getMixedSizes (),
997
997
parallelInsertSliceOp.getMixedStrides ());
998
- Value subview = rewriter. create < memref::SubViewOp> (
999
- parallelInsertSliceOp.getLoc (), subviewMemRefType, *destBuffer ,
1000
- parallelInsertSliceOp.getMixedOffsets (),
998
+ Value subview = memref::SubViewOp::create (
999
+ rewriter, parallelInsertSliceOp.getLoc (), subviewMemRefType,
1000
+ *destBuffer, parallelInsertSliceOp.getMixedOffsets (),
1001
1001
parallelInsertSliceOp.getMixedSizes (),
1002
1002
parallelInsertSliceOp.getMixedStrides ());
1003
1003
@@ -1065,14 +1065,14 @@ struct SplatOpInterface
1065
1065
if (options.defaultMemorySpaceFn (tensorType) != Attribute ())
1066
1066
return op->emitError (" memory space not implemented yet" );
1067
1067
1068
- auto linalgOp =
1069
- rewriter. create <linalg::MapOp>(loc, tensorType, /* inputs=*/ ValueRange (),
1070
- /* init=*/ *tensorAlloc);
1068
+ auto linalgOp = linalg::MapOp::create (rewriter, loc, tensorType,
1069
+ /* inputs=*/ ValueRange (),
1070
+ /* init=*/ *tensorAlloc);
1071
1071
Block &linalgBody = linalgOp.getMapper ().emplaceBlock ();
1072
1072
1073
1073
// Create linalg::IndexOps.
1074
1074
rewriter.setInsertionPointToStart (&linalgBody);
1075
- rewriter. create < linalg::YieldOp>( loc, splatOp.getInput ());
1075
+ linalg::YieldOp::create (rewriter, loc, splatOp.getInput ());
1076
1076
rewriter.replaceOp (splatOp, linalgOp.getResult ()[0 ]);
1077
1077
1078
1078
return success ();
@@ -1126,8 +1126,8 @@ struct ConcatOpInterface
1126
1126
MemRefType memrefType =
1127
1127
MemRefType::get (concatOp.getResultType ().getShape (),
1128
1128
concatOp.getResultType ().getElementType (), layout);
1129
- Value dstBuffer = rewriter. create < bufferization::ToBufferOp> (
1130
- op->getLoc (), memrefType, *tensorAlloc);
1129
+ Value dstBuffer = bufferization::ToBufferOp::create (
1130
+ rewriter, op->getLoc (), memrefType, *tensorAlloc);
1131
1131
1132
1132
// Extract the dimension for the concat op
1133
1133
uint64_t concatDim = concatOp.getDim ();
@@ -1142,7 +1142,7 @@ struct ConcatOpInterface
1142
1142
for (const auto &[dimIdx, dimSize] :
1143
1143
llvm::enumerate (tensorType.getShape ())) {
1144
1144
if (dimSize == ShapedType::kDynamic ) {
1145
- auto dimOp = rewriter. create < memref::DimOp>( loc, dstBuffer, dimIdx);
1145
+ auto dimOp = memref::DimOp::create (rewriter, loc, dstBuffer, dimIdx);
1146
1146
sizes.push_back (dimOp.getResult ());
1147
1147
if (dimIdx == concatDim)
1148
1148
dynamicConcatDim = true ;
@@ -1157,7 +1157,7 @@ struct ConcatOpInterface
1157
1157
if (dynamicConcatDim) {
1158
1158
// One or more operands have dynamic size, so we must accumulate the
1159
1159
// offset with arith ops.
1160
- dynamicOffset = rewriter. create < arith::ConstantIndexOp>( loc, 0 );
1160
+ dynamicOffset = arith::ConstantIndexOp::create (rewriter, loc, 0 );
1161
1161
}
1162
1162
1163
1163
for (auto operand : concatOp.getInputs ()) {
@@ -1174,8 +1174,9 @@ struct ConcatOpInterface
1174
1174
1175
1175
if (dynamicConcatDim) {
1176
1176
offsets[concatDim] = dynamicOffset.value ();
1177
- dynamicSize = rewriter.create <memref::DimOp>(loc, *srcBuffer, concatDim)
1178
- .getResult ();
1177
+ dynamicSize =
1178
+ memref::DimOp::create (rewriter, loc, *srcBuffer, concatDim)
1179
+ .getResult ();
1179
1180
sizes[concatDim] = dynamicSize.value ();
1180
1181
} else {
1181
1182
sizes[concatDim] = rewriter.getIndexAttr (operandConcatDimSize);
@@ -1188,16 +1189,16 @@ struct ConcatOpInterface
1188
1189
memref::SubViewOp::inferRankReducedResultType (
1189
1190
operandTensorType.getShape (), dstMemrefType, offsets, sizes,
1190
1191
strides);
1191
- Value subview = rewriter. create < memref::SubViewOp> (
1192
- loc, subviewMemRefType, dstBuffer, offsets, sizes, strides);
1192
+ Value subview = memref::SubViewOp::create (
1193
+ rewriter, loc, subviewMemRefType, dstBuffer, offsets, sizes, strides);
1193
1194
1194
1195
// Copy the source buffer into the destination subview.
1195
1196
if (failed (options.createMemCpy (rewriter, loc, *srcBuffer, subview)))
1196
1197
return failure ();
1197
1198
1198
1199
if (dynamicConcatDim) {
1199
- dynamicOffset = rewriter. create < arith::AddIOp> (
1200
- loc, dynamicOffset.value (), dynamicSize.value ());
1200
+ dynamicOffset = arith::AddIOp::create (
1201
+ rewriter, loc, dynamicOffset.value (), dynamicSize.value ());
1201
1202
} else {
1202
1203
concatDimOffset += operandConcatDimSize;
1203
1204
}
0 commit comments