Skip to content

Commit

Permalink
formatting
Browse files Browse the repository at this point in the history
  • Loading branch information
kahmed10 committed Aug 11, 2023
1 parent 225cd3a commit 7aee638
Show file tree
Hide file tree
Showing 3 changed files with 18 additions and 19 deletions.
2 changes: 1 addition & 1 deletion src/eliminate_contiguous.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -166,7 +166,7 @@ static void remove_contiguous_noops(const std::string& op_name, module& m)
{
for(auto ins : iterator_for(m))
{
if (ins->name() != op_name)
if(ins->name() != op_name)
continue;
if(ins->inputs().front()->get_shape() != ins->get_shape())
continue;
Expand Down
10 changes: 5 additions & 5 deletions test/auto_contiguous_test.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -149,8 +149,8 @@ TEST_CASE(two_transpose_gather)
auto ctd = m2.add_instruction(migraphx::make_op("contiguous"), td);
auto sd = m2.add_instruction(migraphx::make_op("softmax", {{"axis", 2}}), ctd);
auto csd = m2.add_instruction(migraphx::make_op("contiguous"), sd);
auto bd =
m2.add_instruction(migraphx::make_op("transpose", {{"permutation", {0, 3, 1, 2}}}), csd);
auto bd = m2.add_instruction(
migraphx::make_op("transpose", {{"permutation", {0, 3, 1, 2}}}), csd);
auto cbd = m2.add_instruction(migraphx::make_op("contiguous"), bd);
auto r = m2.add_instruction(migraphx::make_op("gather", {{"axis", 2}}), cbd, ind);
auto cr = m2.add_instruction(migraphx::make_op("contiguous"), r);
Expand All @@ -177,9 +177,9 @@ TEST_CASE(standard_reshape)
auto add = m2.add_instruction(migraphx::make_op("add"), data, data);
auto ca = m2.add_instruction(migraphx::make_op("contiguous"), add);
// extra contiguous coming from reshape logic which has "requires_std_shape" attribute
auto cb = m2.add_instruction(migraphx::make_op("contiguous"), ca);
auto r = m2.add_instruction(migraphx::make_op("reshape", {{"dims", {2, 1, 12, 5}}}), cb);
auto cr = m2.add_instruction(migraphx::make_op("contiguous"), r);
auto cb = m2.add_instruction(migraphx::make_op("contiguous"), ca);
auto r = m2.add_instruction(migraphx::make_op("reshape", {{"dims", {2, 1, 12, 5}}}), cb);
auto cr = m2.add_instruction(migraphx::make_op("contiguous"), r);
m2.add_return({cr});
}

Expand Down
25 changes: 12 additions & 13 deletions test/gpu/pack_int8_args.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -108,7 +108,7 @@ TEST_CASE(quant_dot)
migraphx::make_op("multibroadcast", {{"out_lens", m3_shape.lens()}}), beta);
auto mul_alloc = m.add_instruction(
migraphx::make_op("hip::allocate", {{"shape", migraphx::to_value(m3_shape)}}));
auto m3_beta = m.add_instruction(make_precompile_op("mul"), l3, beta_broadcast, mul_alloc);
auto m3_beta = m.add_instruction(make_precompile_op("mul"), l3, beta_broadcast, mul_alloc);
auto gemm_add = m.add_instruction(make_precompile_op("add"), gemm, m3_beta, output);
m.add_return({gemm_add});

Expand Down Expand Up @@ -179,8 +179,8 @@ TEST_CASE(quant_dot_trans)
auto tl1_alpha_int32 =
m.add_instruction(make_precompile_op("mul"), alpha_broadcast, tl1_convert, mul_alloc);
// convert mul_res to int8
auto tl1_alpha_int8_alloc = m.add_instruction(migraphx::make_op(
"hip::allocate", {{"shape", migraphx::to_value(ts1)}}));
auto tl1_alpha_int8_alloc = m.add_instruction(
migraphx::make_op("hip::allocate", {{"shape", migraphx::to_value(ts1)}}));
auto tl1_alpha_int8 =
m.add_instruction(make_precompile_op(migraphx::make_op(
"convert", {{"target_type", tl1->get_shape().type()}})),
Expand Down Expand Up @@ -291,7 +291,7 @@ TEST_CASE(quant_dot_pad)
m.add_instruction(migraphx::make_op("multibroadcast", {{"out_lens", s3.lens()}}), beta);
auto mul_alloc = m.add_instruction(
migraphx::make_op("hip::allocate", {{"shape", migraphx::to_value(s3)}}));
auto m3_beta = m.add_instruction(make_precompile_op("mul"), l3, beta_broadcast, mul_alloc);
auto m3_beta = m.add_instruction(make_precompile_op("mul"), l3, beta_broadcast, mul_alloc);
auto gemm_add = m.add_instruction(make_precompile_op("add"), gemm, m3_beta, output);
m.add_return({gemm_add});
return m;
Expand Down Expand Up @@ -345,14 +345,14 @@ TEST_CASE(quant_dot_trans_pad)
auto tl2 =
m.add_instruction(migraphx::make_op("transpose", {{"permutation", {0, 1, 3, 2}}}), l2);
migraphx::shape ts2{migraphx::shape::int8_type, {3, 2, 9, 7}};

migraphx::instruction_ref ptb{};
if(int8_x4)
{
ptb = m.add_instruction(
migraphx::make_op("hip::allocate", {{"shape", migraphx::to_value(ps2)}}));
}
auto pb = tl2;
auto pb = tl2;
if(int8_x4)
{
pb = m.add_instruction(
Expand Down Expand Up @@ -381,8 +381,8 @@ TEST_CASE(quant_dot_trans_pad)
auto tl1_alpha_int32 =
m.add_instruction(make_precompile_op("mul"), alpha_broadcast, tl1_convert, mul_alloc);
// convert mul_res to int8
auto tl1_alpha_int8_alloc = m.add_instruction(migraphx::make_op(
"hip::allocate", {{"shape", migraphx::to_value(ts1)}}));
auto tl1_alpha_int8_alloc = m.add_instruction(
migraphx::make_op("hip::allocate", {{"shape", migraphx::to_value(ts1)}}));

migraphx::instruction_ref pta{};
if(int8_x4)
Expand All @@ -391,11 +391,10 @@ TEST_CASE(quant_dot_trans_pad)
migraphx::make_op("hip::allocate", {{"shape", migraphx::to_value(ps1)}}));
}

auto tl1_alpha_int8 =
m.add_instruction(make_precompile_op(migraphx::make_op(
"convert", {{"target_type", ts1.type()}})),
tl1_alpha_int32,
tl1_alpha_int8_alloc);
auto tl1_alpha_int8 = m.add_instruction(
make_precompile_op(migraphx::make_op("convert", {{"target_type", ts1.type()}})),
tl1_alpha_int32,
tl1_alpha_int8_alloc);

auto pa = tl1_alpha_int8;
if(int8_x4)
Expand Down

0 comments on commit 7aee638

Please sign in to comment.