Lines Matching refs:MB
165 constexpr uint32_t kSmallChunkSize = 2 * MB; in Create2MethodsWithGap()
472 constexpr uint32_t just_over_max_negative_disp = 16 * MB + 2 - 4u /* PC adjustment */; in TEST_F()
508 constexpr uint32_t max_positive_disp = 16 * MB - 2u + 4u /* PC adjustment */; in TEST_F()
530 constexpr uint32_t max_negative_disp = 16 * MB - 4u /* PC adjustment */; in TEST_F()
556 constexpr uint32_t just_over_max_positive_disp = 16 * MB + 4u /* PC adjustment */; in TEST_F()
585 ASSERT_GE(diff, 16 * MB - (1u << 22)); // Simple encoding, unknown bits fit into imm10:imm11:0. in TEST_F()
601 constexpr uint32_t just_over_max_negative_disp = 16 * MB + 2 - 4u /* PC adjustment */; in TEST_F()
970 1 * MB - (kLiteralOffset2 + kPcAdjustment) in TEST_F()
989 EXPECT_EQ(2 * MB, last_method_offset - first_method_offset); in TEST_F()
1070 1 * MB - (kReachableFromOffset2 + kPcAdjustment) in TEST_F()
1090 EXPECT_EQ(2 * MB, last_method_offset - first_method_offset); in TEST_F()
1292 code.reserve(1 * MB); in TEST_F()
1293 const size_t num_patches = 1 * MB / 8u; in TEST_F()
1304 ASSERT_EQ(1 * MB, code.size()); in TEST_F()
1312 DCHECK_ALIGNED(1 * MB, kArmAlignment); in TEST_F()
1316 PushBackInsn(&expected_code, BneWWithOffset(8u * i + 4u, 1 * MB)); in TEST_F()
1346 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmAlignment) in TEST_F()
1354 filler_size = 1 * MB - sizeof(OatQuickMethodHeader); in TEST_F()
1393 ASSERT_EQ(15 * MB, GetMethodOffset(method_idx) - GetMethodOffset(1u)); in TEST_F()