/* * Copyright (c) 2021, 2022, Oracle and/or its affiliates. All rights reserved. * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. * * This code is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License version 2 only, as * published by the Free Software Foundation. * * This code is distributed in the hope that it will be useful, but WITHOUT * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License * version 2 for more details (a copy is included in the LICENSE file that * accompanied this code). * * You should have received a copy of the GNU General Public License version * 2 along with this work; if not, write to the Free Software Foundation, * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. * * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA * or visit www.oracle.com if you need additional information or have any * questions. *
*/
// PPC implementation uses VSX load/store instructions (if // SuperwordUseVSX) which support 4 byte but not arbitrary alignment staticconstbool misaligned_vectors_ok() { returnfalse;
}
// Whether code generation need accurate ConvI2L types. staticconstbool convi2l_type_required = true;
// Do we need to mask the count passed to shift instructions or does // the cpu only look at the lower 5/6 bits anyway? // PowerPC requires masked shift counts. staticconstbool need_masked_shift_count = true;
// Power6 requires postalloc expand (see block.cpp for description of postalloc expand). staticconstbool require_postalloc_expand = true;
// No support for generic vector operands. staticconstbool supports_generic_vector_operands = false;
static constexpr bool isSimpleConstant64(jlong value) { // Probably always true, even if a temp register is required. returntrue;
}
// Use conditional move (CMOVL) on Power7. static constexpr int long_cmove_cost() { return 0; } // this only makes long cmoves more expensive than int cmoves
// Suppress CMOVF. Conditional move available (sort of) on PPC64 only from P7 onwards. Not exploited yet. // fsel doesn't accept a condition register as input, so this would be slightly different. staticint float_cmove_cost() { return ConditionalMoveLimit; }
// This affects two different things: // - how Decode nodes are matched // - how ImplicitNullCheck opportunities are recognized // If true, the matcher will try to remove all Decodes and match them // (as operands) into nodes. NullChecks are not prepared to deal with // Decodes by final_graph_reshaping(). // If false, final_graph_reshaping() forces the decode behind the Cmp // for a NullCheck. The matcher matches the Decode node into a register. // Implicit_null_check optimization moves the Decode along with the // memory operation back up before the NullCheck. static constexpr bool narrow_oop_use_complex_address() { // TODO: PPC port if (MatchDecodeNodes) return true; returnfalse;
}
staticbool narrow_klass_use_complex_address() {
NOT_LP64(ShouldNotCallThis());
assert(UseCompressedClassPointers, "only for compressed klass code"); // TODO: PPC port if (MatchDecodeNodes) return true; returnfalse;
}
staticbool const_oop_prefer_decode() { // Prefer ConN+DecodeN over ConP in simple compressed oops mode. return CompressedOops::base() == NULL;
}
staticbool const_klass_prefer_decode() { // Prefer ConNKlass+DecodeNKlass over ConP in simple compressed klass mode. return CompressedKlassPointers::base() == NULL;
}
// Is it better to copy float constants, or load them directly from memory? // Intel can load a float constant from a direct address, requiring no // extra registers. Most RISCs will have to materialize an address into a // register first, so they would do better to copy the constant from stack. staticconstbool rematerialize_float_constants = false;
// If CPU can load and store mis-aligned doubles directly then no fixup is // needed. Else we split the double into 2 integer pieces and move it // piece-by-piece. Only happens when passing doubles into C code as the // Java calling convention forces doubles to be aligned. staticconstbool misaligned_doubles_ok = true;
// Advertise here if the CPU requires explicit rounding operations to implement strictfp mode. staticconstbool strict_fp_requires_explicit_rounding = false;
// Do floats take an entire double register or just half? // // A float occupies a ppc64 double register. For the allocator, a // ppc64 double register appears as a pair of float registers. static constexpr bool float_in_double() { returntrue; }
// Do ints take an entire long register or just half? // The relevant question is how the int is callee-saved: // the whole long is written but de-opt'ing will have to extract // the relevant 32 bits. staticconstbool int_in_long = true;
// Does the CPU supports vector variable shift instructions? static constexpr bool supports_vector_variable_shifts(void) { returnfalse;
}
// Does the CPU supports vector variable rotate instructions? static constexpr bool supports_vector_variable_rotates(void) { returnfalse;
}
// Does the CPU supports vector constant rotate instructions? static constexpr bool supports_vector_constant_rotates(int shift) { returnfalse;
}
// Does the CPU supports vector unsigned comparison instructions? static constexpr bool supports_vector_comparison_unsigned(int vlen, BasicType bt) { returnfalse;
}
// Some microarchitectures have mask registers used on vectors static constexpr bool has_predicated_vectors(void) { returnfalse;
}
// true means we have fast l2f conversion // false means that conversion is done by runtime call staticconstbool convL2FSupported(void) { // fcfids can do the conversion (>= Power7). // fcfid + frsp showed rounding problem when result should be 0x3f800001. return VM_Version::has_fcfids();
}
// Implements a variant of EncodeISOArrayNode that encode ASCII only staticconstbool supports_encode_ascii_array = true;
// Returns pre-selection estimated size of a vector operation. staticint vector_op_pre_select_sz_estimate(int vopc, BasicType ety, int vlen) { switch(vopc) { default: return 0; case Op_RoundVF: // fall through case Op_RoundVD: { return 30;
}
}
} // Returns pre-selection estimated size of a scalar operation. staticint scalar_op_pre_select_sz_estimate(int vopc, BasicType ety) { switch(vopc) { default: return 0; case Op_RoundF: // fall through case Op_RoundD: { return 30;
}
}
}
#endif// CPU_PPC_MATCHER_PPC_HPP
¤ Dauer der Verarbeitung: 0.15 Sekunden
(vorverarbeitet)
¤
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung ist noch experimentell.