@@ -49,6 +49,7 @@ extern rtx cris_gen_movem_load (rtx, rtx
extern rtx cris_emit_movem_store (rtx, rtx, int, bool);
extern void cris_expand_pic_call_address (rtx *);
extern void cris_order_for_addsi3 (rtx *, int);
+extern void cris_emit_trap_for_misalignment (rtx);
#endif /* RTX_CODE */
extern void cris_asm_output_label_ref (FILE *, char *);
extern void cris_target_asm_named_section (const char *, unsigned int, tree);
@@ -1922,6 +1922,39 @@ cris_simple_epilogue (void)
return true;
}
+/* Emit checking that MEM is aligned for an access in MODE, failing
+ that, executing a "break 8" (or call to abort, if "break 8" is
+ disabled). */
+
+void
+cris_emit_trap_for_misalignment (rtx mem)
+{
+ rtx addr, reg, ok_label, and, jmp;
+ int natural_alignment;
+ gcc_assert (MEM_P (mem));
+
+ natural_alignment = GET_MODE_SIZE (GET_MODE (mem));
+ addr = XEXP (mem, 0);
+ reg = force_reg (Pmode, addr);
+ ok_label = gen_label_rtx ();
+
+ /* This will yield a btstq without a separate register used, usually -
+ with the exception for PRE hoisting the "and" but not the branch
+ around the trap: see gcc.dg/target/cris/sync-3s.c. */
+ and = gen_rtx_AND (Pmode, reg, GEN_INT (natural_alignment - 1));
+ emit_cmp_and_jump_insns (force_reg (SImode, and), const0_rtx, EQ,
+ NULL_RTX, Pmode, 1, ok_label);
+ jmp = get_last_insn ();
+ gcc_assert (JUMP_P (jmp));
+
+ /* While this isn't mudflap, it is a similar kind of assertion.
+ If PRED_MUDFLAP stops working, use something else or introduce a
+ more suitable assertion predication type. */
+ predict_insn_def (jmp, PRED_MUDFLAP, TAKEN);
+ expand_builtin_trap ();
+ emit_label (ok_label);
+}
+
/* Expand a return insn (just one insn) marked as using SRP or stack
slot depending on parameter ON_STACK. */
@@ -282,15 +282,25 @@ extern int cris_cpu_version;
#define TARGET_CPU_DEFAULT CRIS_CPU_BASE
#endif
-/* Default target_flags if no switches specified. */
+/* Default target_flags if no switches specified.
+ The alignment-by-32 is to make builtin atomic support for v10 and v32
+ work for *-elf for types without specified alignment (like plain
+ "int"). See top comment in sync.md. */
#ifndef TARGET_DEFAULT
# if TARGET_CPU_DEFAULT == 32
# define TARGET_DEFAULT \
(MASK_STACK_ALIGN \
+ MASK_CONST_ALIGN + MASK_DATA_ALIGN \
+ + MASK_ALIGN_BY_32 \
+ MASK_PROLOGUE_EPILOGUE)
-# else /* 10 */
-# define TARGET_DEFAULT \
+# elif TARGET_CPU_DEFAULT == 10
+# define TARGET_DEFAULT \
+ (MASK_SIDE_EFFECT_PREFIXES + MASK_STACK_ALIGN \
+ + MASK_CONST_ALIGN + MASK_DATA_ALIGN \
+ + MASK_ALIGN_BY_32 \
+ + MASK_PROLOGUE_EPILOGUE + MASK_MUL_BUG)
+# else /* 0 */
+# define TARGET_DEFAULT \
(MASK_SIDE_EFFECT_PREFIXES + MASK_STACK_ALIGN \
+ MASK_CONST_ALIGN + MASK_DATA_ALIGN \
+ MASK_PROLOGUE_EPILOGUE + MASK_MUL_BUG)
@@ -310,6 +320,16 @@ extern int cris_cpu_version;
#define TARGET_TRAP_USING_BREAK8 \
(cris_trap_using_break8 == 2 ? TARGET_HAS_BREAK : cris_trap_using_break8)
+/* The < v10 atomics turn off interrupts, so they don't need alignment.
+ Incidentally, by default alignment is off there causing variables to
+ be default unaligned all over, so we'd have to make support
+ libraries use a proper atomic type (instead of "int"), one we'd
+ specify as aligned. */
+#define TARGET_TRAP_UNALIGNED_ATOMIC \
+ (cris_trap_unaligned_atomic == 2 \
+ ? (TARGET_V32 || cris_cpu_version == 10) \
+ : cris_trap_unaligned_atomic)
+
/* Node: Storage Layout */
#define BITS_BIG_ENDIAN 0
@@ -55,46 +55,46 @@
;; The movsi for a gotless symbol could be split (post reload).
-(define_constants
+(define_c_enum ""
[
;; PLT reference from call expansion: operand 0 is the address,
;; the mode is VOIDmode. Always wrapped in CONST.
;; The value is relative to the GOT.
- (CRIS_UNSPEC_PLT_GOTREL 0)
+ CRIS_UNSPEC_PLT_GOTREL
;; PLT reference from call expansion: operand 0 is the address,
;; the mode is VOIDmode. Always wrapped in CONST.
;; The value is relative to the PC. It's arch-dependent whether
;; the offset counts from the start or the end of the current item.
- (CRIS_UNSPEC_PLT_PCREL 1)
+ CRIS_UNSPEC_PLT_PCREL
;; The address of the global offset table as a source operand.
- (CRIS_UNSPEC_GOT 2)
+ CRIS_UNSPEC_GOT
;; The offset from the global offset table to the operand.
- (CRIS_UNSPEC_GOTREL 3)
+ CRIS_UNSPEC_GOTREL
;; The PC-relative offset to the operand. It's arch-dependent whether
;; the offset counts from the start or the end of the current item.
- (CRIS_UNSPEC_PCREL 4)
+ CRIS_UNSPEC_PCREL
;; The index into the global offset table of a symbol, while
;; also generating a GOT entry for the symbol.
- (CRIS_UNSPEC_GOTREAD 5)
+ CRIS_UNSPEC_GOTREAD
;; Similar to CRIS_UNSPEC_GOTREAD, but also generating a PLT entry.
- (CRIS_UNSPEC_PLTGOTREAD 6)
+ CRIS_UNSPEC_PLTGOTREAD
;; Condition for v32 casesi jump, since it needs to have if_then_else
;; form with register as one branch and default label as other.
;; Operand 0 is const_int 0.
- (CRIS_UNSPEC_CASESI 7)
+ CRIS_UNSPEC_CASESI
;; Stack frame deallocation barrier.
- (CRIS_UNSPEC_FRAME_DEALLOC 8)
+ CRIS_UNSPEC_FRAME_DEALLOC
;; Swap all 32 bits of the operand; 31 <=> 0, 30 <=> 1...
- (CRIS_UNSPEC_SWAP_BITS 9)
+ CRIS_UNSPEC_SWAP_BITS
])
;; Register numbers.
@@ -4165,6 +4165,8 @@
3 [(match_dup 0)
(match_dup 1)]))]
"")
+
+(include "sync.md")
;; Splits for all cases in side-effect insns where (possibly after reload
;; and register allocation) rx and ry in [rx=ry+i] are equal.
@@ -179,6 +179,10 @@ mtrap-using-break8
Target Report Var(cris_trap_using_break8) Init(2)
Emit traps as \"break 8\", default for CRIS v3 and up. If disabled, calls to abort() are used.
+mtrap-unaligned-atomic
+Target Report Var(cris_trap_unaligned_atomic) Init(2)
+Emit checks causing \"break 8\" instructions to execute when applying atomic builtins on misaligned memory
+
; TARGET_SVINTO: Currently this just affects alignment. FIXME:
; Redundant with TARGET_ALIGN_BY_32, or put machine stuff here?
; This and the others below could just as well be variables and
@@ -0,0 +1,288 @@
+;; GCC machine description for CRIS atomic memory sequences.
+;; Copyright (C) 2012
+;; Free Software Foundation, Inc.
+;;
+;; This file is part of GCC.
+;;
+;; GCC is free software; you can redistribute it and/or modify
+;; it under the terms of the GNU General Public License as published by
+;; the Free Software Foundation; either version 3, or (at your option)
+;; any later version.
+;;
+;; GCC is distributed in the hope that it will be useful,
+;; but WITHOUT ANY WARRANTY; without even the implied warranty of
+;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+;; GNU General Public License for more details.
+;;
+;; You should have received a copy of the GNU General Public License
+;; along with GCC; see the file COPYING3. If not see
+;; <http://www.gnu.org/licenses/>.
+
+;; The CRIS atomic support yields code in three flavors, depending on
+;; the CPU for which code is generated:
+;;
+;; - Plain old CRIS v0 (..v8)
+;; - CRIS v10 (as used in ETRAX 100 LX)
+;; - CRIS v32 (as used in ETRAX FS)
+;;
+;; The last two alternatives are similar, of LL/SC type. They may
+;; fail for other reasons; an exception, a cache miss or a bus request
+;; from other parts of the system. The difference between them is
+;; just in what condition-codes are used to track LL and success or
+;; failure for the store. See the chapter on integral read-write
+;; operations, chapter 1.13 in "ETRAX 100LX Programmers Manual",
+;; <http://www.axis.com/files/tech_notes/etrax_100lx_prog_man-050519.pdf>
+;; and chapter 2.1 in "ETRAX FS Designer's reference",
+;; <http://www.axis.com/files/manuals/etrax_fs_des_ref-070821.pdf>.
+;; Note that the datum being stored has to be contained fully within a
+;; cache-line to be integral. A failure to store the data integrally
+;; will be flagged, but the store may still have happened in part,
+;; which translates most usefully into the data having to be
+;; "naturally aligned" to work. Natural alignment is verified in the
+;; generated code and will by default cause for unaligned pointers a
+;; "break 8" to be executed or optionally a call to abort(). Beware
+;; that options -m16bit and -m8bit may cause data to be unaligned
+;; where it was otherwise aligned. Data has a better chance of being
+;; aligned if it is declared with e.g. __attribute__ ((__align__ (4))).
+;;
+;; The "plain old v0..v8 flavor" just assumes there's a single CPU in
+;; the system, that no other parts of the system have access to memory
+;; used for atomic accesses and since there's no user mode without
+;; access to interrupt flags (another assumption), it just turns off
+;; interrupts while doing the access. Here, alignment is neither
+;; required nor asserted.
+
+(define_c_enum ""
+ [
+ CRIS_UNSPEC_ATOMIC_OP
+ CRIS_UNSPEC_ATOMIC_SWAP_MEM
+ CRIS_UNSPEC_ATOMIC_SWAP_BOOL
+ ])
+
+(define_constants [(CRIS_CCR_INTERRUPT_BIT 5)])
+
+;; We use "mult" as a placeholder for "nand" (which does not have a
+;; separate binary rtx operation) so we can use an iterator in the
+;; define_expand and define_insn and avoid having a separate
+;; mostly-identical copy. You will see the "mult" operator in rtl
+;; dumps, but it shouldn't matter as its use has one of its operands
+;; inside an unspec_volatile.
+
+(define_code_iterator atomic_op [plus minus ior and xor mult])
+
+(define_code_attr atomic_op_name
+ [(plus "add") (minus "sub") (and "and") (ior "or") (xor "xor") (mult "nand")])
+
+;; Pairs of these are used to insert the "not" after the "and" for nand.
+(define_code_attr atomic_op_mnem_pre ;; Upper-case only to sinplify testing.
+ [(plus "Add.d") (minus "Sub.d") (and "And.d") (ior "Or.d") (xor "Xor")
+ (mult "aNd.d")])
+(define_code_attr atomic_op_mnem_post_op3
+ [(plus "") (minus "") (and "") (ior "") (xor "") (mult "not %3\;")])
+
+(define_expand "atomic_fetch_<atomic_op_name><mode>"
+ [(match_operand:BWD 0 "register_operand")
+ (match_operand:BWD 1 "memory_operand")
+ (match_operand:BWD 2 "register_operand")
+ (match_operand 3)
+ (atomic_op:BWD (match_dup 0) (match_dup 1))]
+ ""
+{
+ if (<MODE>mode != QImode && TARGET_TRAP_UNALIGNED_ATOMIC)
+ cris_emit_trap_for_misalignment (operands[1]);
+
+ expand_mem_thread_fence (INTVAL (operands[3]));
+ emit_insn (gen_cris_atomic_fetch_<atomic_op_name><mode>_1 (operands[0],
+ operands[1],
+ operands[2]));
+ expand_mem_thread_fence (INTVAL (operands[3]));
+ DONE;
+})
+
+(define_insn "cris_atomic_fetch_<atomic_op_name><mode>_1"
+ [(set (match_operand:BWD 1 "memory_operand" "+Q")
+ (atomic_op:BWD
+ (unspec_volatile:BWD [(match_dup 1)] CRIS_UNSPEC_ATOMIC_OP)
+ ;; FIXME: relax this for plus, minus, and, ior.
+ (match_operand:BWD 2 "register_operand" "r")))
+ (set (match_operand:BWD 0 "register_operand" "=&r")
+ (match_dup 1))
+ (clobber (match_scratch:SI 3 "=&r"))]
+ ""
+{
+ /* Can't be too sure; better ICE if this happens. */
+ gcc_assert (!reg_overlap_mentioned_p (operands[2], operands[1]));
+
+ if (TARGET_V32)
+ return
+ "clearf p\n"
+ ".Lsync.%=:\;"
+ "move<m> %1,%0\;"
+ "move.d %0,%3\;"
+ "<atomic_op_mnem_pre> %2,%3\;<atomic_op_mnem_post_op3>"
+ "ax\;"
+ "move<m> %3,%1\;"
+ "bcs .Lsync.%=\;"
+ "clearf p";
+ else if (cris_cpu_version == 10)
+ return
+ "clearf\n"
+ ".Lsync.%=:\;"
+ "move<m> %1,%0\;"
+ "move.d %0,%3\;"
+ "<atomic_op_mnem_pre> %2,%3\;<atomic_op_mnem_post_op3>"
+ "ax\;"
+ "move<m> %3,%1\;"
+ "bwf .Lsync.%=\;"
+ "clearf";
+ else
+ {
+ /* This one is for CRIS versions without load-locked-store-conditional
+ machinery; assume single-core-non-shared-memory without user
+ mode/supervisor mode distinction, and just disable interrupts
+ while performing the operation.
+ Rather than making this pattern more complex by freeing another
+ register or stack position to save condition codes (the value
+ of the interrupt-enabled bit), we check whether interrupts were
+ enabled before we disabled them and branch to a version
+ with/without afterwards re-enabling them. */
+ rtx ops[5];
+
+ /* We have no available macro to stringify CRIS_CCR_INTERRUPT_BIT. */
+ memcpy (ops, operands, sizeof(ops));
+ ops[4] = GEN_INT (CRIS_CCR_INTERRUPT_BIT);
+
+ output_asm_insn ("move $ccr,%3\;"
+ "di\;"
+ "move<m> %1,%0\;"
+ "btstq %4,%3",
+ ops);
+ return
+ "bmi .Lsync.irqon.%=\;"
+ "move.d %0,%3\;"
+
+ "<atomic_op_mnem_pre> %2,%3\;<atomic_op_mnem_post_op3>"
+ "ba .Lsync.irqoff.%=\;"
+ "move<m> %3,%1\n"
+
+ ".Lsync.irqon.%=:\;"
+ "<atomic_op_mnem_pre> %2,%3\;<atomic_op_mnem_post_op3>"
+ "move<m> %3,%1\;"
+ "ei\n"
+ ".Lsync.irqoff.%=:";
+ }
+})
+
+;; This pattern is more-or-less assumed to always exist if any of the
+;; other atomic patterns exist (see e.g. comment at the
+;; can_compare_and_swap_p call in omp-low.c, 4.8 era). We'd slightly
+;; prefer atomic_exchange<mode> over this, but having both would be
+;; redundant.
+(define_expand "atomic_compare_and_swap<mode>"
+ [(match_operand:SI 0 "register_operand")
+ (match_operand:BWD 1 "register_operand")
+ (match_operand:BWD 2 "memory_operand")
+ (match_operand:BWD 3 "general_operand")
+ (match_operand:BWD 4 "register_operand")
+ (match_operand 5)
+ (match_operand 6)
+ (match_operand 7)]
+ ""
+{
+ if (<MODE>mode != QImode && TARGET_TRAP_UNALIGNED_ATOMIC)
+ cris_emit_trap_for_misalignment (operands[2]);
+
+ expand_mem_thread_fence (INTVAL (operands[6]));
+ emit_insn (gen_cris_atomic_compare_and_swap<mode>_1 (operands[0],
+ operands[1],
+ operands[2],
+ operands[3],
+ operands[4]));
+ expand_mem_thread_fence (INTVAL (operands[6]));
+ DONE;
+})
+
+(define_insn "cris_atomic_compare_and_swap<mode>_1"
+ [(set (match_operand:SI 0 "register_operand" "=&r")
+ (unspec_volatile:SI
+ [(match_operand:BWD 2 "memory_operand" "+Q")
+ (match_operand:BWD 3 "general_operand" "g")]
+ CRIS_UNSPEC_ATOMIC_SWAP_BOOL))
+ (set (match_operand:BWD 1 "register_operand" "=&r") (match_dup 2))
+ (set (match_dup 2)
+ (unspec_volatile:BWD
+ [(match_dup 2)
+ (match_dup 3)
+ (match_operand:BWD 4 "register_operand" "r")]
+ CRIS_UNSPEC_ATOMIC_SWAP_MEM))]
+ ""
+{
+ if (TARGET_V32)
+ return
+ "clearf p\n"
+ ".Lsync.repeat.%=:\;"
+ "move<m> %2,%1\;"
+ "cmp<m> %3,%1\;"
+ "bne .Lsync.after.%=\;"
+ "seq %0\;"
+
+ "ax\;"
+ "move<m> %4,%2\;"
+ "bcs .Lsync.repeat.%=\;"
+ "clearf p\n"
+ ".Lsync.after.%=:";
+ else if (cris_cpu_version == 10)
+ return
+ "clearf\n"
+ ".Lsync.repeat.%=:\;"
+ "move<m> %2,%1\;"
+ "cmp<m> %3,%1\;"
+ "bne .Lsync.after.%=\;"
+ "seq %0\;"
+
+ "ax\;"
+ "move<m> %4,%2\;"
+ "bwf .Lsync.repeat.%=\;"
+ "clearf\n"
+ ".Lsync.after.%=:";
+ else
+ {
+ /* This one is for CRIS versions without load-locked-store-conditional
+ machinery; assume single-core-non-shared-memory without user
+ mode/supervisor mode distinction, and just disable interrupts
+ while performing the operation.
+ Rather than making this pattern more complex by freeing another
+ register or stack position to save condition codes (the value
+ of the interrupt-enabled bit), we check whether interrupts were
+ enabled before we disabled them and branch to a version
+ with/without afterwards re-enabling them. */
+ rtx ops[4];
+
+ /* We have no available macro to stringify CRIS_CCR_INTERRUPT_BIT. */
+ memcpy (ops, operands, sizeof(ops));
+ ops[3] = GEN_INT (CRIS_CCR_INTERRUPT_BIT);
+
+ output_asm_insn ("move $ccr,%0\;"
+ "di\;"
+ "move<m> %2,%1\;"
+ "btstq %3,%0",
+ ops);
+ return
+ "bmi .Lsync.irqon.%=\;"
+ "nop\;"
+
+ "cmp<m> %3,%1\;"
+ "bne .Lsync.after.%=\;"
+ "seq %0\;"
+ "ba .Lsync.after.%=\;"
+ "move<m> %4,%2\n"
+
+ ".Lsync.irqon.%=:\;"
+ "cmp<m> %3,%1\;"
+ "bne .Lsync.after.%=\;"
+ "seq %0\;"
+ "move<m> %4,%2\;"
+ "ei\n"
+ ".Lsync.after.%=:";
+ }
+})
@@ -0,0 +1,4 @@
+/* Check that we can assemble both base atomic variants. */
+/* { dg-do assemble } */
+/* { dg-options "-O2 -march=v10" } */
+#include "sync-1.c"
@@ -0,0 +1,4 @@
+/* Check that we can assemble both base atomic variants. */
+/* { dg-do assemble } */
+/* { dg-options "-O2 -march=v32" } */
+#include "sync-1.c"
@@ -0,0 +1,35 @@
+/* Check that we can assemble both base atomic variants, for v0. */
+/* { dg-do assemble } */
+/* { dg-options "-O2 -march=v0" } */
+
+#ifndef type
+#define type char
+#endif
+
+#if !defined(op) && !defined(xchg)
+#define op 1
+#define xchg 1
+#endif
+
+#ifndef op
+#define op 0
+#endif
+
+#ifndef xchg
+#define xchg 0
+#endif
+
+#if op
+int sfa (type *p, type *q, int a)
+{
+ return __atomic_fetch_nand (p, a, __ATOMIC_ACQ_REL)
+ + __atomic_fetch_add (q, a, __ATOMIC_SEQ_CST);
+}
+#endif
+
+#if xchg
+void acen (type *ptr, type *val, type *ret)
+{
+ __atomic_exchange (ptr, val, ret, __ATOMIC_SEQ_CST);
+}
+#endif
@@ -0,0 +1,8 @@
+/* Check that we don't get alignment-checking code, char. */
+/* { dg-do compile } */
+/* { dg-options "-O2" } */
+/* { dg-final { scan-assembler-not "\tbreak\[ \t\]" } } */
+/* { dg-final { scan-assembler-not "\tbtstq\[ \t\]\[^5\]" } } */
+/* { dg-final { scan-assembler-not "\tand" } } */
+/* { dg-final { scan-assembler-not "\t\[jb\]sr" } } */
+#include "sync-1.c"
@@ -0,0 +1,9 @@
+/* Check that we get the expected alignment-checking code, op variant, int. */
+/* { dg-do compile } */
+/* { dg-options "-O2 -Dop -Dtype=int" } */
+/* { dg-additional-options "-mtrap-using-break8 -mtrap-unaligned-atomic" { target cris-*-elf } } */
+/* { dg-final { scan-assembler "\tbreak 8" } } */
+/* { dg-final { scan-assembler "\tbtstq \\(2-1\\)," } } */
+/* { dg-final { scan-assembler-not "\tand" } } */
+/* { dg-final { scan-assembler-not "\t\[jb\]sr" } } */
+#include "sync-1.c"
@@ -0,0 +1,9 @@
+/* Check that we get the expected alignment-checking code, op variant, short. */
+/* { dg-do compile } */
+/* { dg-options "-O2 -Dop -Dtype=short" } */
+/* { dg-additional-options "-mtrap-using-break8 -mtrap-unaligned-atomic" { target cris-*-elf } } */
+/* { dg-final { scan-assembler "\tbreak 8" } } */
+/* { dg-final { scan-assembler "\tbtstq \\(1-1\\)," } } */
+/* { dg-final { scan-assembler-not "\tand" } } */
+/* { dg-final { scan-assembler-not "\t\[jb\]sr" } } */
+#include "sync-1.c"
@@ -0,0 +1,11 @@
+/* Check that we get the expected alignment-checking code, xchg variant, int.
+ Unfortunately, PRE moves the "and" to a different BB, so combine doesn't
+ see it with the compare to make it a btstq. */
+/* { dg-do compile } */
+/* { dg-options "-O2 -Dxchg -Dtype=int" } */
+/* { dg-additional-options "-mtrap-using-break8 -mtrap-unaligned-atomic" { target cris-*-elf } } */
+/* { dg-final { scan-assembler "\tbreak 8" } } */
+/* { dg-final { scan-assembler "\tbtstq \\(2-1\\)," { xfail *-*-* } } } */
+/* { dg-final { scan-assembler-not "\tand" { xfail *-*-* } } } */
+/* { dg-final { scan-assembler-not "\t\[jb\]sr" } } */
+#include "sync-1.c"
@@ -0,0 +1,11 @@
+/* Check that we get the expected alignment-checking code, xchg variant, short.
+ Unfortunately, PRE moves the "and" to a different BB, so combine doesn't
+ see it with the compare to make it a btstq. */
+/* { dg-do compile } */
+/* { dg-options "-O2 -Dxchg -Dtype=short" } */
+/* { dg-additional-options "-mtrap-using-break8 -mtrap-unaligned-atomic" { target cris-*-elf } } */
+/* { dg-final { scan-assembler "\tbreak 8" } } */
+/* { dg-final { scan-assembler "\tbtstq \\(1-1\\)," { xfail *-*-* } } } */
+/* { dg-final { scan-assembler-not "\tand" { xfail *-*-* } } } */
+/* { dg-final { scan-assembler-not "\t\[jb\]sr" } } */
+#include "sync-1.c"
@@ -0,0 +1,8 @@
+/* Check that we get don't alignment-checking code, xchg variant, char. */
+/* { dg-do compile } */
+/* { dg-options "-O2 -mno-trap-unaligned-atomic" } */
+/* { dg-final { scan-assembler-not "\tbreak\[ \t\]" } } */
+/* { dg-final { scan-assembler-not "\tbtstq\[ \t\]\[^5\]" } } */
+/* { dg-final { scan-assembler-not "\tand" } } */
+/* { dg-final { scan-assembler-not "\t\[jb\]sr" } } */
+#include "sync-1.c"
@@ -0,0 +1,8 @@
+/* Check that we don't get alignment-checking code, int. */
+/* { dg-do compile } */
+/* { dg-options "-O2 -Dtype=int -mno-trap-unaligned-atomic" } */
+/* { dg-final { scan-assembler-not "\tbreak\[ \t\]" } } */
+/* { dg-final { scan-assembler-not "\tbtstq\[ \t\]\[^5\]" } } */
+/* { dg-final { scan-assembler-not "\tand" } } */
+/* { dg-final { scan-assembler-not "\t\[jb\]sr" } } */
+#include "sync-1.c"
@@ -0,0 +1,8 @@
+/* Check that we don't get alignment-checking code, short. */
+/* { dg-do compile } */
+/* { dg-options "-O2 -Dtype=short -mno-trap-unaligned-atomic" } */
+/* { dg-final { scan-assembler-not "\tbreak\[ \t\]" } } */
+/* { dg-final { scan-assembler-not "\tbtstq\[ \t\]\[^5\]" } } */
+/* { dg-final { scan-assembler-not "\tand" } } */
+/* { dg-final { scan-assembler-not "\t\[jb\]sr" } } */
+#include "sync-1.c"
@@ -0,0 +1,4 @@
+/* { dg-do run { target *-*-linux* } } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dop -Dtype=int" } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,5 @@
+/* { dg-do run } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dop -Dtype=int -DTRAP_USING_ABORT -mno-trap-using-break8" } */
+/* { dg-additional-options "-mtrap-unaligned-atomic" { target cris-*-elf } } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,4 @@
+/* { dg-do run { target *-*-linux* } } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dop -Dtype=int -Dmisalignment=2" } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,5 @@
+/* { dg-do run } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dop -Dtype=int -Dmisalignment=2 -DTRAP_USING_ABORT -mno-trap-using-break8" } */
+/* { dg-additional-options "-mtrap-unaligned-atomic" { target cris-*-elf } } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,4 @@
+/* { dg-do run { target *-*-linux* } } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dop -Dtype=int -Dmisalignment=3" } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,5 @@
+/* { dg-do run } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dop -Dtype=int -Dmisalignment=3 -DTRAP_USING_ABORT -mno-trap-using-break8" } */
+/* { dg-additional-options "-mtrap-unaligned-atomic" { target cris-*-elf } } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,111 @@
+/* { dg-do run { target *-*-linux* } } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dop -Dtype=short" } */
+
+/* Make sure we get a SIGTRAP or equivalent when passing unaligned
+ but otherwise valid pointers to the atomic builtins. */
+
+#include <signal.h>
+#include <stdlib.h>
+
+#ifndef type
+#error type not defined
+#endif
+
+#ifndef op
+#define op 0
+#endif
+
+#ifndef xchg
+#define xchg 0
+#endif
+
+#if op
+int sfa (type *p, type *q, int a);
+#endif
+
+#if xchg
+void acen (type *ptr, type *val, type *ret);
+#endif
+
+#ifndef misalignment
+#define misalignment 1
+#endif
+
+volatile int trap_expected = 0;
+
+struct { char x[misalignment]; type i; } s __attribute__ ((__aligned__ (4)))
+ = { {0}, (type) 0xdeadbeef };
+type x = 2;
+type ret = 42;
+
+#ifdef TRAP_USING_ABORT
+#define SYMSTR(x) STR1(__USER_LABEL_PREFIX__, x)
+#define STR1(x,y) STR2(x, y)
+#define STR2(x,y) #x #y
+/* LTO requires marking seemingly-unused-but-used global functions. */
+void my_abort (void) __asm__ (SYMSTR (abort)) __attribute__ ((__used__));
+void my_abort (void)
+#else
+#ifdef __gnu_linux__
+void trap_handler(int signum)
+#else
+#error "can't catch break 8"
+#endif
+#endif
+{
+ if (1
+#ifndef TRAP_USING_ABORT
+ && signum == SIGTRAP
+#endif
+ && trap_expected
+ && s.i == (type) 0xdeadbeef
+ && x == 2 && ret == 42)
+ exit (0);
+
+#ifdef TRAP_USING_ABORT
+ /* We might be able to trust the exit-value getting through, but add
+ a NULL-dereference SEGV just in case. Make sure gcc doesn't
+ understand the NULL. */
+ *({ int *p; asm ("" : "=rm" (p) : "0" (0)); p; }) = 0xdead;
+ exit (2);
+#else
+ abort ();
+#endif
+}
+
+int main(void)
+{
+ type x = 0;
+ type ret;
+
+#ifndef TRAP_USING_ABORT
+#ifdef __gnu_linux__
+ if (signal (SIGTRAP, trap_handler) == SIG_ERR)
+ abort ();
+#endif
+#endif
+
+ trap_expected = 1;
+
+#if op
+ sfa (&s.i, &s.i, 42);
+
+ /* We should have fallen into the trap now. But don't call abort
+ yet: if the trap is implemented as a call to abort, we have to
+ tell the difference. Set a global variable *and* make sure the
+ setting isn't eliminated by optimizers: another call to sfa
+ should do it. */
+ trap_expected = 0;
+
+ sfa (&x, &x, 1);
+#else
+ acen (&s.i, &x, &ret);
+
+ trap_expected = 0;
+
+ acen (&x, &x, &ret);
+#endif
+
+ abort ();
+}
@@ -0,0 +1,5 @@
+/* { dg-do run } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dop -Dtype=short -DTRAP_USING_ABORT -mno-trap-using-break8" } */
+/* { dg-additional-options "-mtrap-unaligned-atomic" { target cris-*-elf } } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,4 @@
+/* { dg-do run { target *-*-linux* } } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dxchg -Dtype=int" } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,5 @@
+/* { dg-do run } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dxchg -Dtype=int -DTRAP_USING_ABORT -mno-trap-using-break8" } */
+/* { dg-additional-options "-mtrap-unaligned-atomic" { target cris-*-elf } } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,4 @@
+/* { dg-do run { target *-*-linux* } } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dxchg -Dtype=int -Dmisalignment=2" } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,5 @@
+/* { dg-do run } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dxchg -Dtype=int -Dmisalignment=2 -DTRAP_USING_ABORT -mno-trap-using-break8" } */
+/* { dg-additional-options "-mtrap-unaligned-atomic" { target cris-*-elf } } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,4 @@
+/* { dg-do run { target *-*-linux* } } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dxchg -Dtype=int -Dmisalignment=3" } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,5 @@
+/* { dg-do run } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dxchg -Dtype=int -Dmisalignment=3 -DTRAP_USING_ABORT -mno-trap-using-break8" } */
+/* { dg-additional-options "-mtrap-unaligned-atomic" { target cris-*-elf } } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,4 @@
+/* { dg-do run { target *-*-linux* } } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dxchg -Dtype=short" } */
+#include "sync-mis-op-s-1.c"
@@ -0,0 +1,5 @@
+/* { dg-do run } */
+/* { dg-additional-sources "../sync-1.c" } */
+/* { dg-options "-Dxchg -Dtype=short -DTRAP_USING_ABORT -mno-trap-using-break8" } */
+/* { dg-additional-options "-mtrap-unaligned-atomic" { target cris-*-elf } } */
+#include "sync-mis-op-s-1.c"