#include <linux/kernel.h> #include <linux/sched.h> #include <linux/types.h> #include <asm/byteorder.h> #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ __asm__ ("addcc %r4,%5,%1\n\t" \ "addx %r2,%3,%0\n" \ : "=r" ((USItype)(sh)), \ "=&r" ((USItype)(sl)) \ : "%rJ" ((USItype)(ah)), \ "rI" ((USItype)(bh)), \ "%rJ" ((USItype)(al)), \ "rI" ((USItype)(bl)) \ : "cc") #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ __asm__ ("subcc %r4,%5,%1\n\t" \ "subx %r2,%3,%0\n" \ : "=r" ((USItype)(sh)), \ "=&r" ((USItype)(sl)) \ : "rJ" ((USItype)(ah)), \ "rI" ((USItype)(bh)), \ "rJ" ((USItype)(al)), \ "rI" ((USItype)(bl)) \ : "cc") #define umul_ppmm(w1, w0, u, v) \ __asm__ ("! Inlined umul_ppmm\n\t" \ "wr %%g0,%2,%%y ! SPARC has 0-3 delay insn after a wr\n\t" \ "sra %3,31,%%g2 ! Don't move this insn\n\t" \ "and %2,%%g2,%%g2 ! Don't move this insn\n\t" \ "andcc %%g0,0,%%g1 ! Don't move this insn\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,%3,%%g1\n\t" \ "mulscc %%g1,0,%%g1\n\t" \ "add %%g1,%%g2,%0\n\t" \ "rd %%y,%1\n" \ : "=r" ((USItype)(w1)), \ "=r" ((USItype)(w0)) \ : "%rI" ((USItype)(u)), \ "r" ((USItype)(v)) \ : "%g1", "%g2", "cc") /* It's quite necessary to add this much assembler for the sparc. The default udiv_qrnnd (in C) is more than 10 times slower! */ #define udiv_qrnnd(q, r, n1, n0, d) \ __asm__ ("! Inlined udiv_qrnnd\n\t" \ "mov 32,%%g1\n\t" \ "subcc %1,%2,%%g0\n\t" \ "1: bcs 5f\n\t" \ "addxcc %0,%0,%0 ! shift n1n0 and a q-bit in lsb\n\t" \ "sub %1,%2,%1 ! this kills msb of n\n\t" \ "addx %1,%1,%1 ! so this can't give carry\n\t" \ "subcc %%g1,1,%%g1\n\t" \ "2: bne 1b\n\t" \ "subcc %1,%2,%%g0\n\t" \ "bcs 3f\n\t" \ "addxcc %0,%0,%0 ! shift n1n0 and a q-bit in lsb\n\t" \ "b 3f\n\t" \ "sub %1,%2,%1 ! this kills msb of n\n\t" \ "4: sub %1,%2,%1\n\t" \ "5: addxcc %1,%1,%1\n\t" \ "bcc 2b\n\t" \ "subcc %%g1,1,%%g1\n\t" \ "! Got carry from n. Subtract next step to cancel this carry.\n\t" \ "bne 4b\n\t" \ "addcc %0,%0,%0 ! shift n1n0 and a 0-bit in lsb\n\t" \ "sub %1,%2,%1\n\t" \ "3: xnor %0,0,%0\n\t" \ "! End of inline udiv_qrnnd\n" \ : "=&r" ((USItype)(q)), \ "=&r" ((USItype)(r)) \ : "r" ((USItype)(d)), \ "1" ((USItype)(n1)), \ "0" ((USItype)(n0)) : "%g1", "cc") #define UDIV_NEEDS_NORMALIZATION 0 #define abort() \ return 0 #ifdef __BIG_ENDIAN #define __BYTE_ORDER __BIG_ENDIAN #else #define __BYTE_ORDER __LITTLE_ENDIAN #endif