/* $NetBSD: in_cksum_arm.c,v 1.3 2001/12/08 21:18:50 chris Exp $ */ /* * ARM version: * * Copyright (c) 1997 Mark Brinicome * Copyright (c) 1997 Causality Limited * * Based on the sparc version. */ /* * Sparc version: * * Copyright (c) 1995 Zubin Dittia. * Copyright (c) 1995 Matthew R. Green. * Copyright (c) 1994 Charles M. Hannum. * Copyright (c) 1992, 1993 * The Regents of the University of California. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: * This product includes software developed by the University of * California, Berkeley and its contributors. * 4. Neither the name of the University nor the names of its contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * @(#)in_cksum.c 8.1 (Berkeley) 6/11/93 */ #include #include #include #include #include #include #include /* * Checksum routine for Internet Protocol family headers. * * This routine is very heavily used in the network * code and should be modified for each CPU to be as fast as possible. * * ARM version. */ #define ADD64 __asm __volatile(" \n\ ldmia %0!, {%2, %3, %4, %5} \n\ adds %1,%7,%2; adcs %1,%1,%3 \n\ adcs %1,%1,%4; adcs %1,%1,%5 \n\ ldmia %0!, {%2, %3, %4, %5} \n\ adcs %1,%1,%2; adcs %1,%1,%3 \n\ adcs %1,%1,%4; adcs %1,%1,%5 \n\ ldmia %0!, {%2, %3, %4, %5} \n\ adcs %1,%1,%2; adcs %1,%1,%3 \n\ adcs %1,%1,%4; adcs %1,%1,%5 \n\ ldmia %0!, {%2, %3, %4, %5} \n\ adcs %1,%1,%2; adcs %1,%1,%3 \n\ adcs %1,%1,%4; adcs %1,%1,%5 \n\ adcs %1,%1,#0\n" \ : "=r" (w), "=r" (sum), "=&r" (tmp1), "=&r" (tmp2), "=&r" (tmp3), "=&r" (tmp4) \ : "0" (w), "r" (sum) \ : "cc") #define ADD32 __asm __volatile(" \n\ ldmia %0!, {%2, %3, %4, %5} \n\ adds %1,%7,%2; adcs %1,%1,%3 \n\ adcs %1,%1,%4; adcs %1,%1,%5 \n\ ldmia %0!, {%2, %3, %4, %5} \n\ adcs %1,%1,%2; adcs %1,%1,%3 \n\ adcs %1,%1,%4; adcs %1,%1,%5 \n\ adcs %1,%1,#0\n" \ : "=r" (w), "=r" (sum), "=&r" (tmp1), "=&r" (tmp2), "=&r" (tmp3), "=&r" (tmp4) \ : "0" (w), "r" (sum) \ : "cc") #define ADD16 __asm __volatile(" \n\ ldmia %0!, {%2, %3, %4, %5} \n\ adds %1,%7,%2; adcs %1,%1,%3 \n\ adcs %1,%1,%4; adcs %1,%1,%5 \n\ adcs %1,%1,#0\n" \ : "=r" (w), "=r" (sum), "=&r" (tmp1), "=&r" (tmp2), "=&r" (tmp3), "=&r" (tmp4) \ : "0" (w), "r" (sum) \ : "cc") #define ADD8 __asm __volatile(" \n\ ldmia %0!, {%2, %3} \n\ adds %1,%5,%2; adcs %1,%1,%3 \n\ adcs %1,%1,#0\n" \ : "=r" (w), "=r" (sum), "=&r" (tmp1), "=&r" (tmp2) \ : "0" (w), "r" (sum) \ : "cc" ) #define ADD4 __asm __volatile(" \n\ ldr %2,[%0],#4 \n\ adds %1,%4,%2 \n\ adcs %1,%1,#0\n" \ : "=r" (w), "=r" (sum), "=&r" (tmp1) \ : "0" (w), "r" (sum) \ : "cc") /*#define REDUCE {sum = (sum & 0xffff) + (sum >> 16);}*/ #define REDUCE __asm __volatile(" \n\ mov %2, #0x00ff \n\ orr %2, %2, #0xff00 \n\ and %2, %0, %2 \n\ add %0, %2, %0, lsr #16\n" \ : "=r" (sum) \ : "0" (sum), "r" (tmp1)) #define ADDCARRY {if (sum > 0xffff) sum -= 0xffff;} #define ROL {sum = sum << 8;} /* depends on recent REDUCE */ #define ADDBYTE {ROL; sum += (*w << 8); byte_swapped ^= 1;} #define ADDSHORT {sum += *(u_short *)w;} #define ADVANCE(n) {w += n; mlen -= n;} #define ADVANCEML(n) {mlen -= n;} static __inline__ int in_cksum_internal(struct mbuf *m, int off, int len, u_int sum) { u_char *w; int mlen = 0; int byte_swapped = 0; /* * Declare four temporary registers for use by the asm code. We * allow the compiler to pick which specific machine registers to * use, instead of hard-coding this in the asm code above. */ register u_int tmp1=0, tmp2, tmp3, tmp4; for (; m && len; m = m->m_next) { if (m->m_len == 0) continue; w = mtod(m, u_char *) + off; mlen = m->m_len - off; off = 0; if (len < mlen) mlen = len; len -= mlen; /* * Ensure that we're aligned on a word boundary here so * that we can do 32 bit operations below. */ if ((3 & (long)w) != 0) { REDUCE; if ((1 & (long)w) != 0 && mlen >= 1) { ADDBYTE; ADVANCE(1); } if ((2 & (long)w) != 0 && mlen >= 2) { ADDSHORT; ADVANCE(2); } } /* * Do as many 32 bit operations as possible using the * 64/32/16/8/4 macro's above, using as many as possible of * these. */ while (mlen >= 64) { ADD64; ADVANCEML(64); } if (mlen >= 32) { ADD32; ADVANCEML(32); } if (mlen >= 16) { ADD16; ADVANCEML(16); } if (mlen >= 8) { ADD8; ADVANCEML(8); } if (mlen >= 4) { ADD4; ADVANCEML(4) } if (mlen == 0) continue; REDUCE; if (mlen >= 2) { ADDSHORT; ADVANCE(2); } if (mlen == 1) { ADDBYTE; } } if (byte_swapped) { REDUCE; ROL; } REDUCE; ADDCARRY; return (0xffff ^ sum); } int in_cksum(m, len) struct mbuf *m; int len; { int cksum; cksum =in_cksum_internal(m, 0, len, 0); return cksum; } int in4_cksum(m, nxt, off, len) struct mbuf *m; u_int8_t nxt; int off, len; { u_int sum = 0; if (nxt != 0) { /* for ADD macros */ register u_int tmp1, tmp2, tmp3, tmp4; u_char *w; struct ipovly ipov; /* pseudo header */ if (off < sizeof(struct ipovly)) panic("in4_cksum: offset too short"); if (m->m_len < sizeof(struct ip)) panic("in4_cksum: bad mbuf chain"); bzero(&ipov, sizeof(ipov)); ipov.ih_len = htons(len); ipov.ih_pr = nxt; ipov.ih_src = mtod(m, struct ip *)->ip_src; ipov.ih_dst = mtod(m, struct ip *)->ip_dst; w = (u_char *)&ipov; /* assumes sizeof(ipov) == 20 */ ADD16; ADD4; } /* skip unnecessary part */ while (m && off > 0) { if (m->m_len > off) break; off -= m->m_len; m = m->m_next; } return (in_cksum_internal(m, off, len, sum)); }