#ifndef __ASM_ARM_CHECKSUM_H
|
#ifndef __ASM_ARM_CHECKSUM_H
|
#define __ASM_ARM_CHECKSUM_H
|
#define __ASM_ARM_CHECKSUM_H
|
|
|
#ifndef __ASM_ARM_SEGMENT_H
|
#ifndef __ASM_ARM_SEGMENT_H
|
#include <asm/segment.h>
|
#include <asm/segment.h>
|
#endif
|
#endif
|
|
|
/*
|
/*
|
* computes the checksum of a memory block at buff, length len,
|
* computes the checksum of a memory block at buff, length len,
|
* and adds in "sum" (32-bit)
|
* and adds in "sum" (32-bit)
|
*
|
*
|
* returns a 32-bit number suitable for feeding into itself
|
* returns a 32-bit number suitable for feeding into itself
|
* or csum_tcpudp_magic
|
* or csum_tcpudp_magic
|
*
|
*
|
* this function must be called with even lengths, except
|
* this function must be called with even lengths, except
|
* for the last fragment, which may be odd
|
* for the last fragment, which may be odd
|
*
|
*
|
* it's best to have buff aligned on a 32-bit boundary
|
* it's best to have buff aligned on a 32-bit boundary
|
*/
|
*/
|
unsigned int csum_partial(const unsigned char * buff, int len, unsigned int sum);
|
unsigned int csum_partial(const unsigned char * buff, int len, unsigned int sum);
|
|
|
/*
|
/*
|
* the same as csum_partial, but copies from src while it
|
* the same as csum_partial, but copies from src while it
|
* checksums
|
* checksums
|
*
|
*
|
* here even more important to align src and dst on a 32-bit (or even
|
* here even more important to align src and dst on a 32-bit (or even
|
* better 64-bit) boundary
|
* better 64-bit) boundary
|
*/
|
*/
|
|
|
unsigned int csum_partial_copy( const char *src, char *dst, int len, int sum);
|
unsigned int csum_partial_copy( const char *src, char *dst, int len, int sum);
|
|
|
|
|
/*
|
/*
|
* the same as csum_partial_copy, but copies from user space.
|
* the same as csum_partial_copy, but copies from user space.
|
*
|
*
|
* here even more important to align src and dst on a 32-bit (or even
|
* here even more important to align src and dst on a 32-bit (or even
|
* better 64-bit) boundary
|
* better 64-bit) boundary
|
*/
|
*/
|
|
|
static __INLINE__ unsigned int csum_partial_copy_fromuser(const char *src, char *dst, int len, int sum)
|
static __INLINE__ unsigned int csum_partial_copy_fromuser(const char *src, char *dst, int len, int sum)
|
{
|
{
|
extern unsigned int __csum_partial_copy_fromuser(const char *src, char *dst, int len, int sum);
|
extern unsigned int __csum_partial_copy_fromuser(const char *src, char *dst, int len, int sum);
|
|
|
if (IS_USER_SEG)
|
if (IS_USER_SEG)
|
return __csum_partial_copy_fromuser (src, dst, len, sum);
|
return __csum_partial_copy_fromuser (src, dst, len, sum);
|
else
|
else
|
return csum_partial_copy (src, dst, len, sum);
|
return csum_partial_copy (src, dst, len, sum);
|
}
|
}
|
|
|
/*
|
/*
|
* This is a version of ip_compute_csum() optimized for IP headers,
|
* This is a version of ip_compute_csum() optimized for IP headers,
|
* which always checksum on 4 octet boundaries.
|
* which always checksum on 4 octet boundaries.
|
*
|
*
|
* By Jorge Cwik <jorge@laser.satlink.net>, adapted for linux by
|
* By Jorge Cwik <jorge@laser.satlink.net>, adapted for linux by
|
* Arnt Gulbrandsen.
|
* Arnt Gulbrandsen.
|
*
|
*
|
* Converted to ARM by R.M.King
|
* Converted to ARM by R.M.King
|
*/
|
*/
|
static inline unsigned short ip_fast_csum(unsigned char * iph,
|
static inline unsigned short ip_fast_csum(unsigned char * iph,
|
unsigned int ihl) {
|
unsigned int ihl) {
|
unsigned int sum, tmp1;
|
unsigned int sum, tmp1;
|
|
|
__asm__ __volatile__("
|
__asm__ __volatile__("
|
sub %2, %2, #5
|
sub %2, %2, #5
|
ldr %0, [%1], #4
|
ldr %0, [%1], #4
|
ldr %3, [%1], #4
|
ldr %3, [%1], #4
|
adds %0, %0, %3
|
adds %0, %0, %3
|
ldr %3, [%1], #4
|
ldr %3, [%1], #4
|
adcs %0, %0, %3
|
adcs %0, %0, %3
|
ldr %3, [%1], #4
|
ldr %3, [%1], #4
|
adcs %0, %0, %3
|
adcs %0, %0, %3
|
1: ldr %3, [%1], #4
|
1: ldr %3, [%1], #4
|
adcs %0, %0, %3
|
adcs %0, %0, %3
|
tst %2, #15
|
tst %2, #15
|
subne %2, %2, #1
|
subne %2, %2, #1
|
bne 1b
|
bne 1b
|
adc %0, %0, #0
|
adc %0, %0, #0
|
adds %0, %0, %0, lsl #16
|
adds %0, %0, %0, lsl #16
|
addcs %0, %0, #0x10000
|
addcs %0, %0, #0x10000
|
mvn %0, %0
|
mvn %0, %0
|
mov %0, %0, lsr #16
|
mov %0, %0, lsr #16
|
"
|
"
|
: "=&r" (sum), "=&r" (iph), "=&r" (ihl), "=&r" (tmp1)
|
: "=&r" (sum), "=&r" (iph), "=&r" (ihl), "=&r" (tmp1)
|
: "1" (iph), "2" (ihl));
|
: "1" (iph), "2" (ihl));
|
return(sum);
|
return(sum);
|
}
|
}
|
|
|
/*
|
/*
|
* computes the checksum of the TCP/UDP pseudo-header
|
* computes the checksum of the TCP/UDP pseudo-header
|
* returns a 16-bit checksum, already complemented
|
* returns a 16-bit checksum, already complemented
|
*/
|
*/
|
static inline unsigned short int csum_tcpudp_magic(unsigned long saddr,
|
static inline unsigned short int csum_tcpudp_magic(unsigned long saddr,
|
unsigned long daddr,
|
unsigned long daddr,
|
unsigned short len,
|
unsigned short len,
|
unsigned short proto,
|
unsigned short proto,
|
unsigned int sum) {
|
unsigned int sum) {
|
__asm__ __volatile__("
|
__asm__ __volatile__("
|
adds %0, %0, %1
|
adds %0, %0, %1
|
adcs %0, %0, %4
|
adcs %0, %0, %4
|
adcs %0, %0, %5
|
adcs %0, %0, %5
|
adc %0, %0, #0
|
adc %0, %0, #0
|
adds %0, %0, %0, lsl #16
|
adds %0, %0, %0, lsl #16
|
addcs %0, %0, #0x10000
|
addcs %0, %0, #0x10000
|
mvn %0, %0
|
mvn %0, %0
|
mov %0, %0, lsr #16
|
mov %0, %0, lsr #16
|
"
|
"
|
: "=&r" (sum), "=&r" (saddr)
|
: "=&r" (sum), "=&r" (saddr)
|
: "0" (daddr), "1"(saddr), "r"((ntohs(len)<<16)+proto*256), "r"(sum));
|
: "0" (daddr), "1"(saddr), "r"((ntohs(len)<<16)+proto*256), "r"(sum));
|
return((unsigned short)sum);
|
return((unsigned short)sum);
|
}
|
}
|
|
|
/*
|
/*
|
* Fold a partial checksum without adding pseudo headers
|
* Fold a partial checksum without adding pseudo headers
|
*/
|
*/
|
static inline unsigned int csum_fold(unsigned int sum)
|
static inline unsigned int csum_fold(unsigned int sum)
|
{
|
{
|
__asm__ __volatile__("
|
__asm__ __volatile__("
|
adds %0, %0, %0, lsl #16
|
adds %0, %0, %0, lsl #16
|
addcss %0, %0, #0x10000
|
addcss %0, %0, #0x10000
|
addcs %0, %0, #0x10000
|
addcs %0, %0, #0x10000
|
mvn %0, %0
|
mvn %0, %0
|
mov %0, %0, lsr #16
|
mov %0, %0, lsr #16
|
"
|
"
|
: "=r" (sum)
|
: "=r" (sum)
|
: "0" (sum));
|
: "0" (sum));
|
return sum;
|
return sum;
|
}
|
}
|
|
|
|
|
/*
|
/*
|
* this routine is used for miscellaneous IP-like checksums, mainly
|
* this routine is used for miscellaneous IP-like checksums, mainly
|
* in icmp.c
|
* in icmp.c
|
*/
|
*/
|
|
|
static inline unsigned short ip_compute_csum(unsigned char * buff, int len) {
|
static inline unsigned short ip_compute_csum(unsigned char * buff, int len) {
|
unsigned int sum;
|
unsigned int sum;
|
|
|
__asm__ __volatile__("
|
__asm__ __volatile__("
|
adds %0, %0, %0, lsl #16
|
adds %0, %0, %0, lsl #16
|
addcs %0, %0, #0x10000
|
addcs %0, %0, #0x10000
|
mvn %0, %0
|
mvn %0, %0
|
mov %0, %0, lsr #16
|
mov %0, %0, lsr #16
|
"
|
"
|
: "=r"(sum)
|
: "=r"(sum)
|
: "0" (csum_partial(buff, len, 0)));
|
: "0" (csum_partial(buff, len, 0)));
|
return(sum);
|
return(sum);
|
}
|
}
|
|
|
#endif
|
#endif
|
|
|