OpenCores
URL https://opencores.org/ocsvn/or1k/or1k/trunk

Subversion Repositories or1k

[/] [or1k/] [trunk/] [linux/] [linux-2.4/] [include/] [asm-s390/] [checksum.h] - Blame information for rev 1774

Go to most recent revision | Details | Compare with Previous | View Log

Line No. Rev Author Line
1 1276 phoenix
#ifndef _S390_CHECKSUM_H
2
#define _S390_CHECKSUM_H
3
 
4
/*
5
 *  include/asm-s390/checksum.h
6
 *    S390 fast network checksum routines
7
 *    see also arch/S390/lib/checksum.c
8
 *
9
 *  S390 version
10
 *    Copyright (C) 1999 IBM Deutschland Entwicklung GmbH, IBM Corporation
11
 *    Author(s): Ulrich Hild        (first version)
12
 *               Martin Schwidefsky (heavily optimized CKSM version)
13
 *               D.J. Barrow        (third attempt)
14
 */
15
 
16
#include <asm/uaccess.h>
17
 
18
/*
19
 * computes the checksum of a memory block at buff, length len,
20
 * and adds in "sum" (32-bit)
21
 *
22
 * returns a 32-bit number suitable for feeding into itself
23
 * or csum_tcpudp_magic
24
 *
25
 * this function must be called with even lengths, except
26
 * for the last fragment, which may be odd
27
 *
28
 * it's best to have buff aligned on a 32-bit boundary
29
 */
30
unsigned int
31
csum_partial(const unsigned char * buff, int len, unsigned int sum);
32
 
33
/*
34
 * csum_partial as an inline function
35
 */
36
extern inline unsigned int
37
csum_partial_inline(const unsigned char * buff, int len, unsigned int sum)
38
{
39
        register_pair rp;
40
 
41
        rp.subreg.even = (unsigned long) buff;
42
        rp.subreg.odd = (unsigned long) len;
43
        __asm__ __volatile__ (
44
                "0:  cksm %0,%1\n"    /* do checksum on longs */
45
                "    jo   0b\n"
46
                : "+&d" (sum), "+&a" (rp) : : "cc" );
47
        return sum;
48
}
49
 
50
/*
51
 * the same as csum_partial, but copies from src while it
52
 * checksums
53
 *
54
 * here even more important to align src and dst on a 32-bit (or even
55
 * better 64-bit) boundary
56
 */
57
 
58
extern inline unsigned int
59
csum_partial_copy(const char *src, char *dst, int len,unsigned int sum)
60
{
61
        memcpy(dst,src,len);
62
        return csum_partial_inline(dst, len, sum);
63
}
64
 
65
/*
66
 * the same as csum_partial_copy, but copies from user space.
67
 *
68
 * here even more important to align src and dst on a 32-bit (or even
69
 * better 64-bit) boundary
70
 *
71
 * Copy from userspace and compute checksum.  If we catch an exception
72
 * then zero the rest of the buffer.
73
 */
74
extern inline unsigned int
75
csum_partial_copy_from_user (const char *src, char *dst,
76
                                          int len, unsigned int sum,
77
                                          int *err_ptr)
78
{
79
        int missing;
80
 
81
        missing = copy_from_user(dst, src, len);
82
        if (missing) {
83
                memset(dst + len - missing, 0, missing);
84
                *err_ptr = -EFAULT;
85
        }
86
 
87
        return csum_partial(dst, len, sum);
88
}
89
 
90
 
91
extern inline unsigned int
92
csum_partial_copy_nocheck (const char *src, char *dst, int len, unsigned int sum)
93
{
94
        memcpy(dst,src,len);
95
        return csum_partial_inline(dst, len, sum);
96
}
97
 
98
/*
99
 *      Fold a partial checksum without adding pseudo headers
100
 */
101
#if 1
102
unsigned short csum_fold(unsigned int sum);
103
#else
104
extern inline unsigned short
105
csum_fold(unsigned int sum)
106
{
107
        register_pair rp;
108
 
109
        __asm__ __volatile__ (
110
                "    slr  %N1,%N1\n" /* %0 = H L */
111
                "    lr   %1,%0\n"   /* %0 = H L, %1 = H L 0 0 */
112
                "    srdl %1,16\n"   /* %0 = H L, %1 = 0 H L 0 */
113
                "    alr  %1,%N1\n"  /* %0 = H L, %1 = L H L 0 */
114
                "    alr  %0,%1\n"   /* %0 = H+L+C L+H */
115
                "    srl  %0,16\n"   /* %0 = H+L+C */
116
                : "+&d" (sum), "=d" (rp) : : "cc" );
117
        return ((unsigned short) ~sum);
118
}
119
#endif
120
 
121
/*
122
 *      This is a version of ip_compute_csum() optimized for IP headers,
123
 *      which always checksum on 4 octet boundaries.
124
 *
125
 */
126
extern inline unsigned short
127
ip_fast_csum(unsigned char *iph, unsigned int ihl)
128
{
129
        register_pair rp;
130
        unsigned long sum;
131
 
132
        rp.subreg.even = (unsigned long) iph;
133
        rp.subreg.odd = (unsigned long) ihl*4;
134
        __asm__ __volatile__ (
135
                "    sr   %0,%0\n"   /* set sum to zero */
136
                "0:  cksm %0,%1\n"   /* do checksum on longs */
137
                "    jo   0b\n"
138
                : "=&d" (sum), "+&a" (rp) : : "cc" );
139
        return csum_fold(sum);
140
}
141
 
142
/*
143
 * computes the checksum of the TCP/UDP pseudo-header
144
 * returns a 32-bit checksum
145
 */
146
extern inline unsigned int
147
csum_tcpudp_nofold(unsigned long saddr, unsigned long daddr,
148
                   unsigned short len, unsigned short proto,
149
                   unsigned int sum)
150
{
151
        __asm__ __volatile__ (
152
                "    alr   %0,%1\n"  /* sum += saddr */
153
                "    brc   12,0f\n"
154
                "    ahi   %0,1\n"   /* add carry */
155
                "0:"
156
                : "+&d" (sum) : "d" (saddr) : "cc" );
157
        __asm__ __volatile__ (
158
                "    alr   %0,%1\n"  /* sum += daddr */
159
                "    brc   12,1f\n"
160
                "    ahi   %0,1\n"   /* add carry */
161
                "1:"
162
                : "+&d" (sum) : "d" (daddr) : "cc" );
163
        __asm__ __volatile__ (
164
                "    alr   %0,%1\n"  /* sum += (len<<16) + (proto<<8) */
165
                "    brc   12,2f\n"
166
                "    ahi   %0,1\n"   /* add carry */
167
                "2:"
168
                : "+&d" (sum)
169
                : "d" (((unsigned int) len<<16) + (unsigned int) proto)
170
                : "cc" );
171
        return sum;
172
}
173
 
174
/*
175
 * computes the checksum of the TCP/UDP pseudo-header
176
 * returns a 16-bit checksum, already complemented
177
 */
178
 
179
extern inline unsigned short int
180
csum_tcpudp_magic(unsigned long saddr, unsigned long daddr,
181
                  unsigned short len, unsigned short proto,
182
                  unsigned int sum)
183
{
184
        return csum_fold(csum_tcpudp_nofold(saddr,daddr,len,proto,sum));
185
}
186
 
187
/*
188
 * this routine is used for miscellaneous IP-like checksums, mainly
189
 * in icmp.c
190
 */
191
 
192
extern inline unsigned short
193
ip_compute_csum(unsigned char * buff, int len)
194
{
195
        return csum_fold(csum_partial(buff, len, 0));
196
}
197
 
198
#endif /* _S390_CHECKSUM_H */
199
 
200
 

powered by: WebSVN 2.1.0

© copyright 1999-2024 OpenCores.org, equivalent to Oliscience, all rights reserved. OpenCores®, registered trademark.