908
|
1 //#include "arm_arch.h" |
|
2 #define __ARM_ARCH__ 6 |
|
3 |
|
4 .text |
|
5 |
|
6 .global sha1_block_data_order |
|
7 .type sha1_block_data_order,%function |
|
8 |
|
9 .align 2 |
|
10 sha1_block_data_order: |
|
11 stmdb sp!,{r4-r12,lr} |
|
12 add r2,r1,r2,lsl#6 @ r2 to point at the end of r1 |
|
13 ldmia r0,{r3,r4,r5,r6,r7} |
|
14 .Lloop: |
|
15 ldr r8,.LK_00_19 |
|
16 mov r14,sp |
|
17 sub sp,sp,#15*4 |
|
18 mov r5,r5,ror#30 |
|
19 mov r6,r6,ror#30 |
|
20 mov r7,r7,ror#30 @ [6] |
|
21 .L_00_15: |
|
22 #if __ARM_ARCH__<7 |
|
23 ldrb r10,[r1,#2] |
|
24 ldrb r9,[r1,#3] |
|
25 ldrb r11,[r1,#1] |
|
26 add r7,r8,r7,ror#2 @ E+=K_00_19 |
|
27 ldrb r12,[r1],#4 |
|
28 orr r9,r9,r10,lsl#8 |
|
29 eor r10,r5,r6 @ F_xx_xx |
|
30 orr r9,r9,r11,lsl#16 |
|
31 add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
|
32 orr r9,r9,r12,lsl#24 |
|
33 #else |
|
34 ldr r9,[r1],#4 @ handles unaligned |
|
35 add r7,r8,r7,ror#2 @ E+=K_00_19 |
|
36 eor r10,r5,r6 @ F_xx_xx |
|
37 add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
|
38 #ifdef __ARMEL__ |
|
39 rev r9,r9 @ byte swap |
|
40 #endif |
|
41 #endif |
|
42 and r10,r4,r10,ror#2 |
|
43 add r7,r7,r9 @ E+=X[i] |
|
44 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) |
|
45 str r9,[r14,#-4]! |
|
46 add r7,r7,r10 @ E+=F_00_19(B,C,D) |
|
47 #if __ARM_ARCH__<7 |
|
48 ldrb r10,[r1,#2] |
|
49 ldrb r9,[r1,#3] |
|
50 ldrb r11,[r1,#1] |
|
51 add r6,r8,r6,ror#2 @ E+=K_00_19 |
|
52 ldrb r12,[r1],#4 |
|
53 orr r9,r9,r10,lsl#8 |
|
54 eor r10,r4,r5 @ F_xx_xx |
|
55 orr r9,r9,r11,lsl#16 |
|
56 add r6,r6,r7,ror#27 @ E+=ROR(A,27) |
|
57 orr r9,r9,r12,lsl#24 |
|
58 #else |
|
59 ldr r9,[r1],#4 @ handles unaligned |
|
60 add r6,r8,r6,ror#2 @ E+=K_00_19 |
|
61 eor r10,r4,r5 @ F_xx_xx |
|
62 add r6,r6,r7,ror#27 @ E+=ROR(A,27) |
|
63 #ifdef __ARMEL__ |
|
64 rev r9,r9 @ byte swap |
|
65 #endif |
|
66 #endif |
|
67 and r10,r3,r10,ror#2 |
|
68 add r6,r6,r9 @ E+=X[i] |
|
69 eor r10,r10,r5,ror#2 @ F_00_19(B,C,D) |
|
70 str r9,[r14,#-4]! |
|
71 add r6,r6,r10 @ E+=F_00_19(B,C,D) |
|
72 #if __ARM_ARCH__<7 |
|
73 ldrb r10,[r1,#2] |
|
74 ldrb r9,[r1,#3] |
|
75 ldrb r11,[r1,#1] |
|
76 add r5,r8,r5,ror#2 @ E+=K_00_19 |
|
77 ldrb r12,[r1],#4 |
|
78 orr r9,r9,r10,lsl#8 |
|
79 eor r10,r3,r4 @ F_xx_xx |
|
80 orr r9,r9,r11,lsl#16 |
|
81 add r5,r5,r6,ror#27 @ E+=ROR(A,27) |
|
82 orr r9,r9,r12,lsl#24 |
|
83 #else |
|
84 ldr r9,[r1],#4 @ handles unaligned |
|
85 add r5,r8,r5,ror#2 @ E+=K_00_19 |
|
86 eor r10,r3,r4 @ F_xx_xx |
|
87 add r5,r5,r6,ror#27 @ E+=ROR(A,27) |
|
88 #ifdef __ARMEL__ |
|
89 rev r9,r9 @ byte swap |
|
90 #endif |
|
91 #endif |
|
92 and r10,r7,r10,ror#2 |
|
93 add r5,r5,r9 @ E+=X[i] |
|
94 eor r10,r10,r4,ror#2 @ F_00_19(B,C,D) |
|
95 str r9,[r14,#-4]! |
|
96 add r5,r5,r10 @ E+=F_00_19(B,C,D) |
|
97 #if __ARM_ARCH__<7 |
|
98 ldrb r10,[r1,#2] |
|
99 ldrb r9,[r1,#3] |
|
100 ldrb r11,[r1,#1] |
|
101 add r4,r8,r4,ror#2 @ E+=K_00_19 |
|
102 ldrb r12,[r1],#4 |
|
103 orr r9,r9,r10,lsl#8 |
|
104 eor r10,r7,r3 @ F_xx_xx |
|
105 orr r9,r9,r11,lsl#16 |
|
106 add r4,r4,r5,ror#27 @ E+=ROR(A,27) |
|
107 orr r9,r9,r12,lsl#24 |
|
108 #else |
|
109 ldr r9,[r1],#4 @ handles unaligned |
|
110 add r4,r8,r4,ror#2 @ E+=K_00_19 |
|
111 eor r10,r7,r3 @ F_xx_xx |
|
112 add r4,r4,r5,ror#27 @ E+=ROR(A,27) |
|
113 #ifdef __ARMEL__ |
|
114 rev r9,r9 @ byte swap |
|
115 #endif |
|
116 #endif |
|
117 and r10,r6,r10,ror#2 |
|
118 add r4,r4,r9 @ E+=X[i] |
|
119 eor r10,r10,r3,ror#2 @ F_00_19(B,C,D) |
|
120 str r9,[r14,#-4]! |
|
121 add r4,r4,r10 @ E+=F_00_19(B,C,D) |
|
122 #if __ARM_ARCH__<7 |
|
123 ldrb r10,[r1,#2] |
|
124 ldrb r9,[r1,#3] |
|
125 ldrb r11,[r1,#1] |
|
126 add r3,r8,r3,ror#2 @ E+=K_00_19 |
|
127 ldrb r12,[r1],#4 |
|
128 orr r9,r9,r10,lsl#8 |
|
129 eor r10,r6,r7 @ F_xx_xx |
|
130 orr r9,r9,r11,lsl#16 |
|
131 add r3,r3,r4,ror#27 @ E+=ROR(A,27) |
|
132 orr r9,r9,r12,lsl#24 |
|
133 #else |
|
134 ldr r9,[r1],#4 @ handles unaligned |
|
135 add r3,r8,r3,ror#2 @ E+=K_00_19 |
|
136 eor r10,r6,r7 @ F_xx_xx |
|
137 add r3,r3,r4,ror#27 @ E+=ROR(A,27) |
|
138 #ifdef __ARMEL__ |
|
139 rev r9,r9 @ byte swap |
|
140 #endif |
|
141 #endif |
|
142 and r10,r5,r10,ror#2 |
|
143 add r3,r3,r9 @ E+=X[i] |
|
144 eor r10,r10,r7,ror#2 @ F_00_19(B,C,D) |
|
145 str r9,[r14,#-4]! |
|
146 add r3,r3,r10 @ E+=F_00_19(B,C,D) |
|
147 teq r14,sp |
|
148 bne .L_00_15 @ [((11+4)*5+2)*3] |
|
149 sub sp,sp,#25*4 |
|
150 #if __ARM_ARCH__<7 |
|
151 ldrb r10,[r1,#2] |
|
152 ldrb r9,[r1,#3] |
|
153 ldrb r11,[r1,#1] |
|
154 add r7,r8,r7,ror#2 @ E+=K_00_19 |
|
155 ldrb r12,[r1],#4 |
|
156 orr r9,r9,r10,lsl#8 |
|
157 eor r10,r5,r6 @ F_xx_xx |
|
158 orr r9,r9,r11,lsl#16 |
|
159 add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
|
160 orr r9,r9,r12,lsl#24 |
|
161 #else |
|
162 ldr r9,[r1],#4 @ handles unaligned |
|
163 add r7,r8,r7,ror#2 @ E+=K_00_19 |
|
164 eor r10,r5,r6 @ F_xx_xx |
|
165 add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
|
166 #ifdef __ARMEL__ |
|
167 rev r9,r9 @ byte swap |
|
168 #endif |
|
169 #endif |
|
170 and r10,r4,r10,ror#2 |
|
171 add r7,r7,r9 @ E+=X[i] |
|
172 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) |
|
173 str r9,[r14,#-4]! |
|
174 add r7,r7,r10 @ E+=F_00_19(B,C,D) |
|
175 ldr r9,[r14,#15*4] |
|
176 ldr r10,[r14,#13*4] |
|
177 ldr r11,[r14,#7*4] |
|
178 add r6,r8,r6,ror#2 @ E+=K_xx_xx |
|
179 ldr r12,[r14,#2*4] |
|
180 eor r9,r9,r10 |
|
181 eor r11,r11,r12 @ 1 cycle stall |
|
182 eor r10,r4,r5 @ F_xx_xx |
|
183 mov r9,r9,ror#31 |
|
184 add r6,r6,r7,ror#27 @ E+=ROR(A,27) |
|
185 eor r9,r9,r11,ror#31 |
|
186 str r9,[r14,#-4]! |
|
187 and r10,r3,r10,ror#2 @ F_xx_xx |
|
188 @ F_xx_xx |
|
189 add r6,r6,r9 @ E+=X[i] |
|
190 eor r10,r10,r5,ror#2 @ F_00_19(B,C,D) |
|
191 add r6,r6,r10 @ E+=F_00_19(B,C,D) |
|
192 ldr r9,[r14,#15*4] |
|
193 ldr r10,[r14,#13*4] |
|
194 ldr r11,[r14,#7*4] |
|
195 add r5,r8,r5,ror#2 @ E+=K_xx_xx |
|
196 ldr r12,[r14,#2*4] |
|
197 eor r9,r9,r10 |
|
198 eor r11,r11,r12 @ 1 cycle stall |
|
199 eor r10,r3,r4 @ F_xx_xx |
|
200 mov r9,r9,ror#31 |
|
201 add r5,r5,r6,ror#27 @ E+=ROR(A,27) |
|
202 eor r9,r9,r11,ror#31 |
|
203 str r9,[r14,#-4]! |
|
204 and r10,r7,r10,ror#2 @ F_xx_xx |
|
205 @ F_xx_xx |
|
206 add r5,r5,r9 @ E+=X[i] |
|
207 eor r10,r10,r4,ror#2 @ F_00_19(B,C,D) |
|
208 add r5,r5,r10 @ E+=F_00_19(B,C,D) |
|
209 ldr r9,[r14,#15*4] |
|
210 ldr r10,[r14,#13*4] |
|
211 ldr r11,[r14,#7*4] |
|
212 add r4,r8,r4,ror#2 @ E+=K_xx_xx |
|
213 ldr r12,[r14,#2*4] |
|
214 eor r9,r9,r10 |
|
215 eor r11,r11,r12 @ 1 cycle stall |
|
216 eor r10,r7,r3 @ F_xx_xx |
|
217 mov r9,r9,ror#31 |
|
218 add r4,r4,r5,ror#27 @ E+=ROR(A,27) |
|
219 eor r9,r9,r11,ror#31 |
|
220 str r9,[r14,#-4]! |
|
221 and r10,r6,r10,ror#2 @ F_xx_xx |
|
222 @ F_xx_xx |
|
223 add r4,r4,r9 @ E+=X[i] |
|
224 eor r10,r10,r3,ror#2 @ F_00_19(B,C,D) |
|
225 add r4,r4,r10 @ E+=F_00_19(B,C,D) |
|
226 ldr r9,[r14,#15*4] |
|
227 ldr r10,[r14,#13*4] |
|
228 ldr r11,[r14,#7*4] |
|
229 add r3,r8,r3,ror#2 @ E+=K_xx_xx |
|
230 ldr r12,[r14,#2*4] |
|
231 eor r9,r9,r10 |
|
232 eor r11,r11,r12 @ 1 cycle stall |
|
233 eor r10,r6,r7 @ F_xx_xx |
|
234 mov r9,r9,ror#31 |
|
235 add r3,r3,r4,ror#27 @ E+=ROR(A,27) |
|
236 eor r9,r9,r11,ror#31 |
|
237 str r9,[r14,#-4]! |
|
238 and r10,r5,r10,ror#2 @ F_xx_xx |
|
239 @ F_xx_xx |
|
240 add r3,r3,r9 @ E+=X[i] |
|
241 eor r10,r10,r7,ror#2 @ F_00_19(B,C,D) |
|
242 add r3,r3,r10 @ E+=F_00_19(B,C,D) |
|
243 |
|
244 ldr r8,.LK_20_39 @ [+15+16*4] |
|
245 cmn sp,#0 @ [+3], clear carry to denote 20_39 |
|
246 .L_20_39_or_60_79: |
|
247 ldr r9,[r14,#15*4] |
|
248 ldr r10,[r14,#13*4] |
|
249 ldr r11,[r14,#7*4] |
|
250 add r7,r8,r7,ror#2 @ E+=K_xx_xx |
|
251 ldr r12,[r14,#2*4] |
|
252 eor r9,r9,r10 |
|
253 eor r11,r11,r12 @ 1 cycle stall |
|
254 eor r10,r5,r6 @ F_xx_xx |
|
255 mov r9,r9,ror#31 |
|
256 add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
|
257 eor r9,r9,r11,ror#31 |
|
258 str r9,[r14,#-4]! |
|
259 eor r10,r4,r10,ror#2 @ F_xx_xx |
|
260 @ F_xx_xx |
|
261 add r7,r7,r9 @ E+=X[i] |
|
262 add r7,r7,r10 @ E+=F_20_39(B,C,D) |
|
263 ldr r9,[r14,#15*4] |
|
264 ldr r10,[r14,#13*4] |
|
265 ldr r11,[r14,#7*4] |
|
266 add r6,r8,r6,ror#2 @ E+=K_xx_xx |
|
267 ldr r12,[r14,#2*4] |
|
268 eor r9,r9,r10 |
|
269 eor r11,r11,r12 @ 1 cycle stall |
|
270 eor r10,r4,r5 @ F_xx_xx |
|
271 mov r9,r9,ror#31 |
|
272 add r6,r6,r7,ror#27 @ E+=ROR(A,27) |
|
273 eor r9,r9,r11,ror#31 |
|
274 str r9,[r14,#-4]! |
|
275 eor r10,r3,r10,ror#2 @ F_xx_xx |
|
276 @ F_xx_xx |
|
277 add r6,r6,r9 @ E+=X[i] |
|
278 add r6,r6,r10 @ E+=F_20_39(B,C,D) |
|
279 ldr r9,[r14,#15*4] |
|
280 ldr r10,[r14,#13*4] |
|
281 ldr r11,[r14,#7*4] |
|
282 add r5,r8,r5,ror#2 @ E+=K_xx_xx |
|
283 ldr r12,[r14,#2*4] |
|
284 eor r9,r9,r10 |
|
285 eor r11,r11,r12 @ 1 cycle stall |
|
286 eor r10,r3,r4 @ F_xx_xx |
|
287 mov r9,r9,ror#31 |
|
288 add r5,r5,r6,ror#27 @ E+=ROR(A,27) |
|
289 eor r9,r9,r11,ror#31 |
|
290 str r9,[r14,#-4]! |
|
291 eor r10,r7,r10,ror#2 @ F_xx_xx |
|
292 @ F_xx_xx |
|
293 add r5,r5,r9 @ E+=X[i] |
|
294 add r5,r5,r10 @ E+=F_20_39(B,C,D) |
|
295 ldr r9,[r14,#15*4] |
|
296 ldr r10,[r14,#13*4] |
|
297 ldr r11,[r14,#7*4] |
|
298 add r4,r8,r4,ror#2 @ E+=K_xx_xx |
|
299 ldr r12,[r14,#2*4] |
|
300 eor r9,r9,r10 |
|
301 eor r11,r11,r12 @ 1 cycle stall |
|
302 eor r10,r7,r3 @ F_xx_xx |
|
303 mov r9,r9,ror#31 |
|
304 add r4,r4,r5,ror#27 @ E+=ROR(A,27) |
|
305 eor r9,r9,r11,ror#31 |
|
306 str r9,[r14,#-4]! |
|
307 eor r10,r6,r10,ror#2 @ F_xx_xx |
|
308 @ F_xx_xx |
|
309 add r4,r4,r9 @ E+=X[i] |
|
310 add r4,r4,r10 @ E+=F_20_39(B,C,D) |
|
311 ldr r9,[r14,#15*4] |
|
312 ldr r10,[r14,#13*4] |
|
313 ldr r11,[r14,#7*4] |
|
314 add r3,r8,r3,ror#2 @ E+=K_xx_xx |
|
315 ldr r12,[r14,#2*4] |
|
316 eor r9,r9,r10 |
|
317 eor r11,r11,r12 @ 1 cycle stall |
|
318 eor r10,r6,r7 @ F_xx_xx |
|
319 mov r9,r9,ror#31 |
|
320 add r3,r3,r4,ror#27 @ E+=ROR(A,27) |
|
321 eor r9,r9,r11,ror#31 |
|
322 str r9,[r14,#-4]! |
|
323 eor r10,r5,r10,ror#2 @ F_xx_xx |
|
324 @ F_xx_xx |
|
325 add r3,r3,r9 @ E+=X[i] |
|
326 add r3,r3,r10 @ E+=F_20_39(B,C,D) |
|
327 teq r14,sp @ preserve carry |
|
328 bne .L_20_39_or_60_79 @ [+((12+3)*5+2)*4] |
|
329 bcs .L_done @ [+((12+3)*5+2)*4], spare 300 bytes |
|
330 |
|
331 ldr r8,.LK_40_59 |
|
332 sub sp,sp,#20*4 @ [+2] |
|
333 .L_40_59: |
|
334 ldr r9,[r14,#15*4] |
|
335 ldr r10,[r14,#13*4] |
|
336 ldr r11,[r14,#7*4] |
|
337 add r7,r8,r7,ror#2 @ E+=K_xx_xx |
|
338 ldr r12,[r14,#2*4] |
|
339 eor r9,r9,r10 |
|
340 eor r11,r11,r12 @ 1 cycle stall |
|
341 eor r10,r5,r6 @ F_xx_xx |
|
342 mov r9,r9,ror#31 |
|
343 add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
|
344 eor r9,r9,r11,ror#31 |
|
345 str r9,[r14,#-4]! |
|
346 and r10,r4,r10,ror#2 @ F_xx_xx |
|
347 and r11,r5,r6 @ F_xx_xx |
|
348 add r7,r7,r9 @ E+=X[i] |
|
349 add r7,r7,r10 @ E+=F_40_59(B,C,D) |
|
350 add r7,r7,r11,ror#2 |
|
351 ldr r9,[r14,#15*4] |
|
352 ldr r10,[r14,#13*4] |
|
353 ldr r11,[r14,#7*4] |
|
354 add r6,r8,r6,ror#2 @ E+=K_xx_xx |
|
355 ldr r12,[r14,#2*4] |
|
356 eor r9,r9,r10 |
|
357 eor r11,r11,r12 @ 1 cycle stall |
|
358 eor r10,r4,r5 @ F_xx_xx |
|
359 mov r9,r9,ror#31 |
|
360 add r6,r6,r7,ror#27 @ E+=ROR(A,27) |
|
361 eor r9,r9,r11,ror#31 |
|
362 str r9,[r14,#-4]! |
|
363 and r10,r3,r10,ror#2 @ F_xx_xx |
|
364 and r11,r4,r5 @ F_xx_xx |
|
365 add r6,r6,r9 @ E+=X[i] |
|
366 add r6,r6,r10 @ E+=F_40_59(B,C,D) |
|
367 add r6,r6,r11,ror#2 |
|
368 ldr r9,[r14,#15*4] |
|
369 ldr r10,[r14,#13*4] |
|
370 ldr r11,[r14,#7*4] |
|
371 add r5,r8,r5,ror#2 @ E+=K_xx_xx |
|
372 ldr r12,[r14,#2*4] |
|
373 eor r9,r9,r10 |
|
374 eor r11,r11,r12 @ 1 cycle stall |
|
375 eor r10,r3,r4 @ F_xx_xx |
|
376 mov r9,r9,ror#31 |
|
377 add r5,r5,r6,ror#27 @ E+=ROR(A,27) |
|
378 eor r9,r9,r11,ror#31 |
|
379 str r9,[r14,#-4]! |
|
380 and r10,r7,r10,ror#2 @ F_xx_xx |
|
381 and r11,r3,r4 @ F_xx_xx |
|
382 add r5,r5,r9 @ E+=X[i] |
|
383 add r5,r5,r10 @ E+=F_40_59(B,C,D) |
|
384 add r5,r5,r11,ror#2 |
|
385 ldr r9,[r14,#15*4] |
|
386 ldr r10,[r14,#13*4] |
|
387 ldr r11,[r14,#7*4] |
|
388 add r4,r8,r4,ror#2 @ E+=K_xx_xx |
|
389 ldr r12,[r14,#2*4] |
|
390 eor r9,r9,r10 |
|
391 eor r11,r11,r12 @ 1 cycle stall |
|
392 eor r10,r7,r3 @ F_xx_xx |
|
393 mov r9,r9,ror#31 |
|
394 add r4,r4,r5,ror#27 @ E+=ROR(A,27) |
|
395 eor r9,r9,r11,ror#31 |
|
396 str r9,[r14,#-4]! |
|
397 and r10,r6,r10,ror#2 @ F_xx_xx |
|
398 and r11,r7,r3 @ F_xx_xx |
|
399 add r4,r4,r9 @ E+=X[i] |
|
400 add r4,r4,r10 @ E+=F_40_59(B,C,D) |
|
401 add r4,r4,r11,ror#2 |
|
402 ldr r9,[r14,#15*4] |
|
403 ldr r10,[r14,#13*4] |
|
404 ldr r11,[r14,#7*4] |
|
405 add r3,r8,r3,ror#2 @ E+=K_xx_xx |
|
406 ldr r12,[r14,#2*4] |
|
407 eor r9,r9,r10 |
|
408 eor r11,r11,r12 @ 1 cycle stall |
|
409 eor r10,r6,r7 @ F_xx_xx |
|
410 mov r9,r9,ror#31 |
|
411 add r3,r3,r4,ror#27 @ E+=ROR(A,27) |
|
412 eor r9,r9,r11,ror#31 |
|
413 str r9,[r14,#-4]! |
|
414 and r10,r5,r10,ror#2 @ F_xx_xx |
|
415 and r11,r6,r7 @ F_xx_xx |
|
416 add r3,r3,r9 @ E+=X[i] |
|
417 add r3,r3,r10 @ E+=F_40_59(B,C,D) |
|
418 add r3,r3,r11,ror#2 |
|
419 teq r14,sp |
|
420 bne .L_40_59 @ [+((12+5)*5+2)*4] |
|
421 |
|
422 ldr r8,.LK_60_79 |
|
423 sub sp,sp,#20*4 |
|
424 cmp sp,#0 @ set carry to denote 60_79 |
|
425 b .L_20_39_or_60_79 @ [+4], spare 300 bytes |
|
426 .L_done: |
|
427 add sp,sp,#80*4 @ "deallocate" stack frame |
|
428 ldmia r0,{r8,r9,r10,r11,r12} |
|
429 add r3,r8,r3 |
|
430 add r4,r9,r4 |
|
431 add r5,r10,r5,ror#2 |
|
432 add r6,r11,r6,ror#2 |
|
433 add r7,r12,r7,ror#2 |
|
434 stmia r0,{r3,r4,r5,r6,r7} |
|
435 teq r1,r2 |
|
436 bne .Lloop @ [+18], total 1307 |
|
437 |
|
438 #if __ARM_ARCH__>=5 |
|
439 ldmia sp!,{r4-r12,pc} |
|
440 #else |
|
441 ldmia sp!,{r4-r12,lr} |
|
442 tst lr,#1 |
|
443 moveq pc,lr @ be binary compatible with V4, yet |
|
444 .word 0xe12fff1e @ interoperable with Thumb ISA:-) |
|
445 #endif |
|
446 .align 2 |
|
447 .LK_00_19: .word 0x5a827999 |
|
448 .LK_20_39: .word 0x6ed9eba1 |
|
449 .LK_40_59: .word 0x8f1bbcdc |
|
450 .LK_60_79: .word 0xca62c1d6 |
|
451 .size sha1_block_data_order,.-sha1_block_data_order |
|
452 .asciz "SHA1 block transform for ARMv4, CRYPTOGAMS by <[email protected]>" |
|
453 .align 2 |