1/*
2 * M7memset.S: SPARC M7 optimized memset.
3 *
4 * Copyright (c) 2016, Oracle and/or its affiliates. All rights reserved.
5 */
6
7/*
8 * M7memset.S: M7 optimized memset.
9 *
10 * char *memset(sp, c, n)
11 *
12 * Set an array of n chars starting at sp to the character c.
13 * Return sp.
14 *
15 * Fast assembler language version of the following C-program for memset
16 * which represents the `standard' for the C-library.
17 *
18 * void *
19 * memset(void *sp1, int c, size_t n)
20 * {
21 * if (n != 0) {
22 * char *sp = sp1;
23 * do {
24 * *sp++ = (char)c;
25 * } while (--n != 0);
26 * }
27 * return (sp1);
28 * }
29 *
30 * The algorithm is as follows :
31 *
32 * For small 6 or fewer bytes stores, bytes will be stored.
33 *
34 * For less than 32 bytes stores, align the address on 4 byte boundary.
35 * Then store as many 4-byte chunks, followed by trailing bytes.
36 *
37 * For sizes greater than 32 bytes, align the address on 8 byte boundary.
38 * if (count >= 64) {
39 * store 8-bytes chunks to align the address on 64 byte boundary
40 * if (value to be set is zero && count >= MIN_ZERO) {
41 * Using BIS stores, set the first long word of each
42 * 64-byte cache line to zero which will also clear the
43 * other seven long words of the cache line.
44 * }
45 * else if (count >= MIN_LOOP) {
46 * Using BIS stores, set the first long word of each of
47 * ST_CHUNK cache lines (64 bytes each) before the main
48 * loop is entered.
49 * In the main loop, continue pre-setting the first long
50 * word of each cache line ST_CHUNK lines in advance while
51 * setting the other seven long words (56 bytes) of each
52 * cache line until fewer than ST_CHUNK*64 bytes remain.
53 * Then set the remaining seven long words of each cache
54 * line that has already had its first long word set.
55 * }
56 * store remaining data in 64-byte chunks until less than
57 * 64 bytes remain.
58 * }
59 * Store as many 8-byte chunks, followed by trailing bytes.
60 *
61 * BIS = Block Init Store
62 * Doing the advance store of the first element of the cache line
63 * initiates the displacement of a cache line while only using a single
64 * instruction in the pipeline. That avoids various pipeline delays,
65 * such as filling the miss buffer. The performance effect is
66 * similar to prefetching for normal stores.
67 * The special case for zero fills runs faster and uses fewer instruction
68 * cycles than the normal memset loop.
69 *
70 * We only use BIS for memset of greater than MIN_LOOP bytes because a sequence
71 * BIS stores must be followed by a membar #StoreStore. The benefit of
72 * the BIS store must be balanced against the cost of the membar operation.
73 */
74
75/*
76 * ASI_STBI_P marks the cache line as "least recently used"
77 * which means if many threads are active, it has a high chance
78 * of being pushed out of the cache between the first initializing
79 * store and the final stores.
80 * Thus, we use ASI_STBIMRU_P which marks the cache line as
81 * "most recently used" for all but the last store to the cache line.
82 */
83
84#include <asm/asi.h>
85#include <asm/page.h>
86
87#define ASI_STBI_P ASI_BLK_INIT_QUAD_LDD_P
88#define ASI_STBIMRU_P ASI_ST_BLKINIT_MRU_P
89
90
91#define ST_CHUNK 24 /* multiple of 4 due to loop unrolling */
92#define MIN_LOOP 16320
93#define MIN_ZERO 512
94
95 .section ".text"
96 .align 32
97
98/*
99 * Define clear_page(dest) as memset(dest, 0, PAGE_SIZE)
100 * (can create a more optimized version later.)
101 */
102 .globl M7clear_page
103 .globl M7clear_user_page
104M7clear_page: /* clear_page(dest) */
105M7clear_user_page:
106 set PAGE_SIZE, %o1
107 /* fall through into bzero code */
108
109 .size M7clear_page,.-M7clear_page
110 .size M7clear_user_page,.-M7clear_user_page
111
112/*
113 * Define bzero(dest, n) as memset(dest, 0, n)
114 * (can create a more optimized version later.)
115 */
116 .globl M7bzero
117M7bzero: /* bzero(dest, size) */
118 mov %o1, %o2
119 mov 0, %o1
120 /* fall through into memset code */
121
122 .size M7bzero,.-M7bzero
123
124 .global M7memset
125 .type M7memset, #function
126 .register %g3, #scratch
127M7memset:
128 mov %o0, %o5 ! copy sp1 before using it
129 cmp %o2, 7 ! if small counts, just write bytes
130 bleu,pn %xcc, .wrchar
131 and %o1, 0xff, %o1 ! o1 is (char)c
132
133 sll %o1, 8, %o3
134 or %o1, %o3, %o1 ! now o1 has 2 bytes of c
135 sll %o1, 16, %o3
136 cmp %o2, 32
137 blu,pn %xcc, .wdalign
138 or %o1, %o3, %o1 ! now o1 has 4 bytes of c
139
140 sllx %o1, 32, %o3
141 or %o1, %o3, %o1 ! now o1 has 8 bytes of c
142
143.dbalign:
144 andcc %o5, 7, %o3 ! is sp1 aligned on a 8 byte bound?
145 bz,pt %xcc, .blkalign ! already long word aligned
146 sub %o3, 8, %o3 ! -(bytes till long word aligned)
147
148 add %o2, %o3, %o2 ! update o2 with new count
149 ! Set -(%o3) bytes till sp1 long word aligned
1501: stb %o1, [%o5] ! there is at least 1 byte to set
151 inccc %o3 ! byte clearing loop
152 bl,pt %xcc, 1b
153 inc %o5
154
155 ! Now sp1 is long word aligned (sp1 is found in %o5)
156.blkalign:
157 cmp %o2, 64 ! check if there are 64 bytes to set
158 blu,pn %xcc, .wrshort
159 mov %o2, %o3
160
161 andcc %o5, 63, %o3 ! is sp1 block aligned?
162 bz,pt %xcc, .blkwr ! now block aligned
163 sub %o3, 64, %o3 ! o3 is -(bytes till block aligned)
164 add %o2, %o3, %o2 ! o2 is the remainder
165
166 ! Store -(%o3) bytes till dst is block (64 byte) aligned.
167 ! Use long word stores.
168 ! Recall that dst is already long word aligned
1691:
170 addcc %o3, 8, %o3
171 stx %o1, [%o5]
172 bl,pt %xcc, 1b
173 add %o5, 8, %o5
174
175 ! Now sp1 is block aligned
176.blkwr:
177 andn %o2, 63, %o4 ! calculate size of blocks in bytes
178 brz,pn %o1, .wrzero ! special case if c == 0
179 and %o2, 63, %o3 ! %o3 = bytes left after blk stores.
180
181 set MIN_LOOP, %g1
182 cmp %o4, %g1 ! check there are enough bytes to set
183 blu,pn %xcc, .short_set ! to justify cost of membar
184 ! must be > pre-cleared lines
185 nop
186
187 ! initial cache-clearing stores
188 ! get store pipeline moving
189 rd %asi, %g3 ! save %asi to be restored later
190 wr %g0, ASI_STBIMRU_P, %asi
191
192 ! Primary memset loop for large memsets
193.wr_loop:
194 sub %o5, 8, %o5 ! adjust %o5 for ASI store alignment
195 mov ST_CHUNK, %g1
196.wr_loop_start:
197 stxa %o1, [%o5+8]%asi
198 subcc %g1, 4, %g1
199 stxa %o1, [%o5+8+64]%asi
200 add %o5, 256, %o5
201 stxa %o1, [%o5+8-128]%asi
202 bgu %xcc, .wr_loop_start
203 stxa %o1, [%o5+8-64]%asi
204
205 sub %o5, ST_CHUNK*64, %o5 ! reset %o5
206 mov ST_CHUNK, %g1
207
208.wr_loop_rest:
209 stxa %o1, [%o5+8+8]%asi
210 sub %o4, 64, %o4
211 stxa %o1, [%o5+16+8]%asi
212 subcc %g1, 1, %g1
213 stxa %o1, [%o5+24+8]%asi
214 stxa %o1, [%o5+32+8]%asi
215 stxa %o1, [%o5+40+8]%asi
216 add %o5, 64, %o5
217 stxa %o1, [%o5-8]%asi
218 bgu %xcc, .wr_loop_rest
219 stxa %o1, [%o5]ASI_STBI_P
220
221 ! If more than ST_CHUNK*64 bytes remain to set, continue
222 ! setting the first long word of each cache line in advance
223 ! to keep the store pipeline moving.
224
225 cmp %o4, ST_CHUNK*64
226 bge,pt %xcc, .wr_loop_start
227 mov ST_CHUNK, %g1
228
229 brz,a,pn %o4, .asi_done
230 add %o5, 8, %o5 ! restore %o5 offset
231
232.wr_loop_small:
233 stxa %o1, [%o5+8]%asi
234 stxa %o1, [%o5+8+8]%asi
235 stxa %o1, [%o5+16+8]%asi
236 stxa %o1, [%o5+24+8]%asi
237 stxa %o1, [%o5+32+8]%asi
238 subcc %o4, 64, %o4
239 stxa %o1, [%o5+40+8]%asi
240 add %o5, 64, %o5
241 stxa %o1, [%o5-8]%asi
242 bgu,pt %xcc, .wr_loop_small
243 stxa %o1, [%o5]ASI_STBI_P
244
245 ba .asi_done
246 add %o5, 8, %o5 ! restore %o5 offset
247
248 ! Special case loop for zero fill memsets
249 ! For each 64 byte cache line, single STBI to first element
250 ! clears line
251.wrzero:
252 cmp %o4, MIN_ZERO ! check if enough bytes to set
253 ! to pay %asi + membar cost
254 blu %xcc, .short_set
255 nop
256 sub %o4, 256, %o4
257
258.wrzero_loop:
259 mov 64, %g3
260 stxa %o1, [%o5]ASI_STBI_P
261 subcc %o4, 256, %o4
262 stxa %o1, [%o5+%g3]ASI_STBI_P
263 add %o5, 256, %o5
264 sub %g3, 192, %g3
265 stxa %o1, [%o5+%g3]ASI_STBI_P
266 add %g3, 64, %g3
267 bge,pt %xcc, .wrzero_loop
268 stxa %o1, [%o5+%g3]ASI_STBI_P
269 add %o4, 256, %o4
270
271 brz,pn %o4, .bsi_done
272 nop
273
274.wrzero_small:
275 stxa %o1, [%o5]ASI_STBI_P
276 subcc %o4, 64, %o4
277 bgu,pt %xcc, .wrzero_small
278 add %o5, 64, %o5
279 ba,a .bsi_done
280
281.asi_done:
282 wr %g3, 0x0, %asi ! restored saved %asi
283.bsi_done:
284 membar #StoreStore ! required by use of Block Store Init
285
286.short_set:
287 cmp %o4, 64 ! check if 64 bytes to set
288 blu %xcc, 5f
289 nop
2904: ! set final blocks of 64 bytes
291 stx %o1, [%o5]
292 stx %o1, [%o5+8]
293 stx %o1, [%o5+16]
294 stx %o1, [%o5+24]
295 subcc %o4, 64, %o4
296 stx %o1, [%o5+32]
297 stx %o1, [%o5+40]
298 add %o5, 64, %o5
299 stx %o1, [%o5-16]
300 bgu,pt %xcc, 4b
301 stx %o1, [%o5-8]
302
3035:
304 ! Set the remaining long words
305.wrshort:
306 subcc %o3, 8, %o3 ! Can we store any long words?
307 blu,pn %xcc, .wrchars
308 and %o2, 7, %o2 ! calc bytes left after long words
3096:
310 subcc %o3, 8, %o3
311 stx %o1, [%o5] ! store the long words
312 bgeu,pt %xcc, 6b
313 add %o5, 8, %o5
314
315.wrchars: ! check for extra chars
316 brnz %o2, .wrfin
317 nop
318 retl
319 nop
320
321.wdalign:
322 andcc %o5, 3, %o3 ! is sp1 aligned on a word boundary
323 bz,pn %xcc, .wrword
324 andn %o2, 3, %o3 ! create word sized count in %o3
325
326 dec %o2 ! decrement count
327 stb %o1, [%o5] ! clear a byte
328 b .wdalign
329 inc %o5 ! next byte
330
331.wrword:
332 subcc %o3, 4, %o3
333 st %o1, [%o5] ! 4-byte writing loop
334 bnz,pt %xcc, .wrword
335 add %o5, 4, %o5
336
337 and %o2, 3, %o2 ! leftover count, if any
338
339.wrchar:
340 ! Set the remaining bytes, if any
341 brz %o2, .exit
342 nop
343.wrfin:
344 deccc %o2
345 stb %o1, [%o5]
346 bgu,pt %xcc, .wrfin
347 inc %o5
348.exit:
349 retl ! %o0 was preserved
350 nop
351
352 .size M7memset,.-M7memset
353

source code of linux/arch/sparc/lib/M7memset.S