@@ -22,7 +22,7 @@ static uint32_t Malloc_Heap_Contiguity_Bitmap[MALLOC_HEAP_BITMAP_SIZE];
22
22
23
23
struct mem_pool Memory_Pool = {
24
24
.start_addr = Malloc_Heap ,
25
- .spinlock = {.head = 0 , .tail = 0 },
25
+ .spinlock = {.head = 0U , .tail = 0U },
26
26
.size = CONFIG_HEAP_SIZE ,
27
27
.buff_size = MALLOC_HEAP_BUFF_SIZE ,
28
28
.total_buffs = MALLOC_HEAP_TOTAL_BUFF ,
@@ -46,7 +46,7 @@ static uint32_t Paging_Heap_Contiguity_Bitmap[MALLOC_HEAP_BITMAP_SIZE];
46
46
47
47
struct mem_pool Paging_Memory_Pool = {
48
48
.start_addr = Paging_Heap ,
49
- .spinlock = {.head = 0 , .tail = 0 },
49
+ .spinlock = {.head = 0U , .tail = 0U },
50
50
.size = CONFIG_NUM_ALLOC_PAGES * CPU_PAGE_SIZE ,
51
51
.buff_size = PAGING_HEAP_BUFF_SIZE ,
52
52
.total_buffs = PAGING_HEAP_TOTAL_BUFF ,
@@ -73,7 +73,7 @@ static void *allocate_mem(struct mem_pool *pool, unsigned int num_bytes)
73
73
/* Calculate number of buffers to be allocated from memory pool */
74
74
requested_buffs = INT_DIV_ROUNDUP (num_bytes , pool -> buff_size );
75
75
76
- for (idx = 0 ; idx < pool -> bmp_size ; idx ++ ) {
76
+ for (idx = 0U ; idx < pool -> bmp_size ; idx ++ ) {
77
77
/* Find the first occurrence of requested_buffs number of free
78
78
* buffers. The 0th bit in bitmap represents a free buffer.
79
79
*/
@@ -125,7 +125,7 @@ static void *allocate_mem(struct mem_pool *pool, unsigned int num_bytes)
125
125
/* Update allocation bitmaps information for
126
126
* selected buffers
127
127
*/
128
- for (i = 0 ; i < requested_buffs ; i ++ ) {
128
+ for (i = 0U ; i < requested_buffs ; i ++ ) {
129
129
/* Set allocation bit in bitmap for
130
130
* this buffer
131
131
*/
@@ -157,7 +157,7 @@ static void *allocate_mem(struct mem_pool *pool, unsigned int num_bytes)
157
157
/* Increment idx */
158
158
idx ++ ;
159
159
/* Reset bit_idx */
160
- bit_idx = 0 ;
160
+ bit_idx = 0U ;
161
161
}
162
162
}
163
163
@@ -343,7 +343,7 @@ void *memcpy_s(void *d, size_t dmax, const void *s, size_t slen)
343
343
uint8_t * dest8 ;
344
344
uint8_t * src8 ;
345
345
346
- if (slen == 0 || dmax == 0 || dmax < slen ) {
346
+ if (slen == 0U || dmax == 0U || dmax < slen ) {
347
347
pr_err ("%s: invalid src, dest buffer or length." , __func__ );
348
348
return NULL ;
349
349
}
@@ -362,7 +362,7 @@ void *memcpy_s(void *d, size_t dmax, const void *s, size_t slen)
362
362
src8 = (uint8_t * )s ;
363
363
364
364
/*small data block*/
365
- if (slen < 8 ) {
365
+ if (slen < 8U ) {
366
366
while (slen != 0U ) {
367
367
* dest8 ++ = * src8 ++ ;
368
368
slen -- ;
@@ -378,15 +378,15 @@ void *memcpy_s(void *d, size_t dmax, const void *s, size_t slen)
378
378
}
379
379
380
380
/*copy main data blocks, with rep prefix*/
381
- if (slen > 8 ) {
381
+ if (slen > 8U ) {
382
382
uint32_t ecx ;
383
383
384
384
asm volatile ("cld; rep; movsq"
385
385
: "=&c" (ecx ), "=&D" (dest8 ), "=&S" (src8 )
386
386
: "0" (slen / 8 ), "1" (dest8 ), "2" (src8 )
387
387
: "memory" );
388
388
389
- slen = slen % 8 ;
389
+ slen = slen % 8U ;
390
390
}
391
391
392
392
/*tail bytes*/
@@ -406,7 +406,7 @@ void *memset(void *base, uint8_t v, size_t n)
406
406
407
407
dest_p = (uint8_t * )base ;
408
408
409
- if ((dest_p == NULL ) || (n == 0 ))
409
+ if ((dest_p == NULL ) || (n == 0U ))
410
410
return NULL ;
411
411
412
412
/*do the few bytes to get uint64_t alignment*/
@@ -415,11 +415,11 @@ void *memset(void *base, uint8_t v, size_t n)
415
415
* dest_p ++ = v ;
416
416
417
417
/*64-bit mode*/
418
- n_q = count >> 3 ;
418
+ n_q = count >> 3U ;
419
419
asm volatile ("cld ; rep ; stosq ; movl %3,%%ecx ; rep ; stosb"
420
420
: "+c" (n_q ), "+D" (dest_p )
421
421
: "a" (v * 0x0101010101010101U ),
422
- "r" ((unsigned int )count & 7 ));
422
+ "r" ((unsigned int )count & 7U ));
423
423
424
424
return (void * )dest_p ;
425
425
}
0 commit comments