@@ -113,8 +113,8 @@ handle_virt_irqline(struct vm *vm, uint16_t target_vmid,
113
113
/* Call vpic for pic injection */
114
114
ret = handle_vpic_irqline (target_vm , param -> pic_irq , mode );
115
115
116
- /* call vioapic for ioapic injection if ioapic_irq != ~0UL */
117
- if (param -> ioapic_irq != (~0UL )) {
116
+ /* call vioapic for ioapic injection if ioapic_irq != ~0U */
117
+ if (param -> ioapic_irq != (~0U )) {
118
118
/* handle IOAPIC irqline */
119
119
ret = handle_vioapic_irqline (target_vm ,
120
120
param -> ioapic_irq , mode );
@@ -172,7 +172,7 @@ int32_t hcall_create_vm(struct vm *vm, uint64_t param)
172
172
return ret ;
173
173
}
174
174
175
- int32_t hcall_destroy_vm (uint64_t vmid )
175
+ int32_t hcall_destroy_vm (uint16_t vmid )
176
176
{
177
177
int32_t ret = 0 ;
178
178
struct vm * target_vm = get_vm_from_vmid (vmid );
@@ -185,7 +185,7 @@ int32_t hcall_destroy_vm(uint64_t vmid)
185
185
return ret ;
186
186
}
187
187
188
- int32_t hcall_resume_vm (uint64_t vmid )
188
+ int32_t hcall_resume_vm (uint16_t vmid )
189
189
{
190
190
int32_t ret = 0 ;
191
191
struct vm * target_vm = get_vm_from_vmid (vmid );
@@ -202,7 +202,7 @@ int32_t hcall_resume_vm(uint64_t vmid)
202
202
return ret ;
203
203
}
204
204
205
- int32_t hcall_pause_vm (uint64_t vmid )
205
+ int32_t hcall_pause_vm (uint16_t vmid )
206
206
{
207
207
struct vm * target_vm = get_vm_from_vmid (vmid );
208
208
@@ -215,7 +215,7 @@ int32_t hcall_pause_vm(uint64_t vmid)
215
215
return 0 ;
216
216
}
217
217
218
- int32_t hcall_create_vcpu (struct vm * vm , uint64_t vmid , uint64_t param )
218
+ int32_t hcall_create_vcpu (struct vm * vm , uint16_t vmid , uint64_t param )
219
219
{
220
220
int32_t ret ;
221
221
uint16_t pcpu_id ;
@@ -242,7 +242,7 @@ int32_t hcall_create_vcpu(struct vm *vm, uint64_t vmid, uint64_t param)
242
242
return ret ;
243
243
}
244
244
245
- int32_t hcall_assert_irqline (struct vm * vm , uint64_t vmid , uint64_t param )
245
+ int32_t hcall_assert_irqline (struct vm * vm , uint16_t vmid , uint64_t param )
246
246
{
247
247
int32_t ret = 0 ;
248
248
struct acrn_irqline irqline ;
@@ -251,12 +251,12 @@ int32_t hcall_assert_irqline(struct vm *vm, uint64_t vmid, uint64_t param)
251
251
pr_err ("%s: Unable copy param to vm\n" , __func__ );
252
252
return -1 ;
253
253
}
254
- ret = handle_virt_irqline (vm , ( uint16_t ) vmid , & irqline , IRQ_ASSERT );
254
+ ret = handle_virt_irqline (vm , vmid , & irqline , IRQ_ASSERT );
255
255
256
256
return ret ;
257
257
}
258
258
259
- int32_t hcall_deassert_irqline (struct vm * vm , uint64_t vmid , uint64_t param )
259
+ int32_t hcall_deassert_irqline (struct vm * vm , uint16_t vmid , uint64_t param )
260
260
{
261
261
int32_t ret = 0 ;
262
262
struct acrn_irqline irqline ;
@@ -265,12 +265,12 @@ int32_t hcall_deassert_irqline(struct vm *vm, uint64_t vmid, uint64_t param)
265
265
pr_err ("%s: Unable copy param to vm\n" , __func__ );
266
266
return -1 ;
267
267
}
268
- ret = handle_virt_irqline (vm , ( uint16_t ) vmid , & irqline , IRQ_DEASSERT );
268
+ ret = handle_virt_irqline (vm , vmid , & irqline , IRQ_DEASSERT );
269
269
270
270
return ret ;
271
271
}
272
272
273
- int32_t hcall_pulse_irqline (struct vm * vm , uint64_t vmid , uint64_t param )
273
+ int32_t hcall_pulse_irqline (struct vm * vm , uint16_t vmid , uint64_t param )
274
274
{
275
275
int32_t ret = 0 ;
276
276
struct acrn_irqline irqline ;
@@ -279,12 +279,12 @@ int32_t hcall_pulse_irqline(struct vm *vm, uint64_t vmid, uint64_t param)
279
279
pr_err ("%s: Unable copy param to vm\n" , __func__ );
280
280
return -1 ;
281
281
}
282
- ret = handle_virt_irqline (vm , ( uint16_t ) vmid , & irqline , IRQ_PULSE );
282
+ ret = handle_virt_irqline (vm , vmid , & irqline , IRQ_PULSE );
283
283
284
284
return ret ;
285
285
}
286
286
287
- int32_t hcall_inject_msi (struct vm * vm , uint64_t vmid , uint64_t param )
287
+ int32_t hcall_inject_msi (struct vm * vm , uint16_t vmid , uint64_t param )
288
288
{
289
289
int32_t ret = 0 ;
290
290
struct acrn_msi_entry msi ;
@@ -304,7 +304,7 @@ int32_t hcall_inject_msi(struct vm *vm, uint64_t vmid, uint64_t param)
304
304
return ret ;
305
305
}
306
306
307
- int32_t hcall_set_ioreq_buffer (struct vm * vm , uint64_t vmid , uint64_t param )
307
+ int32_t hcall_set_ioreq_buffer (struct vm * vm , uint16_t vmid , uint64_t param )
308
308
{
309
309
int32_t ret = 0 ;
310
310
uint64_t hpa = 0UL ;
@@ -370,7 +370,7 @@ static void complete_request(struct vcpu *vcpu)
370
370
resume_vcpu (vcpu );
371
371
}
372
372
373
- int32_t hcall_notify_req_finish (uint64_t vmid , uint64_t vcpu_id )
373
+ int32_t hcall_notify_req_finish (uint16_t vmid , uint16_t vcpu_id )
374
374
{
375
375
union vhm_request_buffer * req_buf ;
376
376
struct vhm_request * req ;
@@ -386,7 +386,7 @@ int32_t hcall_notify_req_finish(uint64_t vmid, uint64_t vcpu_id)
386
386
dev_dbg (ACRN_DBG_HYCALL , "[%d] NOTIFY_FINISH for vcpu %d" ,
387
387
vmid , vcpu_id );
388
388
389
- vcpu = vcpu_from_vid (target_vm , ( uint16_t ) vcpu_id );
389
+ vcpu = vcpu_from_vid (target_vm , vcpu_id );
390
390
if (vcpu == NULL ) {
391
391
pr_err ("%s, failed to get VCPU %d context from VM %d\n" ,
392
392
__func__ , vcpu_id , target_vm -> attr .id );
@@ -410,7 +410,7 @@ _set_vm_memmap(struct vm *vm, struct vm *target_vm,
410
410
struct vm_set_memmap * memmap )
411
411
{
412
412
uint64_t hpa , base_paddr ;
413
- uint32_t attr , prot ;
413
+ uint64_t attr , prot ;
414
414
415
415
if ((memmap -> length & 0xFFFUL ) != 0UL ) {
416
416
pr_err ("%s: ERROR! [vm%d] map size 0x%x is not page aligned" ,
@@ -462,7 +462,7 @@ _set_vm_memmap(struct vm *vm, struct vm *target_vm,
462
462
memmap -> remote_gpa , memmap -> length , memmap -> type , attr );
463
463
}
464
464
465
- int32_t hcall_set_vm_memmap (struct vm * vm , uint64_t vmid , uint64_t param )
465
+ int32_t hcall_set_vm_memmap (struct vm * vm , uint16_t vmid , uint64_t param )
466
466
{
467
467
struct vm_set_memmap memmap ;
468
468
struct vm * target_vm = get_vm_from_vmid (vmid );
@@ -534,7 +534,7 @@ int32_t hcall_set_vm_memmaps(struct vm *vm, uint64_t param)
534
534
return 0 ;
535
535
}
536
536
537
- int32_t hcall_remap_pci_msix (struct vm * vm , uint64_t vmid , uint64_t param )
537
+ int32_t hcall_remap_pci_msix (struct vm * vm , uint16_t vmid , uint64_t param )
538
538
{
539
539
int32_t ret = 0 ;
540
540
struct acrn_vm_pci_msix_remap remap ;
@@ -575,7 +575,7 @@ int32_t hcall_remap_pci_msix(struct vm *vm, uint64_t vmid, uint64_t param)
575
575
return ret ;
576
576
}
577
577
578
- int32_t hcall_gpa_to_hpa (struct vm * vm , uint64_t vmid , uint64_t param )
578
+ int32_t hcall_gpa_to_hpa (struct vm * vm , uint16_t vmid , uint64_t param )
579
579
{
580
580
int32_t ret = 0 ;
581
581
struct vm_gpa2hpa v_gpa2hpa ;
@@ -600,7 +600,7 @@ int32_t hcall_gpa_to_hpa(struct vm *vm, uint64_t vmid, uint64_t param)
600
600
return ret ;
601
601
}
602
602
603
- int32_t hcall_assign_ptdev (struct vm * vm , uint64_t vmid , uint64_t param )
603
+ int32_t hcall_assign_ptdev (struct vm * vm , uint16_t vmid , uint64_t param )
604
604
{
605
605
int32_t ret ;
606
606
uint16_t bdf ;
@@ -638,7 +638,7 @@ int32_t hcall_assign_ptdev(struct vm *vm, uint64_t vmid, uint64_t param)
638
638
return ret ;
639
639
}
640
640
641
- int32_t hcall_deassign_ptdev (struct vm * vm , uint64_t vmid , uint64_t param )
641
+ int32_t hcall_deassign_ptdev (struct vm * vm , uint16_t vmid , uint64_t param )
642
642
{
643
643
int32_t ret = 0 ;
644
644
uint16_t bdf ;
@@ -658,7 +658,7 @@ int32_t hcall_deassign_ptdev(struct vm *vm, uint64_t vmid, uint64_t param)
658
658
return ret ;
659
659
}
660
660
661
- int32_t hcall_set_ptdev_intr_info (struct vm * vm , uint64_t vmid , uint64_t param )
661
+ int32_t hcall_set_ptdev_intr_info (struct vm * vm , uint16_t vmid , uint64_t param )
662
662
{
663
663
int32_t ret = 0 ;
664
664
struct hc_ptdev_irq irq ;
@@ -693,7 +693,7 @@ int32_t hcall_set_ptdev_intr_info(struct vm *vm, uint64_t vmid, uint64_t param)
693
693
}
694
694
695
695
int32_t
696
- hcall_reset_ptdev_intr_info (struct vm * vm , uint64_t vmid , uint64_t param )
696
+ hcall_reset_ptdev_intr_info (struct vm * vm , uint16_t vmid , uint64_t param )
697
697
{
698
698
int32_t ret = 0 ;
699
699
struct hc_ptdev_irq irq ;
@@ -752,7 +752,7 @@ int32_t hcall_get_cpu_pm_state(struct vm *vm, uint64_t cmd, uint64_t param)
752
752
uint16_t target_vm_id ;
753
753
struct vm * target_vm ;
754
754
755
- target_vm_id = (cmd & PMCMD_VMID_MASK ) >> PMCMD_VMID_SHIFT ;
755
+ target_vm_id = (uint16_t )(( cmd & PMCMD_VMID_MASK ) >> PMCMD_VMID_SHIFT ) ;
756
756
target_vm = get_vm_from_vmid (target_vm_id );
757
757
758
758
if (target_vm == NULL ) {
@@ -820,7 +820,8 @@ int32_t hcall_get_cpu_pm_state(struct vm *vm, uint64_t cmd, uint64_t param)
820
820
return -1 ;
821
821
}
822
822
823
- cx_idx = (cmd & PMCMD_STATE_NUM_MASK ) >> PMCMD_STATE_NUM_SHIFT ;
823
+ cx_idx = (uint8_t )
824
+ ((cmd & PMCMD_STATE_NUM_MASK ) >> PMCMD_STATE_NUM_SHIFT );
824
825
if ((cx_idx == 0U ) || (cx_idx > target_vm -> pm .cx_cnt )) {
825
826
return -1 ;
826
827
}
0 commit comments