Skip to content

Commit

Permalink
x86/shadow: shadow_table[] needs only one entry for PV-only configs
Browse files Browse the repository at this point in the history
Furthermore the field isn't needed at all with shadow support disabled -
move it into struct shadow_vcpu.

Introduce for_each_shadow_table(), shortening loops for the 4-level case
at the same time.

Adjust loop variables and a function parameter to be "unsigned int"
where applicable at the same time. Also move a comment that ended up
misplaced due to incremental additions.

Signed-off-by: Jan Beulich <jbeulich@suse.com>
Acked-by: Tim Deegan <tim@xen.org>
  • Loading branch information
jbeulich committed Jul 21, 2020
1 parent ded576c commit ef3b0d8
Show file tree
Hide file tree
Showing 3 changed files with 78 additions and 55 deletions.
20 changes: 12 additions & 8 deletions xen/arch/x86/mm/shadow/common.c
Original file line number Diff line number Diff line change
Expand Up @@ -959,13 +959,15 @@ static void _shadow_prealloc(struct domain *d, unsigned int pages)
perfc_incr(shadow_prealloc_2);

for_each_vcpu(d, v)
for ( i = 0 ; i < 4 ; i++ )
for ( i = 0; i < ARRAY_SIZE(v->arch.paging.shadow.shadow_table); i++ )
{
if ( !pagetable_is_null(v->arch.shadow_table[i]) )
if ( !pagetable_is_null(v->arch.paging.shadow.shadow_table[i]) )
{
TRACE_SHADOW_PATH_FLAG(TRCE_SFLAG_PREALLOC_UNHOOK);
shadow_unhook_mappings(d,
pagetable_get_mfn(v->arch.shadow_table[i]), 0);
shadow_unhook_mappings(
d,
pagetable_get_mfn(v->arch.paging.shadow.shadow_table[i]),
0);

/* See if that freed up enough space */
if ( d->arch.paging.shadow.free_pages >= pages )
Expand Down Expand Up @@ -1018,10 +1020,12 @@ void shadow_blow_tables(struct domain *d)

/* Second pass: unhook entries of in-use shadows */
for_each_vcpu(d, v)
for ( i = 0 ; i < 4 ; i++ )
if ( !pagetable_is_null(v->arch.shadow_table[i]) )
shadow_unhook_mappings(d,
pagetable_get_mfn(v->arch.shadow_table[i]), 0);
for ( i = 0; i < ARRAY_SIZE(v->arch.paging.shadow.shadow_table); i++ )
if ( !pagetable_is_null(v->arch.paging.shadow.shadow_table[i]) )
shadow_unhook_mappings(
d,
pagetable_get_mfn(v->arch.paging.shadow.shadow_table[i]),
0);

/* Make sure everyone sees the unshadowings */
guest_flush_tlb_mask(d, d->dirty_cpumask);
Expand Down
97 changes: 53 additions & 44 deletions xen/arch/x86/mm/shadow/multi.c
Original file line number Diff line number Diff line change
Expand Up @@ -85,6 +85,15 @@ const char *const fetch_type_names[] = {
};
#endif

#if SHADOW_PAGING_LEVELS == 3
# define for_each_shadow_table(v, i) \
for ( (i) = 0; \
(i) < ARRAY_SIZE((v)->arch.paging.shadow.shadow_table); \
++(i) )
#else
# define for_each_shadow_table(v, i) for ( (i) = 0; (i) < 1; ++(i) )
#endif

/* Helper to perform a local TLB flush. */
static void sh_flush_local(const struct domain *d)
{
Expand Down Expand Up @@ -1624,7 +1633,7 @@ static shadow_l4e_t * shadow_get_and_create_l4e(struct vcpu *v,
mfn_t *sl4mfn)
{
/* There is always a shadow of the top level table. Get it. */
*sl4mfn = pagetable_get_mfn(v->arch.shadow_table[0]);
*sl4mfn = pagetable_get_mfn(v->arch.paging.shadow.shadow_table[0]);
/* Reading the top level table is always valid. */
return sh_linear_l4_table(v) + shadow_l4_linear_offset(gw->va);
}
Expand Down Expand Up @@ -1740,7 +1749,7 @@ static shadow_l2e_t * shadow_get_and_create_l2e(struct vcpu *v,
return sh_linear_l2_table(v) + shadow_l2_linear_offset(gw->va);
#else /* 32bit... */
/* There is always a shadow of the top level table. Get it. */
*sl2mfn = pagetable_get_mfn(v->arch.shadow_table[0]);
*sl2mfn = pagetable_get_mfn(v->arch.paging.shadow.shadow_table[0]);
/* This next line is important: the guest l2 has a 16k
* shadow, we need to return the right mfn of the four. This
* call will set it for us as a side-effect. */
Expand Down Expand Up @@ -2333,6 +2342,7 @@ int sh_safe_not_to_sync(struct vcpu *v, mfn_t gl1mfn)
struct domain *d = v->domain;
struct page_info *sp;
mfn_t smfn;
unsigned int i;

if ( !sh_type_has_up_pointer(d, SH_type_l1_shadow) )
return 0;
Expand Down Expand Up @@ -2365,14 +2375,10 @@ int sh_safe_not_to_sync(struct vcpu *v, mfn_t gl1mfn)
ASSERT(mfn_valid(smfn));
#endif

if ( pagetable_get_pfn(v->arch.shadow_table[0]) == mfn_x(smfn)
#if (SHADOW_PAGING_LEVELS == 3)
|| pagetable_get_pfn(v->arch.shadow_table[1]) == mfn_x(smfn)
|| pagetable_get_pfn(v->arch.shadow_table[2]) == mfn_x(smfn)
|| pagetable_get_pfn(v->arch.shadow_table[3]) == mfn_x(smfn)
#endif
)
return 0;
for_each_shadow_table(v, i)
if ( pagetable_get_pfn(v->arch.paging.shadow.shadow_table[i]) ==
mfn_x(smfn) )
return 0;

/* Only in use in one toplevel shadow, and it's not the one we're
* running on */
Expand Down Expand Up @@ -3287,10 +3293,12 @@ static int sh_page_fault(struct vcpu *v,
for_each_vcpu(d, tmp)
{
#if GUEST_PAGING_LEVELS == 3
int i;
for ( i = 0; i < 4; i++ )
unsigned int i;

for_each_shadow_table(v, i)
{
mfn_t smfn = pagetable_get_mfn(v->arch.shadow_table[i]);
mfn_t smfn = pagetable_get_mfn(
v->arch.paging.shadow.shadow_table[i]);

if ( mfn_valid(smfn) && (mfn_x(smfn) != 0) )
{
Expand Down Expand Up @@ -3707,7 +3715,7 @@ sh_update_linear_entries(struct vcpu *v)
*
* Because HVM guests run on the same monitor tables regardless of the
* shadow tables in use, the linear mapping of the shadow tables has to
* be updated every time v->arch.shadow_table changes.
* be updated every time v->arch.paging.shadow.shadow_table changes.
*/

/* Don't try to update the monitor table if it doesn't exist */
Expand All @@ -3723,17 +3731,19 @@ sh_update_linear_entries(struct vcpu *v)
if ( v == current )
{
__linear_l4_table[l4_linear_offset(SH_LINEAR_PT_VIRT_START)] =
l4e_from_pfn(pagetable_get_pfn(v->arch.shadow_table[0]),
__PAGE_HYPERVISOR_RW);
l4e_from_pfn(
pagetable_get_pfn(v->arch.paging.shadow.shadow_table[0]),
__PAGE_HYPERVISOR_RW);
}
else
{
l4_pgentry_t *ml4e;

ml4e = map_domain_page(pagetable_get_mfn(v->arch.hvm.monitor_table));
ml4e[l4_table_offset(SH_LINEAR_PT_VIRT_START)] =
l4e_from_pfn(pagetable_get_pfn(v->arch.shadow_table[0]),
__PAGE_HYPERVISOR_RW);
l4e_from_pfn(
pagetable_get_pfn(v->arch.paging.shadow.shadow_table[0]),
__PAGE_HYPERVISOR_RW);
unmap_domain_page(ml4e);
}

Expand Down Expand Up @@ -3812,46 +3822,42 @@ sh_update_linear_entries(struct vcpu *v)


/*
* Removes vcpu->arch.shadow_table[].
* Removes v->arch.paging.shadow.shadow_table[].
* Does all appropriate management/bookkeeping/refcounting/etc...
*/
static void
sh_detach_old_tables(struct vcpu *v)
{
struct domain *d = v->domain;
mfn_t smfn;
int i = 0;
unsigned int i;

////
//// vcpu->arch.shadow_table[]
//// vcpu->arch.paging.shadow.shadow_table[]
////

#if GUEST_PAGING_LEVELS == 3
/* PAE guests have four shadow_table entries */
for ( i = 0 ; i < 4 ; i++ )
#endif
for_each_shadow_table(v, i)
{
smfn = pagetable_get_mfn(v->arch.shadow_table[i]);
smfn = pagetable_get_mfn(v->arch.paging.shadow.shadow_table[i]);
if ( mfn_x(smfn) )
sh_put_ref(d, smfn, 0);
v->arch.shadow_table[i] = pagetable_null();
v->arch.paging.shadow.shadow_table[i] = pagetable_null();
}
}

/* Set up the top-level shadow and install it in slot 'slot' of shadow_table */
static void
sh_set_toplevel_shadow(struct vcpu *v,
int slot,
unsigned int slot,
mfn_t gmfn,
unsigned int root_type)
{
mfn_t smfn;
pagetable_t old_entry, new_entry;

struct domain *d = v->domain;

/* Remember the old contents of this slot */
old_entry = v->arch.shadow_table[slot];
old_entry = v->arch.paging.shadow.shadow_table[slot];

/* Now figure out the new contents: is this a valid guest MFN? */
if ( !mfn_valid(gmfn) )
Expand Down Expand Up @@ -3893,7 +3899,7 @@ sh_set_toplevel_shadow(struct vcpu *v,
SHADOW_PRINTK("%u/%u [%u] gmfn %#"PRI_mfn" smfn %#"PRI_mfn"\n",
GUEST_PAGING_LEVELS, SHADOW_PAGING_LEVELS, slot,
mfn_x(gmfn), mfn_x(pagetable_get_mfn(new_entry)));
v->arch.shadow_table[slot] = new_entry;
v->arch.paging.shadow.shadow_table[slot] = new_entry;

/* Decrement the refcount of the old contents of this slot */
if ( !pagetable_is_null(old_entry) ) {
Expand Down Expand Up @@ -3999,7 +4005,7 @@ sh_update_cr3(struct vcpu *v, int do_locking, bool noflush)


////
//// vcpu->arch.shadow_table[]
//// vcpu->arch.paging.shadow.shadow_table[]
////

/* We revoke write access to the new guest toplevel page(s) before we
Expand Down Expand Up @@ -4056,7 +4062,7 @@ sh_update_cr3(struct vcpu *v, int do_locking, bool noflush)
sh_set_toplevel_shadow(v, 0, gmfn, SH_type_l4_shadow);
if ( !shadow_mode_external(d) && !is_pv_32bit_domain(d) )
{
mfn_t smfn = pagetable_get_mfn(v->arch.shadow_table[0]);
mfn_t smfn = pagetable_get_mfn(v->arch.paging.shadow.shadow_table[0]);

if ( !(v->arch.flags & TF_kernel_mode) && VM_ASSIST(d, m2p_strict) )
zap_ro_mpt(smfn);
Expand All @@ -4074,17 +4080,18 @@ sh_update_cr3(struct vcpu *v, int do_locking, bool noflush)
///
#if SHADOW_PAGING_LEVELS == 3
{
mfn_t smfn = pagetable_get_mfn(v->arch.shadow_table[0]);
int i;
for ( i = 0; i < 4; i++ )
mfn_t smfn = pagetable_get_mfn(v->arch.paging.shadow.shadow_table[0]);
unsigned int i;

for_each_shadow_table(v, i)
{
#if GUEST_PAGING_LEVELS == 2
/* 2-on-3: make a PAE l3 that points at the four-page l2 */
if ( i != 0 )
smfn = sh_next_page(smfn);
#else
/* 3-on-3: make a PAE l3 that points at the four l2 pages */
smfn = pagetable_get_mfn(v->arch.shadow_table[i]);
smfn = pagetable_get_mfn(v->arch.paging.shadow.shadow_table[i]);
#endif
v->arch.paging.shadow.l3table[i] =
(mfn_x(smfn) == 0)
Expand All @@ -4108,7 +4115,7 @@ sh_update_cr3(struct vcpu *v, int do_locking, bool noflush)
/* We don't support PV except guest == shadow == config levels */
BUILD_BUG_ON(GUEST_PAGING_LEVELS != SHADOW_PAGING_LEVELS);
/* Just use the shadow top-level directly */
make_cr3(v, pagetable_get_mfn(v->arch.shadow_table[0]));
make_cr3(v, pagetable_get_mfn(v->arch.paging.shadow.shadow_table[0]));
}
#endif

Expand All @@ -4124,7 +4131,8 @@ sh_update_cr3(struct vcpu *v, int do_locking, bool noflush)
v->arch.hvm.hw_cr[3] = virt_to_maddr(&v->arch.paging.shadow.l3table);
#else
/* 4-on-4: Just use the shadow top-level directly */
v->arch.hvm.hw_cr[3] = pagetable_get_paddr(v->arch.shadow_table[0]);
v->arch.hvm.hw_cr[3] =
pagetable_get_paddr(v->arch.paging.shadow.shadow_table[0]);
#endif
hvm_update_guest_cr3(v, noflush);
}
Expand Down Expand Up @@ -4443,7 +4451,7 @@ static void sh_pagetable_dying(paddr_t gpa)
{
struct vcpu *v = current;
struct domain *d = v->domain;
int i = 0;
unsigned int i;
int flush = 0;
int fast_path = 0;
paddr_t gcr3 = 0;
Expand Down Expand Up @@ -4474,15 +4482,16 @@ static void sh_pagetable_dying(paddr_t gpa)
gl3pa = map_domain_page(l3mfn);
gl3e = (guest_l3e_t *)(gl3pa + ((unsigned long)gpa & ~PAGE_MASK));
}
for ( i = 0; i < 4; i++ )
for_each_shadow_table(v, i)
{
mfn_t smfn, gmfn;

if ( fast_path ) {
if ( pagetable_is_null(v->arch.shadow_table[i]) )
if ( fast_path )
{
if ( pagetable_is_null(v->arch.paging.shadow.shadow_table[i]) )
smfn = INVALID_MFN;
else
smfn = pagetable_get_mfn(v->arch.shadow_table[i]);
smfn = pagetable_get_mfn(v->arch.paging.shadow.shadow_table[i]);
}
else
{
Expand Down
16 changes: 13 additions & 3 deletions xen/include/asm-x86/domain.h
Original file line number Diff line number Diff line change
Expand Up @@ -135,6 +135,14 @@ struct shadow_vcpu {
l3_pgentry_t l3table[4] __attribute__((__aligned__(32)));
/* PAE guests: per-vcpu cache of the top-level *guest* entries */
l3_pgentry_t gl3e[4] __attribute__((__aligned__(32)));

/* shadow(s) of guest (MFN) */
#ifdef CONFIG_HVM
pagetable_t shadow_table[4];
#else
pagetable_t shadow_table[1];
#endif

/* Last MFN that we emulated a write to as unshadow heuristics. */
unsigned long last_emulated_mfn_for_unshadow;
/* MFN of the last shadow that we shot a writeable mapping in */
Expand Down Expand Up @@ -576,16 +584,18 @@ struct arch_vcpu
struct hvm_vcpu hvm;
};

/*
* guest_table{,_user} hold a ref to the page, and also a type-count
* unless shadow refcounts are in use
*/
pagetable_t guest_table_user; /* (MFN) x86/64 user-space pagetable */
pagetable_t guest_table; /* (MFN) guest notion of cr3 */
struct page_info *old_guest_table; /* partially destructed pagetable */
struct page_info *old_guest_ptpg; /* containing page table of the */
/* former, if any */
bool old_guest_table_partial; /* Are we dropping a type ref, or just
* finishing up a partial de-validation? */
/* guest_table holds a ref to the page, and also a type-count unless
* shadow refcounts are in use */
pagetable_t shadow_table[4]; /* (MFN) shadow(s) of guest */

unsigned long cr3; /* (MA) value to install in HW CR3 */

/*
Expand Down

0 comments on commit ef3b0d8

Please sign in to comment.