@@ -1046,7 +1046,7 @@ int xenmem_add_to_physmap_one(
unsigned int space,
union xen_add_to_physmap_batch_extra extra,
unsigned long idx,
- xen_pfn_t gpfn)
+ gfn_t gfn)
{
unsigned long mfn = 0;
int rc;
@@ -1081,8 +1081,8 @@ int xenmem_add_to_physmap_one(
else
return -EINVAL;
}
-
- d->arch.grant_table_gpfn[idx] = gpfn;
+
+ d->arch.grant_table_gpfn[idx] = gfn_x(gfn);
t = p2m_ram_rw;
@@ -1145,7 +1145,7 @@ int xenmem_add_to_physmap_one(
if ( extra.res0 )
return -EOPNOTSUPP;
- rc = map_dev_mmio_region(d, gpfn, 1, idx);
+ rc = map_dev_mmio_region(d, gfn_x(gfn), 1, idx);
return rc;
default:
@@ -1153,7 +1153,7 @@ int xenmem_add_to_physmap_one(
}
/* Map at new location. */
- rc = guest_physmap_add_entry(d, _gfn(gpfn), _mfn(mfn), 0, t);
+ rc = guest_physmap_add_entry(d, gfn, _mfn(mfn), 0, t);
/* If we fail to add the mapping, we need to drop the reference we
* took earlier on foreign pages */
@@ -4775,7 +4775,7 @@ int xenmem_add_to_physmap_one(
unsigned int space,
union xen_add_to_physmap_batch_extra extra,
unsigned long idx,
- xen_pfn_t gpfn)
+ gfn_t gpfn)
{
struct page_info *page = NULL;
unsigned long gfn = 0; /* gcc ... */
@@ -4834,7 +4834,7 @@ int xenmem_add_to_physmap_one(
break;
}
case XENMAPSPACE_gmfn_foreign:
- return p2m_add_foreign(d, idx, gpfn, extra.foreign_domid);
+ return p2m_add_foreign(d, idx, gfn_x(gpfn), extra.foreign_domid);
default:
break;
}
@@ -4849,19 +4849,18 @@ int xenmem_add_to_physmap_one(
}
/* Remove previously mapped page if it was present. */
- prev_mfn = mfn_x(get_gfn(d, gpfn, &p2mt));
+ prev_mfn = mfn_x(get_gfn(d, gfn_x(gpfn), &p2mt));
if ( mfn_valid(prev_mfn) )
{
if ( is_xen_heap_mfn(prev_mfn) )
/* Xen heap frames are simply unhooked from this phys slot. */
- guest_physmap_remove_page(d, _gfn(gpfn), _mfn(prev_mfn),
- PAGE_ORDER_4K);
+ guest_physmap_remove_page(d, gpfn, _mfn(prev_mfn), PAGE_ORDER_4K);
else
/* Normal domain memory is freed, to avoid leaking memory. */
- guest_remove_page(d, gpfn);
+ guest_remove_page(d, gfn_x(gpfn));
}
/* In the XENMAPSPACE_gmfn case we still hold a ref on the old page. */
- put_gfn(d, gpfn);
+ put_gfn(d, gfn_x(gpfn));
/* Unmap from old location, if any. */
old_gpfn = get_gpfn_from_mfn(mfn);
@@ -4872,7 +4871,7 @@ int xenmem_add_to_physmap_one(
guest_physmap_remove_page(d, _gfn(old_gpfn), _mfn(mfn), PAGE_ORDER_4K);
/* Map at new location. */
- rc = guest_physmap_add_page(d, _gfn(gpfn), _mfn(mfn), PAGE_ORDER_4K);
+ rc = guest_physmap_add_page(d, gpfn, _mfn(mfn), PAGE_ORDER_4K);
/* In the XENMAPSPACE_gmfn, we took a ref of the gfn at the top */
if ( space == XENMAPSPACE_gmfn || space == XENMAPSPACE_gmfn_range )
@@ -649,7 +649,7 @@ static int xenmem_add_to_physmap(struct domain *d,
if ( xatp->space != XENMAPSPACE_gmfn_range )
return xenmem_add_to_physmap_one(d, xatp->space, extra,
- xatp->idx, xatp->gpfn);
+ xatp->idx, _gfn(xatp->gpfn));
if ( xatp->size < start )
return -EILSEQ;
@@ -666,7 +666,7 @@ static int xenmem_add_to_physmap(struct domain *d,
while ( xatp->size > done )
{
rc = xenmem_add_to_physmap_one(d, xatp->space, extra,
- xatp->idx, xatp->gpfn);
+ xatp->idx, _gfn(xatp->gpfn));
if ( rc < 0 )
break;
@@ -735,7 +735,7 @@ static int xenmem_add_to_physmap_batch(struct domain *d,
rc = xenmem_add_to_physmap_one(d, xatpb->space,
xatpb->u,
- idx, gpfn);
+ idx, _gfn(gpfn));
if ( unlikely(__copy_to_guest_offset(xatpb->errs, 0, &rc, 1)) )
{
@@ -548,7 +548,7 @@ void scrub_one_page(struct page_info *);
int xenmem_add_to_physmap_one(struct domain *d, unsigned int space,
union xen_add_to_physmap_batch_extra extra,
- unsigned long idx, xen_pfn_t gpfn);
+ unsigned long idx, gfn_t gfn);
/* Returns 1 on success, 0 on error, negative if the ring
* for event propagation is full in the presence of paging */