Lines Matching defs:uobjpage

491     struct vm_page *uobjpage,
500 if (uobjpage != PGO_DONTCARE)
501 uobj = uobjpage->uobject;
509 (uobjpage == PGO_DONTCARE) ? UVM_PGA_ZERO : 0);
538 if (uobjpage != PGO_DONTCARE)
539 uvm_pagecopy(uobjpage, pg);
1151 struct vm_page *uobjpage = NULL;
1186 * remember this page as "uobjpage." (for later use).
1189 uobjpage = pages[lcv];
1227 return uobjpage;
1249 struct vm_page *uobjpage, *pg = NULL;
1262 uobjpage = NULL;
1264 uobjpage = uvm_fault_lower_lookup(ufi, flt, pages);
1272 * then we've got a pointer to it as "uobjpage" and we've already
1285 * note that uobjpage can not be PGO_DONTCARE at this point. we now
1286 * set uobjpage to PGO_DONTCARE if we are doing a zero fill. if we
1291 uobjpage = PGO_DONTCARE;
1294 KASSERT(uobjpage != PGO_DONTCARE);
1300 * if uobjpage is not null then we do not need to do I/O to get the
1301 * uobjpage.
1303 * if uobjpage is null, then we need to ask the pager to
1307 if (uobjpage) {
1310 if (uobjpage != PGO_DONTCARE) {
1312 uvm_pageactivate(uobjpage);
1316 error = uvm_fault_lower_io(ufi, flt, &uobj, &uobjpage);
1323 * - at this point uobjpage can not be NULL
1324 * - at this point uobjpage could be PG_WANTED (handle later)
1333 * set "pg" to the page we want to map in (uobjpage, usually)
1338 pg = uobjpage; /* map in the actual object */
1340 /* assert(uobjpage != PGO_DONTCARE) */
1357 error = uvmfault_promote(ufi, uobjpage, &anon, &pg);
1364 if (uobjpage != PGO_DONTCARE) {
1372 pmap_page_protect(uobjpage, PROT_NONE);
1384 pmap_page_protect(uobjpage, PROT_NONE);
1387 /* done with copied uobjpage. */
1422 * Note: pg is either the uobjpage or the new page in the new anon.