Lines Matching full:fault

42 	/* Are we prepared to handle this kernel fault? */  in no_context()
62 static inline void mm_fault_error(struct pt_regs *regs, unsigned long addr, vm_fault_t fault) in mm_fault_error() argument
69 if (fault & VM_FAULT_OOM) { in mm_fault_error()
72 * (which will retry the fault, or kill us if we got oom-killed). in mm_fault_error()
76 } else if (fault & (VM_FAULT_SIGBUS | VM_FAULT_HWPOISON | VM_FAULT_HWPOISON_LARGE)) { in mm_fault_error()
80 } else if (fault & VM_FAULT_SIGSEGV) { in mm_fault_error()
231 vm_fault_t fault; in handle_page_fault() local
243 * Fault-in kernel-space virtual memory on-demand. in handle_page_fault()
263 * in an atomic region, then we must not take the fault. in handle_page_fault()
302 fault = handle_mm_fault(vma, addr, flags | FAULT_FLAG_VMA_LOCK, regs); in handle_page_fault()
303 if (!(fault & (VM_FAULT_RETRY | VM_FAULT_COMPLETED))) in handle_page_fault()
306 if (!(fault & VM_FAULT_RETRY)) { in handle_page_fault()
311 if (fault & VM_FAULT_MAJOR) in handle_page_fault()
314 if (fault_signal_pending(fault, regs)) { in handle_page_fault()
342 * If for any reason at all we could not handle the fault, in handle_page_fault()
344 * the fault. in handle_page_fault()
346 fault = handle_mm_fault(vma, addr, flags, regs); in handle_page_fault()
353 if (fault_signal_pending(fault, regs)) { in handle_page_fault()
359 /* The fault is fully completed (including releasing mmap lock) */ in handle_page_fault()
360 if (fault & VM_FAULT_COMPLETED) in handle_page_fault()
363 if (unlikely(fault & VM_FAULT_RETRY)) { in handle_page_fault()
377 if (unlikely(fault & VM_FAULT_ERROR)) { in handle_page_fault()
379 mm_fault_error(regs, addr, fault); in handle_page_fault()