Blame view

kernel/linux-rt-4.4.41/Documentation/vm/userfaultfd.txt 6.95 KB
5113f6f70   김현기   kernel add
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
  = Userfaultfd =
  
  == Objective ==
  
  Userfaults allow the implementation of on-demand paging from userland
  and more generally they allow userland to take control of various
  memory page faults, something otherwise only the kernel code could do.
  
  For example userfaults allows a proper and more optimal implementation
  of the PROT_NONE+SIGSEGV trick.
  
  == Design ==
  
  Userfaults are delivered and resolved through the userfaultfd syscall.
  
  The userfaultfd (aside from registering and unregistering virtual
  memory ranges) provides two primary functionalities:
  
  1) read/POLLIN protocol to notify a userland thread of the faults
     happening
  
  2) various UFFDIO_* ioctls that can manage the virtual memory regions
     registered in the userfaultfd that allows userland to efficiently
     resolve the userfaults it receives via 1) or to manage the virtual
     memory in the background
  
  The real advantage of userfaults if compared to regular virtual memory
  management of mremap/mprotect is that the userfaults in all their
  operations never involve heavyweight structures like vmas (in fact the
  userfaultfd runtime load never takes the mmap_sem for writing).
  
  Vmas are not suitable for page- (or hugepage) granular fault tracking
  when dealing with virtual address spaces that could span
  Terabytes. Too many vmas would be needed for that.
  
  The userfaultfd once opened by invoking the syscall, can also be
  passed using unix domain sockets to a manager process, so the same
  manager process could handle the userfaults of a multitude of
  different processes without them being aware about what is going on
  (well of course unless they later try to use the userfaultfd
  themselves on the same region the manager is already tracking, which
  is a corner case that would currently return -EBUSY).
  
  == API ==
  
  When first opened the userfaultfd must be enabled invoking the
  UFFDIO_API ioctl specifying a uffdio_api.api value set to UFFD_API (or
  a later API version) which will specify the read/POLLIN protocol
  userland intends to speak on the UFFD and the uffdio_api.features
  userland requires. The UFFDIO_API ioctl if successful (i.e. if the
  requested uffdio_api.api is spoken also by the running kernel and the
  requested features are going to be enabled) will return into
  uffdio_api.features and uffdio_api.ioctls two 64bit bitmasks of
  respectively all the available features of the read(2) protocol and
  the generic ioctl available.
  
  Once the userfaultfd has been enabled the UFFDIO_REGISTER ioctl should
  be invoked (if present in the returned uffdio_api.ioctls bitmask) to
  register a memory range in the userfaultfd by setting the
  uffdio_register structure accordingly. The uffdio_register.mode
  bitmask will specify to the kernel which kind of faults to track for
  the range (UFFDIO_REGISTER_MODE_MISSING would track missing
  pages). The UFFDIO_REGISTER ioctl will return the
  uffdio_register.ioctls bitmask of ioctls that are suitable to resolve
  userfaults on the range registered. Not all ioctls will necessarily be
  supported for all memory types depending on the underlying virtual
  memory backend (anonymous memory vs tmpfs vs real filebacked
  mappings).
  
  Userland can use the uffdio_register.ioctls to manage the virtual
  address space in the background (to add or potentially also remove
  memory from the userfaultfd registered range). This means a userfault
  could be triggering just before userland maps in the background the
  user-faulted page.
  
  The primary ioctl to resolve userfaults is UFFDIO_COPY. That
  atomically copies a page into the userfault registered range and wakes
  up the blocked userfaults (unless uffdio_copy.mode &
  UFFDIO_COPY_MODE_DONTWAKE is set). Other ioctl works similarly to
  UFFDIO_COPY. They're atomic as in guaranteeing that nothing can see an
  half copied page since it'll keep userfaulting until the copy has
  finished.
  
  == QEMU/KVM ==
  
  QEMU/KVM is using the userfaultfd syscall to implement postcopy live
  migration. Postcopy live migration is one form of memory
  externalization consisting of a virtual machine running with part or
  all of its memory residing on a different node in the cloud. The
  userfaultfd abstraction is generic enough that not a single line of
  KVM kernel code had to be modified in order to add postcopy live
  migration to QEMU.
  
  Guest async page faults, FOLL_NOWAIT and all other GUP features work
  just fine in combination with userfaults. Userfaults trigger async
  page faults in the guest scheduler so those guest processes that
  aren't waiting for userfaults (i.e. network bound) can keep running in
  the guest vcpus.
  
  It is generally beneficial to run one pass of precopy live migration
  just before starting postcopy live migration, in order to avoid
  generating userfaults for readonly guest regions.
  
  The implementation of postcopy live migration currently uses one
  single bidirectional socket but in the future two different sockets
  will be used (to reduce the latency of the userfaults to the minimum
  possible without having to decrease /proc/sys/net/ipv4/tcp_wmem).
  
  The QEMU in the source node writes all pages that it knows are missing
  in the destination node, into the socket, and the migration thread of
  the QEMU running in the destination node runs UFFDIO_COPY|ZEROPAGE
  ioctls on the userfaultfd in order to map the received pages into the
  guest (UFFDIO_ZEROCOPY is used if the source page was a zero page).
  
  A different postcopy thread in the destination node listens with
  poll() to the userfaultfd in parallel. When a POLLIN event is
  generated after a userfault triggers, the postcopy thread read() from
  the userfaultfd and receives the fault address (or -EAGAIN in case the
  userfault was already resolved and waken by a UFFDIO_COPY|ZEROPAGE run
  by the parallel QEMU migration thread).
  
  After the QEMU postcopy thread (running in the destination node) gets
  the userfault address it writes the information about the missing page
  into the socket. The QEMU source node receives the information and
  roughly "seeks" to that page address and continues sending all
  remaining missing pages from that new page offset. Soon after that
  (just the time to flush the tcp_wmem queue through the network) the
  migration thread in the QEMU running in the destination node will
  receive the page that triggered the userfault and it'll map it as
  usual with the UFFDIO_COPY|ZEROPAGE (without actually knowing if it
  was spontaneously sent by the source or if it was an urgent page
  requested through an userfault).
  
  By the time the userfaults start, the QEMU in the destination node
  doesn't need to keep any per-page state bitmap relative to the live
  migration around and a single per-page bitmap has to be maintained in
  the QEMU running in the source node to know which pages are still
  missing in the destination node. The bitmap in the source node is
  checked to find which missing pages to send in round robin and we seek
  over it when receiving incoming userfaults. After sending each page of
  course the bitmap is updated accordingly. It's also useful to avoid
  sending the same page twice (in case the userfault is read by the
  postcopy thread just before UFFDIO_COPY|ZEROPAGE runs in the migration
  thread).