Blame view

kernel/linux-imx6_3.14.28/Documentation/ww-mutex-design.txt 12.4 KB
6b13f685e   김민수   BSP 최초 추가
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
  Wait/Wound Deadlock-Proof Mutex Design
  ======================================
  
  Please read mutex-design.txt first, as it applies to wait/wound mutexes too.
  
  Motivation for WW-Mutexes
  -------------------------
  
  GPU's do operations that commonly involve many buffers.  Those buffers
  can be shared across contexts/processes, exist in different memory
  domains (for example VRAM vs system memory), and so on.  And with
  PRIME / dmabuf, they can even be shared across devices.  So there are
  a handful of situations where the driver needs to wait for buffers to
  become ready.  If you think about this in terms of waiting on a buffer
  mutex for it to become available, this presents a problem because
  there is no way to guarantee that buffers appear in a execbuf/batch in
  the same order in all contexts.  That is directly under control of
  userspace, and a result of the sequence of GL calls that an application
  makes.	Which results in the potential for deadlock.  The problem gets
  more complex when you consider that the kernel may need to migrate the
  buffer(s) into VRAM before the GPU operates on the buffer(s), which
  may in turn require evicting some other buffers (and you don't want to
  evict other buffers which are already queued up to the GPU), but for a
  simplified understanding of the problem you can ignore this.
  
  The algorithm that the TTM graphics subsystem came up with for dealing with
  this problem is quite simple.  For each group of buffers (execbuf) that need
  to be locked, the caller would be assigned a unique reservation id/ticket,
  from a global counter.  In case of deadlock while locking all the buffers
  associated with a execbuf, the one with the lowest reservation ticket (i.e.
  the oldest task) wins, and the one with the higher reservation id (i.e. the
  younger task) unlocks all of the buffers that it has already locked, and then
  tries again.
  
  In the RDBMS literature this deadlock handling approach is called wait/wound:
  The older tasks waits until it can acquire the contended lock. The younger tasks
  needs to back off and drop all the locks it is currently holding, i.e. the
  younger task is wounded.
  
  Concepts
  --------
  
  Compared to normal mutexes two additional concepts/objects show up in the lock
  interface for w/w mutexes:
  
  Acquire context: To ensure eventual forward progress it is important the a task
  trying to acquire locks doesn't grab a new reservation id, but keeps the one it
  acquired when starting the lock acquisition. This ticket is stored in the
  acquire context. Furthermore the acquire context keeps track of debugging state
  to catch w/w mutex interface abuse.
  
  W/w class: In contrast to normal mutexes the lock class needs to be explicit for
  w/w mutexes, since it is required to initialize the acquire context.
  
  Furthermore there are three different class of w/w lock acquire functions:
  
  * Normal lock acquisition with a context, using ww_mutex_lock.
  
  * Slowpath lock acquisition on the contending lock, used by the wounded task
    after having dropped all already acquired locks. These functions have the
    _slow postfix.
  
    From a simple semantics point-of-view the _slow functions are not strictly
    required, since simply calling the normal ww_mutex_lock functions on the
    contending lock (after having dropped all other already acquired locks) will
    work correctly. After all if no other ww mutex has been acquired yet there's
    no deadlock potential and hence the ww_mutex_lock call will block and not
    prematurely return -EDEADLK. The advantage of the _slow functions is in
    interface safety:
    - ww_mutex_lock has a __must_check int return type, whereas ww_mutex_lock_slow
      has a void return type. Note that since ww mutex code needs loops/retries
      anyway the __must_check doesn't result in spurious warnings, even though the
      very first lock operation can never fail.
    - When full debugging is enabled ww_mutex_lock_slow checks that all acquired
      ww mutex have been released (preventing deadlocks) and makes sure that we
      block on the contending lock (preventing spinning through the -EDEADLK
      slowpath until the contended lock can be acquired).
  
  * Functions to only acquire a single w/w mutex, which results in the exact same
    semantics as a normal mutex. This is done by calling ww_mutex_lock with a NULL
    context.
  
    Again this is not strictly required. But often you only want to acquire a
    single lock in which case it's pointless to set up an acquire context (and so
    better to avoid grabbing a deadlock avoidance ticket).
  
  Of course, all the usual variants for handling wake-ups due to signals are also
  provided.
  
  Usage
  -----
  
  Three different ways to acquire locks within the same w/w class. Common
  definitions for methods #1 and #2:
  
  static DEFINE_WW_CLASS(ww_class);
  
  struct obj {
  	struct ww_mutex lock;
  	/* obj data */
  };
  
  struct obj_entry {
  	struct list_head head;
  	struct obj *obj;
  };
  
  Method 1, using a list in execbuf->buffers that's not allowed to be reordered.
  This is useful if a list of required objects is already tracked somewhere.
  Furthermore the lock helper can use propagate the -EALREADY return code back to
  the caller as a signal that an object is twice on the list. This is useful if
  the list is constructed from userspace input and the ABI requires userspace to
  not have duplicate entries (e.g. for a gpu commandbuffer submission ioctl).
  
  int lock_objs(struct list_head *list, struct ww_acquire_ctx *ctx)
  {
  	struct obj *res_obj = NULL;
  	struct obj_entry *contended_entry = NULL;
  	struct obj_entry *entry;
  
  	ww_acquire_init(ctx, &ww_class);
  
  retry:
  	list_for_each_entry (entry, list, head) {
  		if (entry->obj == res_obj) {
  			res_obj = NULL;
  			continue;
  		}
  		ret = ww_mutex_lock(&entry->obj->lock, ctx);
  		if (ret < 0) {
  			contended_entry = entry;
  			goto err;
  		}
  	}
  
  	ww_acquire_done(ctx);
  	return 0;
  
  err:
  	list_for_each_entry_continue_reverse (entry, list, head)
  		ww_mutex_unlock(&entry->obj->lock);
  
  	if (res_obj)
  		ww_mutex_unlock(&res_obj->lock);
  
  	if (ret == -EDEADLK) {
  		/* we lost out in a seqno race, lock and retry.. */
  		ww_mutex_lock_slow(&contended_entry->obj->lock, ctx);
  		res_obj = contended_entry->obj;
  		goto retry;
  	}
  	ww_acquire_fini(ctx);
  
  	return ret;
  }
  
  Method 2, using a list in execbuf->buffers that can be reordered. Same semantics
  of duplicate entry detection using -EALREADY as method 1 above. But the
  list-reordering allows for a bit more idiomatic code.
  
  int lock_objs(struct list_head *list, struct ww_acquire_ctx *ctx)
  {
  	struct obj_entry *entry, *entry2;
  
  	ww_acquire_init(ctx, &ww_class);
  
  	list_for_each_entry (entry, list, head) {
  		ret = ww_mutex_lock(&entry->obj->lock, ctx);
  		if (ret < 0) {
  			entry2 = entry;
  
  			list_for_each_entry_continue_reverse (entry2, list, head)
  				ww_mutex_unlock(&entry2->obj->lock);
  
  			if (ret != -EDEADLK) {
  				ww_acquire_fini(ctx);
  				return ret;
  			}
  
  			/* we lost out in a seqno race, lock and retry.. */
  			ww_mutex_lock_slow(&entry->obj->lock, ctx);
  
  			/*
  			 * Move buf to head of the list, this will point
  			 * buf->next to the first unlocked entry,
  			 * restarting the for loop.
  			 */
  			list_del(&entry->head);
  			list_add(&entry->head, list);
  		}
  	}
  
  	ww_acquire_done(ctx);
  	return 0;
  }
  
  Unlocking works the same way for both methods #1 and #2:
  
  void unlock_objs(struct list_head *list, struct ww_acquire_ctx *ctx)
  {
  	struct obj_entry *entry;
  
  	list_for_each_entry (entry, list, head)
  		ww_mutex_unlock(&entry->obj->lock);
  
  	ww_acquire_fini(ctx);
  }
  
  Method 3 is useful if the list of objects is constructed ad-hoc and not upfront,
  e.g. when adjusting edges in a graph where each node has its own ww_mutex lock,
  and edges can only be changed when holding the locks of all involved nodes. w/w
  mutexes are a natural fit for such a case for two reasons:
  - They can handle lock-acquisition in any order which allows us to start walking
    a graph from a starting point and then iteratively discovering new edges and
    locking down the nodes those edges connect to.
  - Due to the -EALREADY return code signalling that a given objects is already
    held there's no need for additional book-keeping to break cycles in the graph
    or keep track off which looks are already held (when using more than one node
    as a starting point).
  
  Note that this approach differs in two important ways from the above methods:
  - Since the list of objects is dynamically constructed (and might very well be
    different when retrying due to hitting the -EDEADLK wound condition) there's
    no need to keep any object on a persistent list when it's not locked. We can
    therefore move the list_head into the object itself.
  - On the other hand the dynamic object list construction also means that the -EALREADY return
    code can't be propagated.
  
  Note also that methods #1 and #2 and method #3 can be combined, e.g. to first lock a
  list of starting nodes (passed in from userspace) using one of the above
  methods. And then lock any additional objects affected by the operations using
  method #3 below. The backoff/retry procedure will be a bit more involved, since
  when the dynamic locking step hits -EDEADLK we also need to unlock all the
  objects acquired with the fixed list. But the w/w mutex debug checks will catch
  any interface misuse for these cases.
  
  Also, method 3 can't fail the lock acquisition step since it doesn't return
  -EALREADY. Of course this would be different when using the _interruptible
  variants, but that's outside of the scope of these examples here.
  
  struct obj {
  	struct ww_mutex ww_mutex;
  	struct list_head locked_list;
  };
  
  static DEFINE_WW_CLASS(ww_class);
  
  void __unlock_objs(struct list_head *list)
  {
  	struct obj *entry, *temp;
  
  	list_for_each_entry_safe (entry, temp, list, locked_list) {
  		/* need to do that before unlocking, since only the current lock holder is
  		allowed to use object */
  		list_del(&entry->locked_list);
  		ww_mutex_unlock(entry->ww_mutex)
  	}
  }
  
  void lock_objs(struct list_head *list, struct ww_acquire_ctx *ctx)
  {
  	struct obj *obj;
  
  	ww_acquire_init(ctx, &ww_class);
  
  retry:
  	/* re-init loop start state */
  	loop {
  		/* magic code which walks over a graph and decides which objects
  		 * to lock */
  
  		ret = ww_mutex_lock(obj->ww_mutex, ctx);
  		if (ret == -EALREADY) {
  			/* we have that one already, get to the next object */
  			continue;
  		}
  		if (ret == -EDEADLK) {
  			__unlock_objs(list);
  
  			ww_mutex_lock_slow(obj, ctx);
  			list_add(&entry->locked_list, list);
  			goto retry;
  		}
  
  		/* locked a new object, add it to the list */
  		list_add_tail(&entry->locked_list, list);
  	}
  
  	ww_acquire_done(ctx);
  	return 0;
  }
  
  void unlock_objs(struct list_head *list, struct ww_acquire_ctx *ctx)
  {
  	__unlock_objs(list);
  	ww_acquire_fini(ctx);
  }
  
  Method 4: Only lock one single objects. In that case deadlock detection and
  prevention is obviously overkill, since with grabbing just one lock you can't
  produce a deadlock within just one class. To simplify this case the w/w mutex
  api can be used with a NULL context.
  
  Implementation Details
  ----------------------
  
  Design:
    ww_mutex currently encapsulates a struct mutex, this means no extra overhead for
    normal mutex locks, which are far more common. As such there is only a small
    increase in code size if wait/wound mutexes are not used.
  
    In general, not much contention is expected. The locks are typically used to
    serialize access to resources for devices. The only way to make wakeups
    smarter would be at the cost of adding a field to struct mutex_waiter. This
    would add overhead to all cases where normal mutexes are used, and
    ww_mutexes are generally less performance sensitive.
  
  Lockdep:
    Special care has been taken to warn for as many cases of api abuse
    as possible. Some common api abuses will be caught with
    CONFIG_DEBUG_MUTEXES, but CONFIG_PROVE_LOCKING is recommended.
  
    Some of the errors which will be warned about:
     - Forgetting to call ww_acquire_fini or ww_acquire_init.
     - Attempting to lock more mutexes after ww_acquire_done.
     - Attempting to lock the wrong mutex after -EDEADLK and
       unlocking all mutexes.
     - Attempting to lock the right mutex after -EDEADLK,
       before unlocking all mutexes.
  
     - Calling ww_mutex_lock_slow before -EDEADLK was returned.
  
     - Unlocking mutexes with the wrong unlock function.
     - Calling one of the ww_acquire_* twice on the same context.
     - Using a different ww_class for the mutex than for the ww_acquire_ctx.
     - Normal lockdep errors that can result in deadlocks.
  
    Some of the lockdep errors that can result in deadlocks:
     - Calling ww_acquire_init to initialize a second ww_acquire_ctx before
       having called ww_acquire_fini on the first.
     - 'normal' deadlocks that can occur.
  
  FIXME: Update this section once we have the TASK_DEADLOCK task state flag magic
  implemented.