Blame view

kernel/linux-imx6_3.14.28/Documentation/powerpc/eeh-pci-error-recovery.txt 15 KB
6b13f685e   김민수   BSP 최초 추가
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
  
  
                        PCI Bus EEH Error Recovery
                        --------------------------
                             Linas Vepstas
                         <linas@austin.ibm.com>
                            12 January 2005
  
  
  Overview:
  ---------
  The IBM POWER-based pSeries and iSeries computers include PCI bus
  controller chips that have extended capabilities for detecting and
  reporting a large variety of PCI bus error conditions.  These features
  go under the name of "EEH", for "Extended Error Handling".  The EEH
  hardware features allow PCI bus errors to be cleared and a PCI
  card to be "rebooted", without also having to reboot the operating
  system.
  
  This is in contrast to traditional PCI error handling, where the
  PCI chip is wired directly to the CPU, and an error would cause
  a CPU machine-check/check-stop condition, halting the CPU entirely.
  Another "traditional" technique is to ignore such errors, which
  can lead to data corruption, both of user data or of kernel data,
  hung/unresponsive adapters, or system crashes/lockups.  Thus,
  the idea behind EEH is that the operating system can become more
  reliable and robust by protecting it from PCI errors, and giving
  the OS the ability to "reboot"/recover individual PCI devices.
  
  Future systems from other vendors, based on the PCI-E specification,
  may contain similar features.
  
  
  Causes of EEH Errors
  --------------------
  EEH was originally designed to guard against hardware failure, such
  as PCI cards dying from heat, humidity, dust, vibration and bad
  electrical connections. The vast majority of EEH errors seen in
  "real life" are due to either poorly seated PCI cards, or,
  unfortunately quite commonly, due to device driver bugs, device firmware
  bugs, and sometimes PCI card hardware bugs.
  
  The most common software bug, is one that causes the device to
  attempt to DMA to a location in system memory that has not been
  reserved for DMA access for that card.  This is a powerful feature,
  as it prevents what; otherwise, would have been silent memory
  corruption caused by the bad DMA.  A number of device driver
  bugs have been found and fixed in this way over the past few
  years.  Other possible causes of EEH errors include data or
  address line parity errors (for example, due to poor electrical
  connectivity due to a poorly seated card), and PCI-X split-completion
  errors (due to software, device firmware, or device PCI hardware bugs).
  The vast majority of "true hardware failures" can be cured by
  physically removing and re-seating the PCI card.
  
  
  Detection and Recovery
  ----------------------
  In the following discussion, a generic overview of how to detect
  and recover from EEH errors will be presented. This is followed
  by an overview of how the current implementation in the Linux
  kernel does it.  The actual implementation is subject to change,
  and some of the finer points are still being debated.  These
  may in turn be swayed if or when other architectures implement
  similar functionality.
  
  When a PCI Host Bridge (PHB, the bus controller connecting the
  PCI bus to the system CPU electronics complex) detects a PCI error
  condition, it will "isolate" the affected PCI card.  Isolation
  will block all writes (either to the card from the system, or
  from the card to the system), and it will cause all reads to
  return all-ff's (0xff, 0xffff, 0xffffffff for 8/16/32-bit reads).
  This value was chosen because it is the same value you would
  get if the device was physically unplugged from the slot.
  This includes access to PCI memory, I/O space, and PCI config
  space.  Interrupts; however, will continued to be delivered.
  
  Detection and recovery are performed with the aid of ppc64
  firmware.  The programming interfaces in the Linux kernel
  into the firmware are referred to as RTAS (Run-Time Abstraction
  Services).  The Linux kernel does not (should not) access
  the EEH function in the PCI chipsets directly, primarily because
  there are a number of different chipsets out there, each with
  different interfaces and quirks. The firmware provides a
  uniform abstraction layer that will work with all pSeries
  and iSeries hardware (and be forwards-compatible).
  
  If the OS or device driver suspects that a PCI slot has been
  EEH-isolated, there is a firmware call it can make to determine if
  this is the case. If so, then the device driver should put itself
  into a consistent state (given that it won't be able to complete any
  pending work) and start recovery of the card.  Recovery normally
  would consist of resetting the PCI device (holding the PCI #RST
  line high for two seconds), followed by setting up the device
  config space (the base address registers (BAR's), latency timer,
  cache line size, interrupt line, and so on).  This is followed by a
  reinitialization of the device driver.  In a worst-case scenario,
  the power to the card can be toggled, at least on hot-plug-capable
  slots.  In principle, layers far above the device driver probably
  do not need to know that the PCI card has been "rebooted" in this
  way; ideally, there should be at most a pause in Ethernet/disk/USB
  I/O while the card is being reset.
  
  If the card cannot be recovered after three or four resets, the
  kernel/device driver should assume the worst-case scenario, that the
  card has died completely, and report this error to the sysadmin.
  In addition, error messages are reported through RTAS and also through
  syslogd (/var/log/messages) to alert the sysadmin of PCI resets.
  The correct way to deal with failed adapters is to use the standard
  PCI hotplug tools to remove and replace the dead card.
  
  
  Current PPC64 Linux EEH Implementation
  --------------------------------------
  At this time, a generic EEH recovery mechanism has been implemented,
  so that individual device drivers do not need to be modified to support
  EEH recovery.  This generic mechanism piggy-backs on the PCI hotplug
  infrastructure,  and percolates events up through the userspace/udev
  infrastructure.  Following is a detailed description of how this is
  accomplished.
  
  EEH must be enabled in the PHB's very early during the boot process,
  and if a PCI slot is hot-plugged. The former is performed by
  eeh_init() in arch/powerpc/platforms/pseries/eeh.c, and the later by
  drivers/pci/hotplug/pSeries_pci.c calling in to the eeh.c code.
  EEH must be enabled before a PCI scan of the device can proceed.
  Current Power5 hardware will not work unless EEH is enabled;
  although older Power4 can run with it disabled.  Effectively,
  EEH can no longer be turned off.  PCI devices *must* be
  registered with the EEH code; the EEH code needs to know about
  the I/O address ranges of the PCI device in order to detect an
  error.  Given an arbitrary address, the routine
  pci_get_device_by_addr() will find the pci device associated
  with that address (if any).
  
  The default arch/powerpc/include/asm/io.h macros readb(), inb(), insb(),
  etc. include a check to see if the i/o read returned all-0xff's.
  If so, these make a call to eeh_dn_check_failure(), which in turn
  asks the firmware if the all-ff's value is the sign of a true EEH
  error.  If it is not, processing continues as normal.  The grand
  total number of these false alarms or "false positives" can be
  seen in /proc/ppc64/eeh (subject to change).  Normally, almost
  all of these occur during boot, when the PCI bus is scanned, where
  a large number of 0xff reads are part of the bus scan procedure.
  
  If a frozen slot is detected, code in 
  arch/powerpc/platforms/pseries/eeh.c will print a stack trace to 
  syslog (/var/log/messages).  This stack trace has proven to be very 
  useful to device-driver authors for finding out at what point the EEH 
  error was detected, as the error itself usually occurs slightly 
  beforehand.
  
  Next, it uses the Linux kernel notifier chain/work queue mechanism to
  allow any interested parties to find out about the failure.  Device
  drivers, or other parts of the kernel, can use
  eeh_register_notifier(struct notifier_block *) to find out about EEH
  events.  The event will include a pointer to the pci device, the
  device node and some state info.  Receivers of the event can "do as
  they wish"; the default handler will be described further in this
  section.
  
  To assist in the recovery of the device, eeh.c exports the
  following functions:
  
  rtas_set_slot_reset() -- assert the  PCI #RST line for 1/8th of a second
  rtas_configure_bridge() -- ask firmware to configure any PCI bridges
     located topologically under the pci slot.
  eeh_save_bars() and eeh_restore_bars(): save and restore the PCI
     config-space info for a device and any devices under it.
  
  
  A handler for the EEH notifier_block events is implemented in
  drivers/pci/hotplug/pSeries_pci.c, called handle_eeh_events().
  It saves the device BAR's and then calls rpaphp_unconfig_pci_adapter().
  This last call causes the device driver for the card to be stopped,
  which causes uevents to go out to user space. This triggers
  user-space scripts that might issue commands such as "ifdown eth0"
  for ethernet cards, and so on.  This handler then sleeps for 5 seconds,
  hoping to give the user-space scripts enough time to complete.
  It then resets the PCI card, reconfigures the device BAR's, and
  any bridges underneath. It then calls rpaphp_enable_pci_slot(),
  which restarts the device driver and triggers more user-space
  events (for example, calling "ifup eth0" for ethernet cards).
  
  
  Device Shutdown and User-Space Events
  -------------------------------------
  This section documents what happens when a pci slot is unconfigured,
  focusing on how the device driver gets shut down, and on how the
  events get delivered to user-space scripts.
  
  Following is an example sequence of events that cause a device driver
  close function to be called during the first phase of an EEH reset.
  The following sequence is an example of the pcnet32 device driver.
  
      rpa_php_unconfig_pci_adapter (struct slot *)  // in rpaphp_pci.c
      {
        calls
        pci_remove_bus_device (struct pci_dev *) // in /drivers/pci/remove.c
        {
          calls
          pci_destroy_dev (struct pci_dev *)
          {
            calls
            device_unregister (&dev->dev) // in /drivers/base/core.c
            {
              calls
              device_del (struct device *)
              {
                calls
                bus_remove_device() // in /drivers/base/bus.c
                {
                  calls
                  device_release_driver()
                  {
                    calls
                    struct device_driver->remove() which is just
                    pci_device_remove()  // in /drivers/pci/pci_driver.c
                    {
                      calls
                      struct pci_driver->remove() which is just
                      pcnet32_remove_one() // in /drivers/net/pcnet32.c
                      {
                        calls
                        unregister_netdev() // in /net/core/dev.c
                        {
                          calls
                          dev_close()  // in /net/core/dev.c
                          {
                             calls dev->stop();
                             which is just pcnet32_close() // in pcnet32.c
                             {
                               which does what you wanted
                               to stop the device
                             }
                          }
                       }
                     which
                     frees pcnet32 device driver memory
                  }
       }}}}}}
  
  
      in drivers/pci/pci_driver.c,
      struct device_driver->remove() is just pci_device_remove()
      which calls struct pci_driver->remove() which is pcnet32_remove_one()
      which calls unregister_netdev()  (in net/core/dev.c)
      which calls dev_close()  (in net/core/dev.c)
      which calls dev->stop() which is pcnet32_close()
      which then does the appropriate shutdown.
  
  ---
  Following is the analogous stack trace for events sent to user-space
  when the pci device is unconfigured.
  
  rpa_php_unconfig_pci_adapter() {             // in rpaphp_pci.c
    calls
    pci_remove_bus_device (struct pci_dev *) { // in /drivers/pci/remove.c
      calls
      pci_destroy_dev (struct pci_dev *) {
        calls
        device_unregister (&dev->dev) {        // in /drivers/base/core.c
          calls
          device_del(struct device * dev) {    // in /drivers/base/core.c
            calls
            kobject_del() {                    //in /libs/kobject.c
              calls
              kobject_uevent() {               // in /libs/kobject.c
                calls
                kset_uevent() {                // in /lib/kobject.c
                  calls
                  kset->uevent_ops->uevent()   // which is really just
                  a call to
                  dev_uevent() {               // in /drivers/base/core.c
                    calls
                    dev->bus->uevent() which is really just a call to
                    pci_uevent () {            // in drivers/pci/hotplug.c
                      which prints device name, etc....
                   }
                 }
                 then kobject_uevent() sends a netlink uevent to userspace
                 --> userspace uevent
                 (during early boot, nobody listens to netlink events and
                 kobject_uevent() executes uevent_helper[], which runs the
                 event process /sbin/hotplug)
             }
           }
           kobject_del() then calls sysfs_remove_dir(), which would
           trigger any user-space daemon that was watching /sysfs,
           and notice the delete event.
  
  
  Pro's and Con's of the Current Design
  -------------------------------------
  There are several issues with the current EEH software recovery design,
  which may be addressed in future revisions.  But first, note that the
  big plus of the current design is that no changes need to be made to
  individual device drivers, so that the current design throws a wide net.
  The biggest negative of the design is that it potentially disturbs
  network daemons and file systems that didn't need to be disturbed.
  
  -- A minor complaint is that resetting the network card causes
     user-space back-to-back ifdown/ifup burps that potentially disturb
     network daemons, that didn't need to even know that the pci
     card was being rebooted.
  
  -- A more serious concern is that the same reset, for SCSI devices,
     causes havoc to mounted file systems.  Scripts cannot post-facto
     unmount a file system without flushing pending buffers, but this
     is impossible, because I/O has already been stopped.  Thus,
     ideally, the reset should happen at or below the block layer,
     so that the file systems are not disturbed.
  
     Reiserfs does not tolerate errors returned from the block device.
     Ext3fs seems to be tolerant, retrying reads/writes until it does
     succeed. Both have been only lightly tested in this scenario.
  
     The SCSI-generic subsystem already has built-in code for performing
     SCSI device resets, SCSI bus resets, and SCSI host-bus-adapter
     (HBA) resets.  These are cascaded into a chain of attempted
     resets if a SCSI command fails. These are completely hidden
     from the block layer.  It would be very natural to add an EEH
     reset into this chain of events.
  
  -- If a SCSI error occurs for the root device, all is lost unless
     the sysadmin had the foresight to run /bin, /sbin, /etc, /var
     and so on, out of ramdisk/tmpfs.
  
  
  Conclusions
  -----------
  There's forward progress ...