rf_reconstruct.c revision 1.81.2.1 1 1.81.2.1 yamt /* $NetBSD: rf_reconstruct.c,v 1.81.2.1 2005/02/12 18:17:50 yamt Exp $ */
2 1.1 oster /*
3 1.1 oster * Copyright (c) 1995 Carnegie-Mellon University.
4 1.1 oster * All rights reserved.
5 1.1 oster *
6 1.1 oster * Author: Mark Holland
7 1.1 oster *
8 1.1 oster * Permission to use, copy, modify and distribute this software and
9 1.1 oster * its documentation is hereby granted, provided that both the copyright
10 1.1 oster * notice and this permission notice appear in all copies of the
11 1.1 oster * software, derivative works or modified versions, and any portions
12 1.1 oster * thereof, and that both notices appear in supporting documentation.
13 1.1 oster *
14 1.1 oster * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
15 1.1 oster * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND
16 1.1 oster * FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
17 1.1 oster *
18 1.1 oster * Carnegie Mellon requests users of this software to return to
19 1.1 oster *
20 1.1 oster * Software Distribution Coordinator or Software.Distribution (at) CS.CMU.EDU
21 1.1 oster * School of Computer Science
22 1.1 oster * Carnegie Mellon University
23 1.1 oster * Pittsburgh PA 15213-3890
24 1.1 oster *
25 1.1 oster * any improvements or extensions that they make and grant Carnegie the
26 1.1 oster * rights to redistribute these changes.
27 1.1 oster */
28 1.1 oster
29 1.1 oster /************************************************************
30 1.1 oster *
31 1.1 oster * rf_reconstruct.c -- code to perform on-line reconstruction
32 1.1 oster *
33 1.1 oster ************************************************************/
34 1.31 lukem
35 1.31 lukem #include <sys/cdefs.h>
36 1.81.2.1 yamt __KERNEL_RCSID(0, "$NetBSD: rf_reconstruct.c,v 1.81.2.1 2005/02/12 18:17:50 yamt Exp $");
37 1.1 oster
38 1.1 oster #include <sys/time.h>
39 1.1 oster #include <sys/buf.h>
40 1.1 oster #include <sys/errno.h>
41 1.5 oster
42 1.5 oster #include <sys/param.h>
43 1.5 oster #include <sys/systm.h>
44 1.5 oster #include <sys/proc.h>
45 1.5 oster #include <sys/ioctl.h>
46 1.5 oster #include <sys/fcntl.h>
47 1.5 oster #include <sys/vnode.h>
48 1.30 oster #include <dev/raidframe/raidframevar.h>
49 1.5 oster
50 1.1 oster #include "rf_raid.h"
51 1.1 oster #include "rf_reconutil.h"
52 1.1 oster #include "rf_revent.h"
53 1.1 oster #include "rf_reconbuffer.h"
54 1.1 oster #include "rf_acctrace.h"
55 1.1 oster #include "rf_etimer.h"
56 1.1 oster #include "rf_dag.h"
57 1.1 oster #include "rf_desc.h"
58 1.36 oster #include "rf_debugprint.h"
59 1.1 oster #include "rf_general.h"
60 1.1 oster #include "rf_driver.h"
61 1.1 oster #include "rf_utils.h"
62 1.1 oster #include "rf_shutdown.h"
63 1.1 oster
64 1.1 oster #include "rf_kintf.h"
65 1.1 oster
66 1.1 oster /* setting these to -1 causes them to be set to their default values if not set by debug options */
67 1.1 oster
68 1.41 oster #if RF_DEBUG_RECON
69 1.1 oster #define Dprintf(s) if (rf_reconDebug) rf_debug_printf(s,NULL,NULL,NULL,NULL,NULL,NULL,NULL,NULL)
70 1.1 oster #define Dprintf1(s,a) if (rf_reconDebug) rf_debug_printf(s,(void *)((unsigned long)a),NULL,NULL,NULL,NULL,NULL,NULL,NULL)
71 1.1 oster #define Dprintf2(s,a,b) if (rf_reconDebug) rf_debug_printf(s,(void *)((unsigned long)a),(void *)((unsigned long)b),NULL,NULL,NULL,NULL,NULL,NULL)
72 1.1 oster #define Dprintf3(s,a,b,c) if (rf_reconDebug) rf_debug_printf(s,(void *)((unsigned long)a),(void *)((unsigned long)b),(void *)((unsigned long)c),NULL,NULL,NULL,NULL,NULL)
73 1.1 oster #define Dprintf4(s,a,b,c,d) if (rf_reconDebug) rf_debug_printf(s,(void *)((unsigned long)a),(void *)((unsigned long)b),(void *)((unsigned long)c),(void *)((unsigned long)d),NULL,NULL,NULL,NULL)
74 1.1 oster #define Dprintf5(s,a,b,c,d,e) if (rf_reconDebug) rf_debug_printf(s,(void *)((unsigned long)a),(void *)((unsigned long)b),(void *)((unsigned long)c),(void *)((unsigned long)d),(void *)((unsigned long)e),NULL,NULL,NULL)
75 1.1 oster #define Dprintf6(s,a,b,c,d,e,f) if (rf_reconDebug) rf_debug_printf(s,(void *)((unsigned long)a),(void *)((unsigned long)b),(void *)((unsigned long)c),(void *)((unsigned long)d),(void *)((unsigned long)e),(void *)((unsigned long)f),NULL,NULL)
76 1.1 oster #define Dprintf7(s,a,b,c,d,e,f,g) if (rf_reconDebug) rf_debug_printf(s,(void *)((unsigned long)a),(void *)((unsigned long)b),(void *)((unsigned long)c),(void *)((unsigned long)d),(void *)((unsigned long)e),(void *)((unsigned long)f),(void *)((unsigned long)g),NULL)
77 1.1 oster
78 1.1 oster #define DDprintf1(s,a) if (rf_reconDebug) rf_debug_printf(s,(void *)((unsigned long)a),NULL,NULL,NULL,NULL,NULL,NULL,NULL)
79 1.1 oster #define DDprintf2(s,a,b) if (rf_reconDebug) rf_debug_printf(s,(void *)((unsigned long)a),(void *)((unsigned long)b),NULL,NULL,NULL,NULL,NULL,NULL)
80 1.33 oster
81 1.41 oster #else /* RF_DEBUG_RECON */
82 1.33 oster
83 1.33 oster #define Dprintf(s) {}
84 1.33 oster #define Dprintf1(s,a) {}
85 1.33 oster #define Dprintf2(s,a,b) {}
86 1.33 oster #define Dprintf3(s,a,b,c) {}
87 1.33 oster #define Dprintf4(s,a,b,c,d) {}
88 1.33 oster #define Dprintf5(s,a,b,c,d,e) {}
89 1.33 oster #define Dprintf6(s,a,b,c,d,e,f) {}
90 1.33 oster #define Dprintf7(s,a,b,c,d,e,f,g) {}
91 1.33 oster
92 1.33 oster #define DDprintf1(s,a) {}
93 1.33 oster #define DDprintf2(s,a,b) {}
94 1.33 oster
95 1.41 oster #endif /* RF_DEBUG_RECON */
96 1.33 oster
97 1.81.2.1 yamt #define RF_RECON_DONE_READS 1
98 1.81.2.1 yamt #define RF_RECON_READ_ERROR 2
99 1.81.2.1 yamt #define RF_RECON_WRITE_ERROR 3
100 1.81.2.1 yamt #define RF_RECON_READ_STOPPED 4
101 1.81.2.1 yamt
102 1.73 oster #define RF_MAX_FREE_RECONBUFFER 32
103 1.73 oster #define RF_MIN_FREE_RECONBUFFER 16
104 1.1 oster
105 1.69 oster static RF_RaidReconDesc_t *AllocRaidReconDesc(RF_Raid_t *, RF_RowCol_t,
106 1.69 oster RF_RaidDisk_t *, int, RF_RowCol_t);
107 1.69 oster static void FreeReconDesc(RF_RaidReconDesc_t *);
108 1.69 oster static int ProcessReconEvent(RF_Raid_t *, RF_ReconEvent_t *);
109 1.69 oster static int IssueNextReadRequest(RF_Raid_t *, RF_RowCol_t);
110 1.69 oster static int TryToRead(RF_Raid_t *, RF_RowCol_t);
111 1.69 oster static int ComputePSDiskOffsets(RF_Raid_t *, RF_StripeNum_t, RF_RowCol_t,
112 1.69 oster RF_SectorNum_t *, RF_SectorNum_t *, RF_RowCol_t *,
113 1.69 oster RF_SectorNum_t *);
114 1.69 oster static int IssueNextWriteRequest(RF_Raid_t *);
115 1.69 oster static int ReconReadDoneProc(void *, int);
116 1.69 oster static int ReconWriteDoneProc(void *, int);
117 1.69 oster static void CheckForNewMinHeadSep(RF_Raid_t *, RF_HeadSepLimit_t);
118 1.69 oster static int CheckHeadSeparation(RF_Raid_t *, RF_PerDiskReconCtrl_t *,
119 1.69 oster RF_RowCol_t, RF_HeadSepLimit_t,
120 1.69 oster RF_ReconUnitNum_t);
121 1.69 oster static int CheckForcedOrBlockedReconstruction(RF_Raid_t *,
122 1.69 oster RF_ReconParityStripeStatus_t *,
123 1.69 oster RF_PerDiskReconCtrl_t *,
124 1.69 oster RF_RowCol_t, RF_StripeNum_t,
125 1.69 oster RF_ReconUnitNum_t);
126 1.69 oster static void ForceReconReadDoneProc(void *, int);
127 1.1 oster static void rf_ShutdownReconstruction(void *);
128 1.1 oster
129 1.1 oster struct RF_ReconDoneProc_s {
130 1.4 oster void (*proc) (RF_Raid_t *, void *);
131 1.4 oster void *arg;
132 1.4 oster RF_ReconDoneProc_t *next;
133 1.1 oster };
134 1.1 oster
135 1.13 oster /**************************************************************************
136 1.1 oster *
137 1.1 oster * sets up the parameters that will be used by the reconstruction process
138 1.1 oster * currently there are none, except for those that the layout-specific
139 1.1 oster * configuration (e.g. rf_ConfigureDeclustered) routine sets up.
140 1.1 oster *
141 1.1 oster * in the kernel, we fire off the recon thread.
142 1.1 oster *
143 1.13 oster **************************************************************************/
144 1.4 oster static void
145 1.60 oster rf_ShutdownReconstruction(void *ignored)
146 1.4 oster {
147 1.74 oster pool_destroy(&rf_pools.reconbuffer);
148 1.4 oster }
149 1.4 oster
150 1.4 oster int
151 1.60 oster rf_ConfigureReconstruction(RF_ShutdownList_t **listp)
152 1.4 oster {
153 1.4 oster
154 1.74 oster rf_pool_init(&rf_pools.reconbuffer, sizeof(RF_ReconBuffer_t),
155 1.74 oster "rf_reconbuffer_pl", RF_MIN_FREE_RECONBUFFER, RF_MAX_FREE_RECONBUFFER);
156 1.66 oster rf_ShutdownCreate(listp, rf_ShutdownReconstruction, NULL);
157 1.66 oster
158 1.4 oster return (0);
159 1.4 oster }
160 1.4 oster
161 1.4 oster static RF_RaidReconDesc_t *
162 1.60 oster AllocRaidReconDesc(RF_Raid_t *raidPtr, RF_RowCol_t col,
163 1.60 oster RF_RaidDisk_t *spareDiskPtr, int numDisksDone,
164 1.60 oster RF_RowCol_t scol)
165 1.1 oster {
166 1.1 oster
167 1.4 oster RF_RaidReconDesc_t *reconDesc;
168 1.4 oster
169 1.80 oster RF_Malloc(reconDesc, sizeof(RF_RaidReconDesc_t),
170 1.80 oster (RF_RaidReconDesc_t *));
171 1.4 oster reconDesc->raidPtr = raidPtr;
172 1.4 oster reconDesc->col = col;
173 1.4 oster reconDesc->spareDiskPtr = spareDiskPtr;
174 1.4 oster reconDesc->numDisksDone = numDisksDone;
175 1.4 oster reconDesc->scol = scol;
176 1.4 oster reconDesc->next = NULL;
177 1.1 oster
178 1.4 oster return (reconDesc);
179 1.1 oster }
180 1.1 oster
181 1.4 oster static void
182 1.60 oster FreeReconDesc(RF_RaidReconDesc_t *reconDesc)
183 1.1 oster {
184 1.1 oster #if RF_RECON_STATS > 0
185 1.50 oster printf("raid%d: %lu recon event waits, %lu recon delays\n",
186 1.50 oster reconDesc->raidPtr->raidid,
187 1.50 oster (long) reconDesc->numReconEventWaits,
188 1.50 oster (long) reconDesc->numReconExecDelays);
189 1.4 oster #endif /* RF_RECON_STATS > 0 */
190 1.50 oster printf("raid%d: %lu max exec ticks\n",
191 1.50 oster reconDesc->raidPtr->raidid,
192 1.50 oster (long) reconDesc->maxReconExecTicks);
193 1.1 oster #if (RF_RECON_STATS > 0) || defined(KERNEL)
194 1.4 oster printf("\n");
195 1.4 oster #endif /* (RF_RECON_STATS > 0) || KERNEL */
196 1.80 oster RF_Free(reconDesc, sizeof(RF_RaidReconDesc_t));
197 1.1 oster }
198 1.1 oster
199 1.1 oster
200 1.13 oster /*****************************************************************************
201 1.1 oster *
202 1.1 oster * primary routine to reconstruct a failed disk. This should be called from
203 1.1 oster * within its own thread. It won't return until reconstruction completes,
204 1.1 oster * fails, or is aborted.
205 1.13 oster *****************************************************************************/
206 1.4 oster int
207 1.60 oster rf_ReconstructFailedDisk(RF_Raid_t *raidPtr, RF_RowCol_t col)
208 1.4 oster {
209 1.52 jdolecek const RF_LayoutSW_t *lp;
210 1.4 oster int rc;
211 1.4 oster
212 1.4 oster lp = raidPtr->Layout.map;
213 1.4 oster if (lp->SubmitReconBuffer) {
214 1.4 oster /*
215 1.4 oster * The current infrastructure only supports reconstructing one
216 1.4 oster * disk at a time for each array.
217 1.4 oster */
218 1.4 oster RF_LOCK_MUTEX(raidPtr->mutex);
219 1.4 oster while (raidPtr->reconInProgress) {
220 1.4 oster RF_WAIT_COND(raidPtr->waitForReconCond, raidPtr->mutex);
221 1.4 oster }
222 1.4 oster raidPtr->reconInProgress++;
223 1.4 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
224 1.57 oster rc = rf_ReconstructFailedDiskBasic(raidPtr, col);
225 1.6 oster RF_LOCK_MUTEX(raidPtr->mutex);
226 1.6 oster raidPtr->reconInProgress--;
227 1.6 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
228 1.4 oster } else {
229 1.4 oster RF_ERRORMSG1("RECON: no way to reconstruct failed disk for arch %c\n",
230 1.4 oster lp->parityConfig);
231 1.4 oster rc = EIO;
232 1.4 oster }
233 1.4 oster RF_SIGNAL_COND(raidPtr->waitForReconCond);
234 1.4 oster return (rc);
235 1.4 oster }
236 1.4 oster
237 1.4 oster int
238 1.60 oster rf_ReconstructFailedDiskBasic(RF_Raid_t *raidPtr, RF_RowCol_t col)
239 1.4 oster {
240 1.5 oster RF_ComponentLabel_t c_label;
241 1.4 oster RF_RaidDisk_t *spareDiskPtr = NULL;
242 1.4 oster RF_RaidReconDesc_t *reconDesc;
243 1.57 oster RF_RowCol_t scol;
244 1.4 oster int numDisksDone = 0, rc;
245 1.4 oster
246 1.4 oster /* first look for a spare drive onto which to reconstruct the data */
247 1.4 oster /* spare disk descriptors are stored in row 0. This may have to
248 1.4 oster * change eventually */
249 1.4 oster
250 1.4 oster RF_LOCK_MUTEX(raidPtr->mutex);
251 1.57 oster RF_ASSERT(raidPtr->Disks[col].status == rf_ds_failed);
252 1.72 oster #if RF_INCLUDE_PARITY_DECLUSTERING_DS > 0
253 1.4 oster if (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) {
254 1.57 oster if (raidPtr->status != rf_rs_degraded) {
255 1.57 oster RF_ERRORMSG1("Unable to reconstruct disk at col %d because status not degraded\n", col);
256 1.4 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
257 1.4 oster return (EINVAL);
258 1.4 oster }
259 1.4 oster scol = (-1);
260 1.4 oster } else {
261 1.72 oster #endif
262 1.4 oster for (scol = raidPtr->numCol; scol < raidPtr->numCol + raidPtr->numSpare; scol++) {
263 1.57 oster if (raidPtr->Disks[scol].status == rf_ds_spare) {
264 1.57 oster spareDiskPtr = &raidPtr->Disks[scol];
265 1.4 oster spareDiskPtr->status = rf_ds_used_spare;
266 1.4 oster break;
267 1.4 oster }
268 1.4 oster }
269 1.4 oster if (!spareDiskPtr) {
270 1.57 oster RF_ERRORMSG1("Unable to reconstruct disk at col %d because no spares are available\n", col);
271 1.4 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
272 1.4 oster return (ENOSPC);
273 1.4 oster }
274 1.57 oster printf("RECON: initiating reconstruction on col %d -> spare at col %d\n", col, scol);
275 1.72 oster #if RF_INCLUDE_PARITY_DECLUSTERING_DS > 0
276 1.4 oster }
277 1.72 oster #endif
278 1.4 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
279 1.1 oster
280 1.57 oster reconDesc = AllocRaidReconDesc((void *) raidPtr, col, spareDiskPtr, numDisksDone, scol);
281 1.4 oster raidPtr->reconDesc = (void *) reconDesc;
282 1.1 oster #if RF_RECON_STATS > 0
283 1.4 oster reconDesc->hsStallCount = 0;
284 1.4 oster reconDesc->numReconExecDelays = 0;
285 1.4 oster reconDesc->numReconEventWaits = 0;
286 1.4 oster #endif /* RF_RECON_STATS > 0 */
287 1.4 oster reconDesc->reconExecTimerRunning = 0;
288 1.4 oster reconDesc->reconExecTicks = 0;
289 1.4 oster reconDesc->maxReconExecTicks = 0;
290 1.4 oster rc = rf_ContinueReconstructFailedDisk(reconDesc);
291 1.5 oster
292 1.5 oster if (!rc) {
293 1.5 oster /* fix up the component label */
294 1.5 oster /* Don't actually need the read here.. */
295 1.5 oster raidread_component_label(
296 1.57 oster raidPtr->raid_cinfo[scol].ci_dev,
297 1.57 oster raidPtr->raid_cinfo[scol].ci_vp,
298 1.5 oster &c_label);
299 1.5 oster
300 1.15 oster raid_init_component_label( raidPtr, &c_label);
301 1.57 oster c_label.row = 0;
302 1.5 oster c_label.column = col;
303 1.5 oster c_label.clean = RF_RAID_DIRTY;
304 1.5 oster c_label.status = rf_ds_optimal;
305 1.57 oster c_label.partitionSize = raidPtr->Disks[scol].partitionSize;
306 1.15 oster
307 1.28 oster /* We've just done a rebuild based on all the other
308 1.28 oster disks, so at this point the parity is known to be
309 1.28 oster clean, even if it wasn't before. */
310 1.28 oster
311 1.28 oster /* XXX doesn't hold for RAID 6!!*/
312 1.28 oster
313 1.48 oster RF_LOCK_MUTEX(raidPtr->mutex);
314 1.28 oster raidPtr->parity_good = RF_RAID_CLEAN;
315 1.48 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
316 1.28 oster
317 1.15 oster /* XXXX MORE NEEDED HERE */
318 1.5 oster
319 1.5 oster raidwrite_component_label(
320 1.57 oster raidPtr->raid_cinfo[scol].ci_dev,
321 1.57 oster raidPtr->raid_cinfo[scol].ci_vp,
322 1.5 oster &c_label);
323 1.49 oster
324 1.81.2.1 yamt } else {
325 1.81.2.1 yamt /* Reconstruct failed. */
326 1.81.2.1 yamt
327 1.81.2.1 yamt RF_LOCK_MUTEX(raidPtr->mutex);
328 1.81.2.1 yamt /* Failed disk goes back to "failed" status */
329 1.81.2.1 yamt raidPtr->Disks[col].status = rf_ds_failed;
330 1.81.2.1 yamt
331 1.81.2.1 yamt /* Spare disk goes back to "spare" status. */
332 1.81.2.1 yamt spareDiskPtr->status = rf_ds_spare;
333 1.81.2.1 yamt RF_UNLOCK_MUTEX(raidPtr->mutex);
334 1.49 oster
335 1.5 oster }
336 1.81.2.1 yamt rf_update_component_labels(raidPtr, RF_NORMAL_COMPONENT_UPDATE);
337 1.5 oster return (rc);
338 1.5 oster }
339 1.5 oster
340 1.5 oster /*
341 1.5 oster
342 1.5 oster Allow reconstructing a disk in-place -- i.e. component /dev/sd2e goes AWOL,
343 1.5 oster and you don't get a spare until the next Monday. With this function
344 1.5 oster (and hot-swappable drives) you can now put your new disk containing
345 1.5 oster /dev/sd2e on the bus, scsictl it alive, and then use raidctl(8) to
346 1.5 oster rebuild the data "on the spot".
347 1.5 oster
348 1.5 oster */
349 1.5 oster
350 1.5 oster int
351 1.60 oster rf_ReconstructInPlace(RF_Raid_t *raidPtr, RF_RowCol_t col)
352 1.5 oster {
353 1.5 oster RF_RaidDisk_t *spareDiskPtr = NULL;
354 1.5 oster RF_RaidReconDesc_t *reconDesc;
355 1.52 jdolecek const RF_LayoutSW_t *lp;
356 1.5 oster RF_ComponentLabel_t c_label;
357 1.5 oster int numDisksDone = 0, rc;
358 1.5 oster struct partinfo dpart;
359 1.5 oster struct vnode *vp;
360 1.5 oster struct vattr va;
361 1.56 fvdl struct proc *proc;
362 1.5 oster int retcode;
363 1.21 oster int ac;
364 1.5 oster
365 1.5 oster lp = raidPtr->Layout.map;
366 1.61 oster if (!lp->SubmitReconBuffer) {
367 1.61 oster RF_ERRORMSG1("RECON: no way to reconstruct failed disk for arch %c\n",
368 1.61 oster lp->parityConfig);
369 1.61 oster /* wakeup anyone who might be waiting to do a reconstruct */
370 1.61 oster RF_SIGNAL_COND(raidPtr->waitForReconCond);
371 1.61 oster return(EIO);
372 1.62 oster }
373 1.5 oster
374 1.62 oster /*
375 1.62 oster * The current infrastructure only supports reconstructing one
376 1.62 oster * disk at a time for each array.
377 1.62 oster */
378 1.62 oster RF_LOCK_MUTEX(raidPtr->mutex);
379 1.5 oster
380 1.62 oster if (raidPtr->Disks[col].status != rf_ds_failed) {
381 1.62 oster /* "It's gone..." */
382 1.62 oster raidPtr->numFailures++;
383 1.62 oster raidPtr->Disks[col].status = rf_ds_failed;
384 1.62 oster raidPtr->status = rf_rs_degraded;
385 1.62 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
386 1.62 oster rf_update_component_labels(raidPtr,
387 1.62 oster RF_NORMAL_COMPONENT_UPDATE);
388 1.62 oster RF_LOCK_MUTEX(raidPtr->mutex);
389 1.62 oster }
390 1.62 oster
391 1.62 oster while (raidPtr->reconInProgress) {
392 1.62 oster RF_WAIT_COND(raidPtr->waitForReconCond, raidPtr->mutex);
393 1.62 oster }
394 1.62 oster
395 1.62 oster raidPtr->reconInProgress++;
396 1.64 oster
397 1.62 oster /* first look for a spare drive onto which to reconstruct the
398 1.62 oster data. spare disk descriptors are stored in row 0. This
399 1.62 oster may have to change eventually */
400 1.62 oster
401 1.62 oster /* Actually, we don't care if it's failed or not... On a RAID
402 1.62 oster set with correct parity, this function should be callable
403 1.62 oster on any component without ill affects. */
404 1.62 oster /* RF_ASSERT(raidPtr->Disks[col].status == rf_ds_failed); */
405 1.62 oster
406 1.72 oster #if RF_INCLUDE_PARITY_DECLUSTERING_DS > 0
407 1.62 oster if (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) {
408 1.62 oster RF_ERRORMSG1("Unable to reconstruct to disk at col %d: operation not supported for RF_DISTRIBUTE_SPARE\n", col);
409 1.62 oster
410 1.62 oster raidPtr->reconInProgress--;
411 1.62 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
412 1.64 oster RF_SIGNAL_COND(raidPtr->waitForReconCond);
413 1.62 oster return (EINVAL);
414 1.62 oster }
415 1.72 oster #endif
416 1.62 oster proc = raidPtr->engine_thread;
417 1.62 oster
418 1.62 oster /* This device may have been opened successfully the
419 1.62 oster first time. Close it before trying to open it again.. */
420 1.62 oster
421 1.62 oster if (raidPtr->raid_cinfo[col].ci_vp != NULL) {
422 1.37 oster #if 0
423 1.62 oster printf("Closed the open device: %s\n",
424 1.62 oster raidPtr->Disks[col].devname);
425 1.37 oster #endif
426 1.62 oster vp = raidPtr->raid_cinfo[col].ci_vp;
427 1.62 oster ac = raidPtr->Disks[col].auto_configured;
428 1.62 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
429 1.62 oster rf_close_component(raidPtr, vp, ac);
430 1.62 oster RF_LOCK_MUTEX(raidPtr->mutex);
431 1.62 oster raidPtr->raid_cinfo[col].ci_vp = NULL;
432 1.62 oster }
433 1.62 oster /* note that this disk was *not* auto_configured (any longer)*/
434 1.62 oster raidPtr->Disks[col].auto_configured = 0;
435 1.62 oster
436 1.37 oster #if 0
437 1.62 oster printf("About to (re-)open the device for rebuilding: %s\n",
438 1.62 oster raidPtr->Disks[col].devname);
439 1.37 oster #endif
440 1.62 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
441 1.62 oster retcode = raidlookup(raidPtr->Disks[col].devname, proc, &vp);
442 1.62 oster
443 1.62 oster if (retcode) {
444 1.62 oster printf("raid%d: rebuilding: raidlookup on device: %s failed: %d!\n",raidPtr->raidid,
445 1.62 oster raidPtr->Disks[col].devname, retcode);
446 1.62 oster
447 1.62 oster /* the component isn't responding properly...
448 1.62 oster must be still dead :-( */
449 1.62 oster RF_LOCK_MUTEX(raidPtr->mutex);
450 1.62 oster raidPtr->reconInProgress--;
451 1.48 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
452 1.64 oster RF_SIGNAL_COND(raidPtr->waitForReconCond);
453 1.62 oster return(retcode);
454 1.63 oster }
455 1.63 oster
456 1.63 oster /* Ok, so we can at least do a lookup...
457 1.63 oster How about actually getting a vp for it? */
458 1.63 oster
459 1.63 oster if ((retcode = VOP_GETATTR(vp, &va, proc->p_ucred, proc)) != 0) {
460 1.63 oster RF_LOCK_MUTEX(raidPtr->mutex);
461 1.63 oster raidPtr->reconInProgress--;
462 1.63 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
463 1.64 oster RF_SIGNAL_COND(raidPtr->waitForReconCond);
464 1.63 oster return(retcode);
465 1.63 oster }
466 1.63 oster
467 1.64 oster retcode = VOP_IOCTL(vp, DIOCGPART, &dpart, FREAD, proc->p_ucred, proc);
468 1.63 oster if (retcode) {
469 1.62 oster RF_LOCK_MUTEX(raidPtr->mutex);
470 1.63 oster raidPtr->reconInProgress--;
471 1.62 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
472 1.64 oster RF_SIGNAL_COND(raidPtr->waitForReconCond);
473 1.63 oster return(retcode);
474 1.62 oster }
475 1.63 oster RF_LOCK_MUTEX(raidPtr->mutex);
476 1.63 oster raidPtr->Disks[col].blockSize = dpart.disklab->d_secsize;
477 1.62 oster
478 1.64 oster raidPtr->Disks[col].numBlocks = dpart.part->p_size -
479 1.63 oster rf_protectedSectors;
480 1.62 oster
481 1.63 oster raidPtr->raid_cinfo[col].ci_vp = vp;
482 1.63 oster raidPtr->raid_cinfo[col].ci_dev = va.va_rdev;
483 1.63 oster
484 1.63 oster raidPtr->Disks[col].dev = va.va_rdev;
485 1.63 oster
486 1.63 oster /* we allow the user to specify that only a fraction
487 1.63 oster of the disks should be used this is just for debug:
488 1.63 oster it speeds up * the parity scan */
489 1.63 oster raidPtr->Disks[col].numBlocks = raidPtr->Disks[col].numBlocks *
490 1.63 oster rf_sizePercentage / 100;
491 1.63 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
492 1.62 oster
493 1.62 oster spareDiskPtr = &raidPtr->Disks[col];
494 1.62 oster spareDiskPtr->status = rf_ds_used_spare;
495 1.62 oster
496 1.62 oster printf("raid%d: initiating in-place reconstruction on column %d\n",
497 1.62 oster raidPtr->raidid, col);
498 1.5 oster
499 1.62 oster reconDesc = AllocRaidReconDesc((void *) raidPtr, col, spareDiskPtr,
500 1.62 oster numDisksDone, col);
501 1.62 oster raidPtr->reconDesc = (void *) reconDesc;
502 1.5 oster #if RF_RECON_STATS > 0
503 1.62 oster reconDesc->hsStallCount = 0;
504 1.62 oster reconDesc->numReconExecDelays = 0;
505 1.62 oster reconDesc->numReconEventWaits = 0;
506 1.5 oster #endif /* RF_RECON_STATS > 0 */
507 1.62 oster reconDesc->reconExecTimerRunning = 0;
508 1.62 oster reconDesc->reconExecTicks = 0;
509 1.62 oster reconDesc->maxReconExecTicks = 0;
510 1.62 oster rc = rf_ContinueReconstructFailedDisk(reconDesc);
511 1.62 oster
512 1.5 oster if (!rc) {
513 1.48 oster RF_LOCK_MUTEX(raidPtr->mutex);
514 1.5 oster /* Need to set these here, as at this point it'll be claiming
515 1.5 oster that the disk is in rf_ds_spared! But we know better :-) */
516 1.5 oster
517 1.57 oster raidPtr->Disks[col].status = rf_ds_optimal;
518 1.57 oster raidPtr->status = rf_rs_optimal;
519 1.48 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
520 1.5 oster
521 1.5 oster /* fix up the component label */
522 1.5 oster /* Don't actually need the read here.. */
523 1.57 oster raidread_component_label(raidPtr->raid_cinfo[col].ci_dev,
524 1.57 oster raidPtr->raid_cinfo[col].ci_vp,
525 1.5 oster &c_label);
526 1.16 oster
527 1.48 oster RF_LOCK_MUTEX(raidPtr->mutex);
528 1.16 oster raid_init_component_label(raidPtr, &c_label);
529 1.16 oster
530 1.57 oster c_label.row = 0;
531 1.5 oster c_label.column = col;
532 1.28 oster
533 1.28 oster /* We've just done a rebuild based on all the other
534 1.28 oster disks, so at this point the parity is known to be
535 1.28 oster clean, even if it wasn't before. */
536 1.28 oster
537 1.28 oster /* XXX doesn't hold for RAID 6!!*/
538 1.28 oster
539 1.28 oster raidPtr->parity_good = RF_RAID_CLEAN;
540 1.48 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
541 1.28 oster
542 1.57 oster raidwrite_component_label(raidPtr->raid_cinfo[col].ci_dev,
543 1.57 oster raidPtr->raid_cinfo[col].ci_vp,
544 1.5 oster &c_label);
545 1.49 oster
546 1.81.2.1 yamt } else {
547 1.81.2.1 yamt /* Reconstruct-in-place failed. Disk goes back to
548 1.81.2.1 yamt "failed" status, regardless of what it was before. */
549 1.81.2.1 yamt RF_LOCK_MUTEX(raidPtr->mutex);
550 1.81.2.1 yamt raidPtr->Disks[col].status = rf_ds_failed;
551 1.81.2.1 yamt RF_UNLOCK_MUTEX(raidPtr->mutex);
552 1.5 oster }
553 1.81.2.1 yamt
554 1.81.2.1 yamt rf_update_component_labels(raidPtr, RF_NORMAL_COMPONENT_UPDATE);
555 1.81.2.1 yamt
556 1.81.2.1 yamt RF_LOCK_MUTEX(raidPtr->mutex);
557 1.81.2.1 yamt raidPtr->reconInProgress--;
558 1.81.2.1 yamt RF_UNLOCK_MUTEX(raidPtr->mutex);
559 1.81.2.1 yamt
560 1.5 oster RF_SIGNAL_COND(raidPtr->waitForReconCond);
561 1.4 oster return (rc);
562 1.4 oster }
563 1.4 oster
564 1.4 oster
565 1.4 oster int
566 1.60 oster rf_ContinueReconstructFailedDisk(RF_RaidReconDesc_t *reconDesc)
567 1.4 oster {
568 1.4 oster RF_Raid_t *raidPtr = reconDesc->raidPtr;
569 1.4 oster RF_RowCol_t col = reconDesc->col;
570 1.4 oster RF_RowCol_t scol = reconDesc->scol;
571 1.4 oster RF_ReconMap_t *mapPtr;
572 1.46 oster RF_ReconCtrl_t *tmp_reconctrl;
573 1.4 oster RF_ReconEvent_t *event;
574 1.81.2.1 yamt RF_CallbackDesc_t *p;
575 1.4 oster struct timeval etime, elpsd;
576 1.4 oster unsigned long xor_s, xor_resid_us;
577 1.54 simonb int i, ds;
578 1.81.2.1 yamt int status;
579 1.81.2.1 yamt int recon_error, write_error;
580 1.4 oster
581 1.78 oster raidPtr->accumXorTimeUs = 0;
582 1.67 oster #if RF_ACC_TRACE > 0
583 1.78 oster /* create one trace record per physical disk */
584 1.78 oster RF_Malloc(raidPtr->recon_tracerecs, raidPtr->numCol * sizeof(RF_AccTraceEntry_t), (RF_AccTraceEntry_t *));
585 1.67 oster #endif
586 1.78 oster
587 1.78 oster /* quiesce the array prior to starting recon. this is needed
588 1.78 oster * to assure no nasty interactions with pending user writes.
589 1.78 oster * We need to do this before we change the disk or row status. */
590 1.78 oster
591 1.78 oster Dprintf("RECON: begin request suspend\n");
592 1.78 oster rf_SuspendNewRequestsAndWait(raidPtr);
593 1.78 oster Dprintf("RECON: end request suspend\n");
594 1.78 oster
595 1.78 oster /* allocate our RF_ReconCTRL_t before we protect raidPtr->reconControl[row] */
596 1.78 oster tmp_reconctrl = rf_MakeReconControl(reconDesc, col, scol);
597 1.78 oster
598 1.78 oster RF_LOCK_MUTEX(raidPtr->mutex);
599 1.78 oster
600 1.78 oster /* create the reconstruction control pointer and install it in
601 1.78 oster * the right slot */
602 1.78 oster raidPtr->reconControl = tmp_reconctrl;
603 1.78 oster mapPtr = raidPtr->reconControl->reconMap;
604 1.78 oster raidPtr->status = rf_rs_reconstructing;
605 1.78 oster raidPtr->Disks[col].status = rf_ds_reconstructing;
606 1.78 oster raidPtr->Disks[col].spareCol = scol;
607 1.78 oster
608 1.78 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
609 1.78 oster
610 1.78 oster RF_GETTIME(raidPtr->reconControl->starttime);
611 1.78 oster
612 1.78 oster /* now start up the actual reconstruction: issue a read for
613 1.78 oster * each surviving disk */
614 1.78 oster
615 1.78 oster reconDesc->numDisksDone = 0;
616 1.78 oster for (i = 0; i < raidPtr->numCol; i++) {
617 1.78 oster if (i != col) {
618 1.78 oster /* find and issue the next I/O on the
619 1.78 oster * indicated disk */
620 1.78 oster if (IssueNextReadRequest(raidPtr, i)) {
621 1.78 oster Dprintf1("RECON: done issuing for c%d\n", i);
622 1.78 oster reconDesc->numDisksDone++;
623 1.4 oster }
624 1.4 oster }
625 1.78 oster }
626 1.4 oster
627 1.78 oster Dprintf("RECON: resume requests\n");
628 1.78 oster rf_ResumeNewRequests(raidPtr);
629 1.78 oster
630 1.78 oster /* process reconstruction events until all disks report that
631 1.78 oster * they've completed all work */
632 1.4 oster
633 1.78 oster mapPtr = raidPtr->reconControl->reconMap;
634 1.81.2.1 yamt recon_error = 0;
635 1.81.2.1 yamt write_error = 0;
636 1.81.2.1 yamt
637 1.78 oster while (reconDesc->numDisksDone < raidPtr->numCol - 1) {
638 1.78 oster
639 1.78 oster event = rf_GetNextReconEvent(reconDesc);
640 1.81.2.1 yamt status = ProcessReconEvent(raidPtr, event);
641 1.81.2.1 yamt
642 1.81.2.1 yamt /* the normal case is that a read completes, and all is well. */
643 1.81.2.1 yamt if (status == RF_RECON_DONE_READS) {
644 1.78 oster reconDesc->numDisksDone++;
645 1.81.2.1 yamt } else if ((status == RF_RECON_READ_ERROR) ||
646 1.81.2.1 yamt (status == RF_RECON_WRITE_ERROR)) {
647 1.81.2.1 yamt /* an error was encountered while reconstructing...
648 1.81.2.1 yamt Pretend we've finished this disk.
649 1.81.2.1 yamt */
650 1.81.2.1 yamt recon_error = 1;
651 1.81.2.1 yamt raidPtr->reconControl->error = 1;
652 1.81.2.1 yamt
653 1.81.2.1 yamt /* bump the numDisksDone count for reads,
654 1.81.2.1 yamt but not for writes */
655 1.81.2.1 yamt if (status == RF_RECON_READ_ERROR)
656 1.81.2.1 yamt reconDesc->numDisksDone++;
657 1.81.2.1 yamt
658 1.81.2.1 yamt /* write errors are special -- when we are
659 1.81.2.1 yamt done dealing with the reads that are
660 1.81.2.1 yamt finished, we don't want to wait for any
661 1.81.2.1 yamt writes */
662 1.81.2.1 yamt if (status == RF_RECON_WRITE_ERROR)
663 1.81.2.1 yamt write_error = 1;
664 1.81.2.1 yamt
665 1.81.2.1 yamt } else if (status == RF_RECON_READ_STOPPED) {
666 1.81.2.1 yamt /* count this component as being "done" */
667 1.81.2.1 yamt reconDesc->numDisksDone++;
668 1.81.2.1 yamt }
669 1.81.2.1 yamt
670 1.81.2.1 yamt if (recon_error) {
671 1.81.2.1 yamt
672 1.81.2.1 yamt /* make sure any stragglers are woken up so that
673 1.81.2.1 yamt their theads will complete, and we can get out
674 1.81.2.1 yamt of here with all IO processed */
675 1.81.2.1 yamt
676 1.81.2.1 yamt while (raidPtr->reconControl->headSepCBList) {
677 1.81.2.1 yamt p = raidPtr->reconControl->headSepCBList;
678 1.81.2.1 yamt raidPtr->reconControl->headSepCBList = p->next;
679 1.81.2.1 yamt p->next = NULL;
680 1.81.2.1 yamt rf_CauseReconEvent(raidPtr, p->col, NULL, RF_REVENT_HEADSEPCLEAR);
681 1.81.2.1 yamt rf_FreeCallbackDesc(p);
682 1.81.2.1 yamt }
683 1.81.2.1 yamt }
684 1.81.2.1 yamt
685 1.78 oster raidPtr->reconControl->numRUsTotal =
686 1.78 oster mapPtr->totalRUs;
687 1.78 oster raidPtr->reconControl->numRUsComplete =
688 1.78 oster mapPtr->totalRUs -
689 1.78 oster rf_UnitsLeftToReconstruct(mapPtr);
690 1.81.2.1 yamt
691 1.65 oster #if RF_DEBUG_RECON
692 1.78 oster raidPtr->reconControl->percentComplete =
693 1.78 oster (raidPtr->reconControl->numRUsComplete * 100 / raidPtr->reconControl->numRUsTotal);
694 1.78 oster if (rf_prReconSched) {
695 1.78 oster rf_PrintReconSchedule(raidPtr->reconControl->reconMap, &(raidPtr->reconControl->starttime));
696 1.78 oster }
697 1.41 oster #endif
698 1.78 oster }
699 1.78 oster
700 1.78 oster mapPtr = raidPtr->reconControl->reconMap;
701 1.78 oster if (rf_reconDebug) {
702 1.78 oster printf("RECON: all reads completed\n");
703 1.78 oster }
704 1.78 oster /* at this point all the reads have completed. We now wait
705 1.78 oster * for any pending writes to complete, and then we're done */
706 1.81.2.1 yamt
707 1.81.2.1 yamt while (!recon_error && rf_UnitsLeftToReconstruct(raidPtr->reconControl->reconMap) > 0) {
708 1.78 oster
709 1.78 oster event = rf_GetNextReconEvent(reconDesc);
710 1.81.2.1 yamt status = ProcessReconEvent(raidPtr, event);
711 1.81.2.1 yamt
712 1.81.2.1 yamt if (status == RF_RECON_WRITE_ERROR) {
713 1.81.2.1 yamt recon_error = 1;
714 1.81.2.1 yamt raidPtr->reconControl->error = 1;
715 1.81.2.1 yamt /* an error was encountered at the very end... bail */
716 1.81.2.1 yamt } else {
717 1.78 oster #if RF_DEBUG_RECON
718 1.81.2.1 yamt raidPtr->reconControl->percentComplete = 100 - (rf_UnitsLeftToReconstruct(mapPtr) * 100 / mapPtr->totalRUs);
719 1.81.2.1 yamt if (rf_prReconSched) {
720 1.81.2.1 yamt rf_PrintReconSchedule(raidPtr->reconControl->reconMap, &(raidPtr->reconControl->starttime));
721 1.81.2.1 yamt }
722 1.81.2.1 yamt #endif
723 1.81.2.1 yamt }
724 1.81.2.1 yamt }
725 1.81.2.1 yamt
726 1.81.2.1 yamt if (recon_error) {
727 1.81.2.1 yamt /* we've encountered an error in reconstructing. */
728 1.81.2.1 yamt printf("raid%d: reconstruction failed.\n", raidPtr->raidid);
729 1.81.2.1 yamt
730 1.81.2.1 yamt /* we start by blocking IO to the RAID set. */
731 1.81.2.1 yamt rf_SuspendNewRequestsAndWait(raidPtr);
732 1.81.2.1 yamt
733 1.81.2.1 yamt RF_LOCK_MUTEX(raidPtr->mutex);
734 1.81.2.1 yamt /* mark set as being degraded, rather than
735 1.81.2.1 yamt rf_rs_reconstructing as we were before the problem.
736 1.81.2.1 yamt After this is done we can update status of the
737 1.81.2.1 yamt component disks without worrying about someone
738 1.81.2.1 yamt trying to read from a failed component.
739 1.81.2.1 yamt */
740 1.81.2.1 yamt raidPtr->status = rf_rs_degraded;
741 1.81.2.1 yamt RF_UNLOCK_MUTEX(raidPtr->mutex);
742 1.81.2.1 yamt
743 1.81.2.1 yamt /* resume IO */
744 1.81.2.1 yamt rf_ResumeNewRequests(raidPtr);
745 1.81.2.1 yamt
746 1.81.2.1 yamt /* At this point there are two cases:
747 1.81.2.1 yamt 1) If we've experienced a read error, then we've
748 1.81.2.1 yamt already waited for all the reads we're going to get,
749 1.81.2.1 yamt and we just need to wait for the writes.
750 1.81.2.1 yamt
751 1.81.2.1 yamt 2) If we've experienced a write error, we've also
752 1.81.2.1 yamt already waited for all the reads to complete,
753 1.81.2.1 yamt but there is little point in waiting for the writes --
754 1.81.2.1 yamt when they do complete, they will just be ignored.
755 1.81.2.1 yamt
756 1.81.2.1 yamt So we just wait for writes to complete if we didn't have a
757 1.81.2.1 yamt write error.
758 1.81.2.1 yamt */
759 1.81.2.1 yamt
760 1.81.2.1 yamt if (!write_error) {
761 1.81.2.1 yamt /* wait for writes to complete */
762 1.81.2.1 yamt while (raidPtr->reconControl->pending_writes > 0) {
763 1.81.2.1 yamt
764 1.81.2.1 yamt event = rf_GetNextReconEvent(reconDesc);
765 1.81.2.1 yamt status = ProcessReconEvent(raidPtr, event);
766 1.81.2.1 yamt
767 1.81.2.1 yamt if (status == RF_RECON_WRITE_ERROR) {
768 1.81.2.1 yamt raidPtr->reconControl->error = 1;
769 1.81.2.1 yamt /* an error was encountered at the very end... bail.
770 1.81.2.1 yamt This will be very bad news for the user, since
771 1.81.2.1 yamt at this point there will have been a read error
772 1.81.2.1 yamt on one component, and a write error on another!
773 1.81.2.1 yamt */
774 1.81.2.1 yamt break;
775 1.81.2.1 yamt }
776 1.81.2.1 yamt }
777 1.4 oster }
778 1.81.2.1 yamt
779 1.81.2.1 yamt
780 1.81.2.1 yamt /* cleanup */
781 1.81.2.1 yamt
782 1.81.2.1 yamt /* drain the event queue - after waiting for the writes above,
783 1.81.2.1 yamt there shouldn't be much (if anything!) left in the queue. */
784 1.81.2.1 yamt
785 1.81.2.1 yamt rf_DrainReconEventQueue(reconDesc);
786 1.81.2.1 yamt
787 1.81.2.1 yamt /* XXX As much as we'd like to free the recon control structure
788 1.81.2.1 yamt and the reconDesc, we have no way of knowing if/when those will
789 1.81.2.1 yamt be touched by IO that has yet to occur. It is rather poor to be
790 1.81.2.1 yamt basically causing a 'memory leak' here, but there doesn't seem to be
791 1.81.2.1 yamt a cleaner alternative at this time. Perhaps when the reconstruct code
792 1.81.2.1 yamt gets a makeover this problem will go away.
793 1.81.2.1 yamt */
794 1.81.2.1 yamt #if 0
795 1.81.2.1 yamt rf_FreeReconControl(raidPtr);
796 1.41 oster #endif
797 1.81.2.1 yamt
798 1.81.2.1 yamt #if RF_ACC_TRACE > 0
799 1.81.2.1 yamt RF_Free(raidPtr->recon_tracerecs, raidPtr->numCol * sizeof(RF_AccTraceEntry_t));
800 1.81.2.1 yamt #endif
801 1.81.2.1 yamt /* XXX see comment above */
802 1.81.2.1 yamt #if 0
803 1.81.2.1 yamt FreeReconDesc(reconDesc);
804 1.81.2.1 yamt #endif
805 1.81.2.1 yamt
806 1.81.2.1 yamt return (1);
807 1.78 oster }
808 1.14 oster
809 1.78 oster /* Success: mark the dead disk as reconstructed. We quiesce
810 1.78 oster * the array here to assure no nasty interactions with pending
811 1.78 oster * user accesses when we free up the psstatus structure as
812 1.78 oster * part of FreeReconControl() */
813 1.78 oster
814 1.78 oster rf_SuspendNewRequestsAndWait(raidPtr);
815 1.78 oster
816 1.78 oster RF_LOCK_MUTEX(raidPtr->mutex);
817 1.78 oster raidPtr->numFailures--;
818 1.78 oster ds = (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE);
819 1.78 oster raidPtr->Disks[col].status = (ds) ? rf_ds_dist_spared : rf_ds_spared;
820 1.78 oster raidPtr->status = (ds) ? rf_rs_reconfigured : rf_rs_optimal;
821 1.78 oster RF_UNLOCK_MUTEX(raidPtr->mutex);
822 1.78 oster RF_GETTIME(etime);
823 1.78 oster RF_TIMEVAL_DIFF(&(raidPtr->reconControl->starttime), &etime, &elpsd);
824 1.78 oster
825 1.78 oster rf_ResumeNewRequests(raidPtr);
826 1.78 oster
827 1.78 oster printf("raid%d: Reconstruction of disk at col %d completed\n",
828 1.78 oster raidPtr->raidid, col);
829 1.78 oster xor_s = raidPtr->accumXorTimeUs / 1000000;
830 1.78 oster xor_resid_us = raidPtr->accumXorTimeUs % 1000000;
831 1.78 oster printf("raid%d: Recon time was %d.%06d seconds, accumulated XOR time was %ld us (%ld.%06ld)\n",
832 1.78 oster raidPtr->raidid,
833 1.78 oster (int) elpsd.tv_sec, (int) elpsd.tv_usec,
834 1.78 oster raidPtr->accumXorTimeUs, xor_s, xor_resid_us);
835 1.78 oster printf("raid%d: (start time %d sec %d usec, end time %d sec %d usec)\n",
836 1.78 oster raidPtr->raidid,
837 1.78 oster (int) raidPtr->reconControl->starttime.tv_sec,
838 1.78 oster (int) raidPtr->reconControl->starttime.tv_usec,
839 1.78 oster (int) etime.tv_sec, (int) etime.tv_usec);
840 1.1 oster #if RF_RECON_STATS > 0
841 1.78 oster printf("raid%d: Total head-sep stall count was %d\n",
842 1.78 oster raidPtr->raidid, (int) reconDesc->hsStallCount);
843 1.4 oster #endif /* RF_RECON_STATS > 0 */
844 1.78 oster rf_FreeReconControl(raidPtr);
845 1.67 oster #if RF_ACC_TRACE > 0
846 1.78 oster RF_Free(raidPtr->recon_tracerecs, raidPtr->numCol * sizeof(RF_AccTraceEntry_t));
847 1.67 oster #endif
848 1.78 oster FreeReconDesc(reconDesc);
849 1.78 oster
850 1.4 oster return (0);
851 1.81.2.1 yamt
852 1.1 oster }
853 1.13 oster /*****************************************************************************
854 1.1 oster * do the right thing upon each reconstruction event.
855 1.13 oster *****************************************************************************/
856 1.4 oster static int
857 1.60 oster ProcessReconEvent(RF_Raid_t *raidPtr, RF_ReconEvent_t *event)
858 1.4 oster {
859 1.4 oster int retcode = 0, submitblocked;
860 1.4 oster RF_ReconBuffer_t *rbuf;
861 1.4 oster RF_SectorCount_t sectorsPerRU;
862 1.4 oster
863 1.81.2.1 yamt retcode = RF_RECON_READ_STOPPED;
864 1.81.2.1 yamt
865 1.4 oster Dprintf1("RECON: ProcessReconEvent type %d\n", event->type);
866 1.4 oster switch (event->type) {
867 1.4 oster
868 1.4 oster /* a read I/O has completed */
869 1.4 oster case RF_REVENT_READDONE:
870 1.57 oster rbuf = raidPtr->reconControl->perDiskInfo[event->col].rbuf;
871 1.57 oster Dprintf2("RECON: READDONE EVENT: col %d psid %ld\n",
872 1.57 oster event->col, rbuf->parityStripeID);
873 1.4 oster Dprintf7("RECON: done read psid %ld buf %lx %02x %02x %02x %02x %02x\n",
874 1.4 oster rbuf->parityStripeID, rbuf->buffer, rbuf->buffer[0] & 0xff, rbuf->buffer[1] & 0xff,
875 1.4 oster rbuf->buffer[2] & 0xff, rbuf->buffer[3] & 0xff, rbuf->buffer[4] & 0xff);
876 1.4 oster rf_FreeDiskQueueData((RF_DiskQueueData_t *) rbuf->arg);
877 1.81.2.1 yamt if (!raidPtr->reconControl->error) {
878 1.81.2.1 yamt submitblocked = rf_SubmitReconBuffer(rbuf, 0, 0);
879 1.81.2.1 yamt Dprintf1("RECON: submitblocked=%d\n", submitblocked);
880 1.81.2.1 yamt if (!submitblocked)
881 1.81.2.1 yamt retcode = IssueNextReadRequest(raidPtr, event->col);
882 1.81.2.1 yamt }
883 1.4 oster break;
884 1.4 oster
885 1.4 oster /* a write I/O has completed */
886 1.4 oster case RF_REVENT_WRITEDONE:
887 1.40 oster #if RF_DEBUG_RECON
888 1.4 oster if (rf_floatingRbufDebug) {
889 1.4 oster rf_CheckFloatingRbufCount(raidPtr, 1);
890 1.4 oster }
891 1.38 oster #endif
892 1.4 oster sectorsPerRU = raidPtr->Layout.sectorsPerStripeUnit * raidPtr->Layout.SUsPerRU;
893 1.4 oster rbuf = (RF_ReconBuffer_t *) event->arg;
894 1.4 oster rf_FreeDiskQueueData((RF_DiskQueueData_t *) rbuf->arg);
895 1.4 oster Dprintf3("RECON: WRITEDONE EVENT: psid %d ru %d (%d %% complete)\n",
896 1.57 oster rbuf->parityStripeID, rbuf->which_ru, raidPtr->reconControl->percentComplete);
897 1.57 oster rf_ReconMapUpdate(raidPtr, raidPtr->reconControl->reconMap,
898 1.4 oster rbuf->failedDiskSectorOffset, rbuf->failedDiskSectorOffset + sectorsPerRU - 1);
899 1.57 oster rf_RemoveFromActiveReconTable(raidPtr, rbuf->parityStripeID, rbuf->which_ru);
900 1.4 oster
901 1.81.2.1 yamt RF_LOCK_MUTEX(raidPtr->reconControl->rb_mutex);
902 1.81.2.1 yamt raidPtr->reconControl->pending_writes--;
903 1.81.2.1 yamt RF_UNLOCK_MUTEX(raidPtr->reconControl->rb_mutex);
904 1.81.2.1 yamt
905 1.4 oster if (rbuf->type == RF_RBUF_TYPE_FLOATING) {
906 1.57 oster RF_LOCK_MUTEX(raidPtr->reconControl->rb_mutex);
907 1.76 oster while(raidPtr->reconControl->rb_lock) {
908 1.76 oster ltsleep(&raidPtr->reconControl->rb_lock, PRIBIO, "reconctrlpre1", 0,
909 1.76 oster &raidPtr->reconControl->rb_mutex);
910 1.76 oster }
911 1.76 oster raidPtr->reconControl->rb_lock = 1;
912 1.76 oster RF_UNLOCK_MUTEX(raidPtr->reconControl->rb_mutex);
913 1.76 oster
914 1.4 oster raidPtr->numFullReconBuffers--;
915 1.57 oster rf_ReleaseFloatingReconBuffer(raidPtr, rbuf);
916 1.76 oster
917 1.76 oster RF_LOCK_MUTEX(raidPtr->reconControl->rb_mutex);
918 1.76 oster raidPtr->reconControl->rb_lock = 0;
919 1.76 oster wakeup(&raidPtr->reconControl->rb_lock);
920 1.57 oster RF_UNLOCK_MUTEX(raidPtr->reconControl->rb_mutex);
921 1.4 oster } else
922 1.4 oster if (rbuf->type == RF_RBUF_TYPE_FORCED)
923 1.4 oster rf_FreeReconBuffer(rbuf);
924 1.4 oster else
925 1.4 oster RF_ASSERT(0);
926 1.81.2.1 yamt retcode = 0;
927 1.4 oster break;
928 1.4 oster
929 1.4 oster case RF_REVENT_BUFCLEAR: /* A buffer-stall condition has been
930 1.4 oster * cleared */
931 1.57 oster Dprintf1("RECON: BUFCLEAR EVENT: col %d\n", event->col);
932 1.81.2.1 yamt if (!raidPtr->reconControl->error) {
933 1.81.2.1 yamt submitblocked = rf_SubmitReconBuffer(raidPtr->reconControl->perDiskInfo[event->col].rbuf,
934 1.81.2.1 yamt 0, (int) (long) event->arg);
935 1.81.2.1 yamt RF_ASSERT(!submitblocked); /* we wouldn't have gotten the
936 1.81.2.1 yamt * BUFCLEAR event if we
937 1.81.2.1 yamt * couldn't submit */
938 1.81.2.1 yamt retcode = IssueNextReadRequest(raidPtr, event->col);
939 1.81.2.1 yamt }
940 1.4 oster break;
941 1.4 oster
942 1.4 oster case RF_REVENT_BLOCKCLEAR: /* A user-write reconstruction
943 1.4 oster * blockage has been cleared */
944 1.57 oster DDprintf1("RECON: BLOCKCLEAR EVENT: col %d\n", event->col);
945 1.81.2.1 yamt if (!raidPtr->reconControl->error) {
946 1.81.2.1 yamt retcode = TryToRead(raidPtr, event->col);
947 1.81.2.1 yamt }
948 1.4 oster break;
949 1.4 oster
950 1.4 oster case RF_REVENT_HEADSEPCLEAR: /* A max-head-separation
951 1.4 oster * reconstruction blockage has been
952 1.4 oster * cleared */
953 1.57 oster Dprintf1("RECON: HEADSEPCLEAR EVENT: col %d\n", event->col);
954 1.81.2.1 yamt if (!raidPtr->reconControl->error) {
955 1.81.2.1 yamt retcode = TryToRead(raidPtr, event->col);
956 1.81.2.1 yamt }
957 1.4 oster break;
958 1.4 oster
959 1.4 oster /* a buffer has become ready to write */
960 1.4 oster case RF_REVENT_BUFREADY:
961 1.57 oster Dprintf1("RECON: BUFREADY EVENT: col %d\n", event->col);
962 1.81.2.1 yamt if (!raidPtr->reconControl->error) {
963 1.81.2.1 yamt retcode = IssueNextWriteRequest(raidPtr);
964 1.40 oster #if RF_DEBUG_RECON
965 1.81.2.1 yamt if (rf_floatingRbufDebug) {
966 1.81.2.1 yamt rf_CheckFloatingRbufCount(raidPtr, 1);
967 1.81.2.1 yamt }
968 1.38 oster #endif
969 1.81.2.1 yamt }
970 1.4 oster break;
971 1.4 oster
972 1.4 oster /* we need to skip the current RU entirely because it got
973 1.4 oster * recon'd while we were waiting for something else to happen */
974 1.4 oster case RF_REVENT_SKIP:
975 1.57 oster DDprintf1("RECON: SKIP EVENT: col %d\n", event->col);
976 1.81.2.1 yamt if (!raidPtr->reconControl->error) {
977 1.81.2.1 yamt retcode = IssueNextReadRequest(raidPtr, event->col);
978 1.81.2.1 yamt }
979 1.4 oster break;
980 1.4 oster
981 1.4 oster /* a forced-reconstruction read access has completed. Just
982 1.4 oster * submit the buffer */
983 1.4 oster case RF_REVENT_FORCEDREADDONE:
984 1.4 oster rbuf = (RF_ReconBuffer_t *) event->arg;
985 1.4 oster rf_FreeDiskQueueData((RF_DiskQueueData_t *) rbuf->arg);
986 1.57 oster DDprintf1("RECON: FORCEDREADDONE EVENT: col %d\n", event->col);
987 1.81.2.1 yamt if (!raidPtr->reconControl->error) {
988 1.81.2.1 yamt submitblocked = rf_SubmitReconBuffer(rbuf, 1, 0);
989 1.81.2.1 yamt RF_ASSERT(!submitblocked);
990 1.81.2.1 yamt }
991 1.4 oster break;
992 1.4 oster
993 1.70 oster /* A read I/O failed to complete */
994 1.70 oster case RF_REVENT_READ_FAILED:
995 1.81.2.1 yamt retcode = RF_RECON_READ_ERROR;
996 1.81.2.1 yamt break;
997 1.70 oster
998 1.70 oster /* A write I/O failed to complete */
999 1.70 oster case RF_REVENT_WRITE_FAILED:
1000 1.81.2.1 yamt retcode = RF_RECON_WRITE_ERROR;
1001 1.81.2.1 yamt
1002 1.81.2.1 yamt rbuf = (RF_ReconBuffer_t *) event->arg;
1003 1.81.2.1 yamt
1004 1.81.2.1 yamt /* cleanup the disk queue data */
1005 1.81.2.1 yamt rf_FreeDiskQueueData((RF_DiskQueueData_t *) rbuf->arg);
1006 1.81.2.1 yamt
1007 1.81.2.1 yamt /* At this point we're erroring out, badly, and floatingRbufs
1008 1.81.2.1 yamt may not even be valid. Rather than putting this back onto
1009 1.81.2.1 yamt the floatingRbufs list, just arrange for its immediate
1010 1.81.2.1 yamt destruction.
1011 1.81.2.1 yamt */
1012 1.81.2.1 yamt rf_FreeReconBuffer(rbuf);
1013 1.81.2.1 yamt break;
1014 1.70 oster
1015 1.70 oster /* a forced read I/O failed to complete */
1016 1.70 oster case RF_REVENT_FORCEDREAD_FAILED:
1017 1.81.2.1 yamt retcode = RF_RECON_READ_ERROR;
1018 1.81.2.1 yamt break;
1019 1.70 oster
1020 1.4 oster default:
1021 1.4 oster RF_PANIC();
1022 1.4 oster }
1023 1.4 oster rf_FreeReconEventDesc(event);
1024 1.4 oster return (retcode);
1025 1.1 oster }
1026 1.13 oster /*****************************************************************************
1027 1.1 oster *
1028 1.13 oster * find the next thing that's needed on the indicated disk, and issue
1029 1.13 oster * a read request for it. We assume that the reconstruction buffer
1030 1.13 oster * associated with this process is free to receive the data. If
1031 1.13 oster * reconstruction is blocked on the indicated RU, we issue a
1032 1.13 oster * blockage-release request instead of a physical disk read request.
1033 1.13 oster * If the current disk gets too far ahead of the others, we issue a
1034 1.13 oster * head-separation wait request and return.
1035 1.13 oster *
1036 1.13 oster * ctrl->{ru_count, curPSID, diskOffset} and
1037 1.22 soren * rbuf->failedDiskSectorOffset are maintained to point to the unit
1038 1.13 oster * we're currently accessing. Note that this deviates from the
1039 1.13 oster * standard C idiom of having counters point to the next thing to be
1040 1.13 oster * accessed. This allows us to easily retry when we're blocked by
1041 1.13 oster * head separation or reconstruction-blockage events.
1042 1.1 oster *
1043 1.13 oster *****************************************************************************/
1044 1.4 oster static int
1045 1.60 oster IssueNextReadRequest(RF_Raid_t *raidPtr, RF_RowCol_t col)
1046 1.4 oster {
1047 1.57 oster RF_PerDiskReconCtrl_t *ctrl = &raidPtr->reconControl->perDiskInfo[col];
1048 1.4 oster RF_RaidLayout_t *layoutPtr = &raidPtr->Layout;
1049 1.4 oster RF_ReconBuffer_t *rbuf = ctrl->rbuf;
1050 1.4 oster RF_ReconUnitCount_t RUsPerPU = layoutPtr->SUsPerPU / layoutPtr->SUsPerRU;
1051 1.4 oster RF_SectorCount_t sectorsPerRU = layoutPtr->sectorsPerStripeUnit * layoutPtr->SUsPerRU;
1052 1.4 oster int do_new_check = 0, retcode = 0, status;
1053 1.4 oster
1054 1.4 oster /* if we are currently the slowest disk, mark that we have to do a new
1055 1.4 oster * check */
1056 1.57 oster if (ctrl->headSepCounter <= raidPtr->reconControl->minHeadSepCounter)
1057 1.4 oster do_new_check = 1;
1058 1.4 oster
1059 1.4 oster while (1) {
1060 1.4 oster
1061 1.4 oster ctrl->ru_count++;
1062 1.4 oster if (ctrl->ru_count < RUsPerPU) {
1063 1.4 oster ctrl->diskOffset += sectorsPerRU;
1064 1.4 oster rbuf->failedDiskSectorOffset += sectorsPerRU;
1065 1.4 oster } else {
1066 1.4 oster ctrl->curPSID++;
1067 1.4 oster ctrl->ru_count = 0;
1068 1.4 oster /* code left over from when head-sep was based on
1069 1.4 oster * parity stripe id */
1070 1.57 oster if (ctrl->curPSID >= raidPtr->reconControl->lastPSID) {
1071 1.57 oster CheckForNewMinHeadSep(raidPtr, ++(ctrl->headSepCounter));
1072 1.81.2.1 yamt return (RF_RECON_DONE_READS); /* finito! */
1073 1.4 oster }
1074 1.4 oster /* find the disk offsets of the start of the parity
1075 1.4 oster * stripe on both the current disk and the failed
1076 1.4 oster * disk. skip this entire parity stripe if either disk
1077 1.4 oster * does not appear in the indicated PS */
1078 1.57 oster status = ComputePSDiskOffsets(raidPtr, ctrl->curPSID, col, &ctrl->diskOffset, &rbuf->failedDiskSectorOffset,
1079 1.57 oster &rbuf->spCol, &rbuf->spOffset);
1080 1.4 oster if (status) {
1081 1.4 oster ctrl->ru_count = RUsPerPU - 1;
1082 1.4 oster continue;
1083 1.4 oster }
1084 1.4 oster }
1085 1.4 oster rbuf->which_ru = ctrl->ru_count;
1086 1.4 oster
1087 1.4 oster /* skip this RU if it's already been reconstructed */
1088 1.57 oster if (rf_CheckRUReconstructed(raidPtr->reconControl->reconMap, rbuf->failedDiskSectorOffset)) {
1089 1.4 oster Dprintf2("Skipping psid %ld ru %d: already reconstructed\n", ctrl->curPSID, ctrl->ru_count);
1090 1.4 oster continue;
1091 1.4 oster }
1092 1.4 oster break;
1093 1.4 oster }
1094 1.4 oster ctrl->headSepCounter++;
1095 1.4 oster if (do_new_check)
1096 1.57 oster CheckForNewMinHeadSep(raidPtr, ctrl->headSepCounter); /* update min if needed */
1097 1.4 oster
1098 1.4 oster
1099 1.4 oster /* at this point, we have definitely decided what to do, and we have
1100 1.4 oster * only to see if we can actually do it now */
1101 1.4 oster rbuf->parityStripeID = ctrl->curPSID;
1102 1.4 oster rbuf->which_ru = ctrl->ru_count;
1103 1.67 oster #if RF_ACC_TRACE > 0
1104 1.29 thorpej memset((char *) &raidPtr->recon_tracerecs[col], 0,
1105 1.29 thorpej sizeof(raidPtr->recon_tracerecs[col]));
1106 1.4 oster raidPtr->recon_tracerecs[col].reconacc = 1;
1107 1.4 oster RF_ETIMER_START(raidPtr->recon_tracerecs[col].recon_timer);
1108 1.67 oster #endif
1109 1.57 oster retcode = TryToRead(raidPtr, col);
1110 1.4 oster return (retcode);
1111 1.1 oster }
1112 1.13 oster
1113 1.13 oster /*
1114 1.13 oster * tries to issue the next read on the indicated disk. We may be
1115 1.13 oster * blocked by (a) the heads being too far apart, or (b) recon on the
1116 1.13 oster * indicated RU being blocked due to a write by a user thread. In
1117 1.13 oster * this case, we issue a head-sep or blockage wait request, which will
1118 1.13 oster * cause this same routine to be invoked again later when the blockage
1119 1.13 oster * has cleared.
1120 1.1 oster */
1121 1.13 oster
1122 1.4 oster static int
1123 1.60 oster TryToRead(RF_Raid_t *raidPtr, RF_RowCol_t col)
1124 1.4 oster {
1125 1.57 oster RF_PerDiskReconCtrl_t *ctrl = &raidPtr->reconControl->perDiskInfo[col];
1126 1.4 oster RF_SectorCount_t sectorsPerRU = raidPtr->Layout.sectorsPerStripeUnit * raidPtr->Layout.SUsPerRU;
1127 1.4 oster RF_StripeNum_t psid = ctrl->curPSID;
1128 1.4 oster RF_ReconUnitNum_t which_ru = ctrl->ru_count;
1129 1.4 oster RF_DiskQueueData_t *req;
1130 1.68 oster int status;
1131 1.68 oster RF_ReconParityStripeStatus_t *pssPtr, *newpssPtr;
1132 1.4 oster
1133 1.4 oster /* if the current disk is too far ahead of the others, issue a
1134 1.4 oster * head-separation wait and return */
1135 1.57 oster if (CheckHeadSeparation(raidPtr, ctrl, col, ctrl->headSepCounter, which_ru))
1136 1.4 oster return (0);
1137 1.68 oster
1138 1.68 oster /* allocate a new PSS in case we need it */
1139 1.68 oster newpssPtr = rf_AllocPSStatus(raidPtr);
1140 1.68 oster
1141 1.57 oster RF_LOCK_PSS_MUTEX(raidPtr, psid);
1142 1.68 oster pssPtr = rf_LookupRUStatus(raidPtr, raidPtr->reconControl->pssTable, psid, which_ru, RF_PSS_CREATE, newpssPtr);
1143 1.68 oster
1144 1.68 oster if (pssPtr != newpssPtr) {
1145 1.68 oster rf_FreePSStatus(raidPtr, newpssPtr);
1146 1.68 oster }
1147 1.4 oster
1148 1.4 oster /* if recon is blocked on the indicated parity stripe, issue a
1149 1.4 oster * block-wait request and return. this also must mark the indicated RU
1150 1.4 oster * in the stripe as under reconstruction if not blocked. */
1151 1.57 oster status = CheckForcedOrBlockedReconstruction(raidPtr, pssPtr, ctrl, col, psid, which_ru);
1152 1.4 oster if (status == RF_PSS_RECON_BLOCKED) {
1153 1.4 oster Dprintf2("RECON: Stalling psid %ld ru %d: recon blocked\n", psid, which_ru);
1154 1.4 oster goto out;
1155 1.4 oster } else
1156 1.4 oster if (status == RF_PSS_FORCED_ON_WRITE) {
1157 1.57 oster rf_CauseReconEvent(raidPtr, col, NULL, RF_REVENT_SKIP);
1158 1.4 oster goto out;
1159 1.4 oster }
1160 1.4 oster /* make one last check to be sure that the indicated RU didn't get
1161 1.4 oster * reconstructed while we were waiting for something else to happen.
1162 1.4 oster * This is unfortunate in that it causes us to make this check twice
1163 1.4 oster * in the normal case. Might want to make some attempt to re-work
1164 1.4 oster * this so that we only do this check if we've definitely blocked on
1165 1.4 oster * one of the above checks. When this condition is detected, we may
1166 1.4 oster * have just created a bogus status entry, which we need to delete. */
1167 1.57 oster if (rf_CheckRUReconstructed(raidPtr->reconControl->reconMap, ctrl->rbuf->failedDiskSectorOffset)) {
1168 1.4 oster Dprintf2("RECON: Skipping psid %ld ru %d: prior recon after stall\n", psid, which_ru);
1169 1.68 oster if (pssPtr == newpssPtr)
1170 1.57 oster rf_PSStatusDelete(raidPtr, raidPtr->reconControl->pssTable, pssPtr);
1171 1.57 oster rf_CauseReconEvent(raidPtr, col, NULL, RF_REVENT_SKIP);
1172 1.4 oster goto out;
1173 1.4 oster }
1174 1.4 oster /* found something to read. issue the I/O */
1175 1.57 oster Dprintf4("RECON: Read for psid %ld on col %d offset %ld buf %lx\n",
1176 1.57 oster psid, col, ctrl->diskOffset, ctrl->rbuf->buffer);
1177 1.67 oster #if RF_ACC_TRACE > 0
1178 1.4 oster RF_ETIMER_STOP(raidPtr->recon_tracerecs[col].recon_timer);
1179 1.4 oster RF_ETIMER_EVAL(raidPtr->recon_tracerecs[col].recon_timer);
1180 1.4 oster raidPtr->recon_tracerecs[col].specific.recon.recon_start_to_fetch_us =
1181 1.4 oster RF_ETIMER_VAL_US(raidPtr->recon_tracerecs[col].recon_timer);
1182 1.4 oster RF_ETIMER_START(raidPtr->recon_tracerecs[col].recon_timer);
1183 1.67 oster #endif
1184 1.4 oster /* should be ok to use a NULL proc pointer here, all the bufs we use
1185 1.4 oster * should be in kernel space */
1186 1.4 oster req = rf_CreateDiskQueueData(RF_IO_TYPE_READ, ctrl->diskOffset, sectorsPerRU, ctrl->rbuf->buffer, psid, which_ru,
1187 1.81.2.1 yamt ReconReadDoneProc, (void *) ctrl,
1188 1.67 oster #if RF_ACC_TRACE > 0
1189 1.67 oster &raidPtr->recon_tracerecs[col],
1190 1.67 oster #else
1191 1.67 oster NULL,
1192 1.67 oster #endif
1193 1.81.2.1 yamt (void *) raidPtr, 0, NULL, PR_WAITOK);
1194 1.4 oster
1195 1.4 oster ctrl->rbuf->arg = (void *) req;
1196 1.57 oster rf_DiskIOEnqueue(&raidPtr->Queues[col], req, RF_IO_RECON_PRIORITY);
1197 1.4 oster pssPtr->issued[col] = 1;
1198 1.1 oster
1199 1.1 oster out:
1200 1.57 oster RF_UNLOCK_PSS_MUTEX(raidPtr, psid);
1201 1.4 oster return (0);
1202 1.1 oster }
1203 1.1 oster
1204 1.1 oster
1205 1.13 oster /*
1206 1.13 oster * given a parity stripe ID, we want to find out whether both the
1207 1.13 oster * current disk and the failed disk exist in that parity stripe. If
1208 1.13 oster * not, we want to skip this whole PS. If so, we want to find the
1209 1.13 oster * disk offset of the start of the PS on both the current disk and the
1210 1.13 oster * failed disk.
1211 1.13 oster *
1212 1.13 oster * this works by getting a list of disks comprising the indicated
1213 1.13 oster * parity stripe, and searching the list for the current and failed
1214 1.13 oster * disks. Once we've decided they both exist in the parity stripe, we
1215 1.13 oster * need to decide whether each is data or parity, so that we'll know
1216 1.13 oster * which mapping function to call to get the corresponding disk
1217 1.1 oster * offsets.
1218 1.1 oster *
1219 1.13 oster * this is kind of unpleasant, but doing it this way allows the
1220 1.13 oster * reconstruction code to use parity stripe IDs rather than physical
1221 1.13 oster * disks address to march through the failed disk, which greatly
1222 1.13 oster * simplifies a lot of code, as well as eliminating the need for a
1223 1.13 oster * reverse-mapping function. I also think it will execute faster,
1224 1.13 oster * since the calls to the mapping module are kept to a minimum.
1225 1.1 oster *
1226 1.13 oster * ASSUMES THAT THE STRIPE IDENTIFIER IDENTIFIES THE DISKS COMPRISING
1227 1.60 oster * THE STRIPE IN THE CORRECT ORDER
1228 1.60 oster *
1229 1.60 oster * raidPtr - raid descriptor
1230 1.60 oster * psid - parity stripe identifier
1231 1.60 oster * col - column of disk to find the offsets for
1232 1.60 oster * spCol - out: col of spare unit for failed unit
1233 1.60 oster * spOffset - out: offset into disk containing spare unit
1234 1.60 oster *
1235 1.60 oster */
1236 1.13 oster
1237 1.13 oster
1238 1.4 oster static int
1239 1.60 oster ComputePSDiskOffsets(RF_Raid_t *raidPtr, RF_StripeNum_t psid,
1240 1.60 oster RF_RowCol_t col, RF_SectorNum_t *outDiskOffset,
1241 1.60 oster RF_SectorNum_t *outFailedDiskSectorOffset,
1242 1.60 oster RF_RowCol_t *spCol, RF_SectorNum_t *spOffset)
1243 1.60 oster {
1244 1.4 oster RF_RaidLayout_t *layoutPtr = &raidPtr->Layout;
1245 1.57 oster RF_RowCol_t fcol = raidPtr->reconControl->fcol;
1246 1.4 oster RF_RaidAddr_t sosRaidAddress; /* start-of-stripe */
1247 1.4 oster RF_RowCol_t *diskids;
1248 1.4 oster u_int i, j, k, i_offset, j_offset;
1249 1.57 oster RF_RowCol_t pcol;
1250 1.57 oster int testcol;
1251 1.4 oster RF_SectorNum_t poffset;
1252 1.4 oster char i_is_parity = 0, j_is_parity = 0;
1253 1.4 oster RF_RowCol_t stripeWidth = layoutPtr->numDataCol + layoutPtr->numParityCol;
1254 1.4 oster
1255 1.4 oster /* get a listing of the disks comprising that stripe */
1256 1.4 oster sosRaidAddress = rf_ParityStripeIDToRaidAddress(layoutPtr, psid);
1257 1.57 oster (layoutPtr->map->IdentifyStripe) (raidPtr, sosRaidAddress, &diskids);
1258 1.4 oster RF_ASSERT(diskids);
1259 1.4 oster
1260 1.4 oster /* reject this entire parity stripe if it does not contain the
1261 1.4 oster * indicated disk or it does not contain the failed disk */
1262 1.57 oster
1263 1.4 oster for (i = 0; i < stripeWidth; i++) {
1264 1.4 oster if (col == diskids[i])
1265 1.4 oster break;
1266 1.4 oster }
1267 1.4 oster if (i == stripeWidth)
1268 1.4 oster goto skipit;
1269 1.4 oster for (j = 0; j < stripeWidth; j++) {
1270 1.4 oster if (fcol == diskids[j])
1271 1.4 oster break;
1272 1.4 oster }
1273 1.4 oster if (j == stripeWidth) {
1274 1.4 oster goto skipit;
1275 1.4 oster }
1276 1.4 oster /* find out which disk the parity is on */
1277 1.57 oster (layoutPtr->map->MapParity) (raidPtr, sosRaidAddress, &pcol, &poffset, RF_DONT_REMAP);
1278 1.4 oster
1279 1.4 oster /* find out if either the current RU or the failed RU is parity */
1280 1.4 oster /* also, if the parity occurs in this stripe prior to the data and/or
1281 1.4 oster * failed col, we need to decrement i and/or j */
1282 1.4 oster for (k = 0; k < stripeWidth; k++)
1283 1.4 oster if (diskids[k] == pcol)
1284 1.4 oster break;
1285 1.4 oster RF_ASSERT(k < stripeWidth);
1286 1.4 oster i_offset = i;
1287 1.4 oster j_offset = j;
1288 1.4 oster if (k < i)
1289 1.4 oster i_offset--;
1290 1.4 oster else
1291 1.4 oster if (k == i) {
1292 1.4 oster i_is_parity = 1;
1293 1.4 oster i_offset = 0;
1294 1.4 oster } /* set offsets to zero to disable multiply
1295 1.4 oster * below */
1296 1.4 oster if (k < j)
1297 1.4 oster j_offset--;
1298 1.4 oster else
1299 1.4 oster if (k == j) {
1300 1.4 oster j_is_parity = 1;
1301 1.4 oster j_offset = 0;
1302 1.4 oster }
1303 1.4 oster /* at this point, [ij]_is_parity tells us whether the [current,failed]
1304 1.4 oster * disk is parity at the start of this RU, and, if data, "[ij]_offset"
1305 1.4 oster * tells us how far into the stripe the [current,failed] disk is. */
1306 1.4 oster
1307 1.4 oster /* call the mapping routine to get the offset into the current disk,
1308 1.4 oster * repeat for failed disk. */
1309 1.4 oster if (i_is_parity)
1310 1.57 oster layoutPtr->map->MapParity(raidPtr, sosRaidAddress + i_offset * layoutPtr->sectorsPerStripeUnit, &testcol, outDiskOffset, RF_DONT_REMAP);
1311 1.4 oster else
1312 1.57 oster layoutPtr->map->MapSector(raidPtr, sosRaidAddress + i_offset * layoutPtr->sectorsPerStripeUnit, &testcol, outDiskOffset, RF_DONT_REMAP);
1313 1.4 oster
1314 1.57 oster RF_ASSERT(col == testcol);
1315 1.4 oster
1316 1.4 oster if (j_is_parity)
1317 1.57 oster layoutPtr->map->MapParity(raidPtr, sosRaidAddress + j_offset * layoutPtr->sectorsPerStripeUnit, &testcol, outFailedDiskSectorOffset, RF_DONT_REMAP);
1318 1.4 oster else
1319 1.57 oster layoutPtr->map->MapSector(raidPtr, sosRaidAddress + j_offset * layoutPtr->sectorsPerStripeUnit, &testcol, outFailedDiskSectorOffset, RF_DONT_REMAP);
1320 1.57 oster RF_ASSERT(fcol == testcol);
1321 1.4 oster
1322 1.4 oster /* now locate the spare unit for the failed unit */
1323 1.72 oster #if RF_INCLUDE_PARITY_DECLUSTERING_DS > 0
1324 1.4 oster if (layoutPtr->map->flags & RF_DISTRIBUTE_SPARE) {
1325 1.4 oster if (j_is_parity)
1326 1.57 oster layoutPtr->map->MapParity(raidPtr, sosRaidAddress + j_offset * layoutPtr->sectorsPerStripeUnit, spCol, spOffset, RF_REMAP);
1327 1.4 oster else
1328 1.57 oster layoutPtr->map->MapSector(raidPtr, sosRaidAddress + j_offset * layoutPtr->sectorsPerStripeUnit, spCol, spOffset, RF_REMAP);
1329 1.4 oster } else {
1330 1.72 oster #endif
1331 1.57 oster *spCol = raidPtr->reconControl->spareCol;
1332 1.4 oster *spOffset = *outFailedDiskSectorOffset;
1333 1.72 oster #if RF_INCLUDE_PARITY_DECLUSTERING_DS > 0
1334 1.4 oster }
1335 1.72 oster #endif
1336 1.4 oster return (0);
1337 1.1 oster
1338 1.1 oster skipit:
1339 1.57 oster Dprintf2("RECON: Skipping psid %ld: nothing needed from r%d c%d\n",
1340 1.57 oster psid, col);
1341 1.4 oster return (1);
1342 1.1 oster }
1343 1.4 oster /* this is called when a buffer has become ready to write to the replacement disk */
1344 1.4 oster static int
1345 1.60 oster IssueNextWriteRequest(RF_Raid_t *raidPtr)
1346 1.4 oster {
1347 1.4 oster RF_RaidLayout_t *layoutPtr = &raidPtr->Layout;
1348 1.4 oster RF_SectorCount_t sectorsPerRU = layoutPtr->sectorsPerStripeUnit * layoutPtr->SUsPerRU;
1349 1.67 oster #if RF_ACC_TRACE > 0
1350 1.57 oster RF_RowCol_t fcol = raidPtr->reconControl->fcol;
1351 1.67 oster #endif
1352 1.4 oster RF_ReconBuffer_t *rbuf;
1353 1.4 oster RF_DiskQueueData_t *req;
1354 1.4 oster
1355 1.57 oster rbuf = rf_GetFullReconBuffer(raidPtr->reconControl);
1356 1.4 oster RF_ASSERT(rbuf); /* there must be one available, or we wouldn't
1357 1.4 oster * have gotten the event that sent us here */
1358 1.4 oster RF_ASSERT(rbuf->pssPtr);
1359 1.4 oster
1360 1.4 oster rbuf->pssPtr->writeRbuf = rbuf;
1361 1.4 oster rbuf->pssPtr = NULL;
1362 1.4 oster
1363 1.57 oster Dprintf6("RECON: New write (c %d offs %d) for psid %ld ru %d (failed disk offset %ld) buf %lx\n",
1364 1.57 oster rbuf->spCol, rbuf->spOffset, rbuf->parityStripeID,
1365 1.4 oster rbuf->which_ru, rbuf->failedDiskSectorOffset, rbuf->buffer);
1366 1.4 oster Dprintf6("RECON: new write psid %ld %02x %02x %02x %02x %02x\n",
1367 1.4 oster rbuf->parityStripeID, rbuf->buffer[0] & 0xff, rbuf->buffer[1] & 0xff,
1368 1.4 oster rbuf->buffer[2] & 0xff, rbuf->buffer[3] & 0xff, rbuf->buffer[4] & 0xff);
1369 1.4 oster
1370 1.4 oster /* should be ok to use a NULL b_proc here b/c all addrs should be in
1371 1.4 oster * kernel space */
1372 1.4 oster req = rf_CreateDiskQueueData(RF_IO_TYPE_WRITE, rbuf->spOffset,
1373 1.4 oster sectorsPerRU, rbuf->buffer,
1374 1.4 oster rbuf->parityStripeID, rbuf->which_ru,
1375 1.81.2.1 yamt ReconWriteDoneProc, (void *) rbuf,
1376 1.67 oster #if RF_ACC_TRACE > 0
1377 1.4 oster &raidPtr->recon_tracerecs[fcol],
1378 1.67 oster #else
1379 1.67 oster NULL,
1380 1.67 oster #endif
1381 1.81.2.1 yamt (void *) raidPtr, 0, NULL, PR_WAITOK);
1382 1.1 oster
1383 1.4 oster rbuf->arg = (void *) req;
1384 1.81.2.1 yamt RF_LOCK_MUTEX(raidPtr->reconControl->rb_mutex);
1385 1.81.2.1 yamt raidPtr->reconControl->pending_writes++;
1386 1.81.2.1 yamt RF_UNLOCK_MUTEX(raidPtr->reconControl->rb_mutex);
1387 1.57 oster rf_DiskIOEnqueue(&raidPtr->Queues[rbuf->spCol], req, RF_IO_RECON_PRIORITY);
1388 1.1 oster
1389 1.4 oster return (0);
1390 1.1 oster }
1391 1.13 oster
1392 1.13 oster /*
1393 1.13 oster * this gets called upon the completion of a reconstruction read
1394 1.13 oster * operation the arg is a pointer to the per-disk reconstruction
1395 1.13 oster * control structure for the process that just finished a read.
1396 1.1 oster *
1397 1.13 oster * called at interrupt context in the kernel, so don't do anything
1398 1.13 oster * illegal here.
1399 1.1 oster */
1400 1.4 oster static int
1401 1.60 oster ReconReadDoneProc(void *arg, int status)
1402 1.4 oster {
1403 1.4 oster RF_PerDiskReconCtrl_t *ctrl = (RF_PerDiskReconCtrl_t *) arg;
1404 1.81.2.1 yamt RF_Raid_t *raidPtr;
1405 1.81.2.1 yamt
1406 1.81.2.1 yamt /* Detect that reconCtrl is no longer valid, and if that
1407 1.81.2.1 yamt is the case, bail without calling rf_CauseReconEvent().
1408 1.81.2.1 yamt There won't be anyone listening for this event anyway */
1409 1.81.2.1 yamt
1410 1.81.2.1 yamt if (ctrl->reconCtrl == NULL)
1411 1.81.2.1 yamt return(0);
1412 1.81.2.1 yamt
1413 1.81.2.1 yamt raidPtr = ctrl->reconCtrl->reconDesc->raidPtr;
1414 1.4 oster
1415 1.4 oster if (status) {
1416 1.70 oster printf("raid%d: Recon read failed!\n", raidPtr->raidid);
1417 1.70 oster rf_CauseReconEvent(raidPtr, ctrl->col, NULL, RF_REVENT_READ_FAILED);
1418 1.70 oster return(0);
1419 1.4 oster }
1420 1.67 oster #if RF_ACC_TRACE > 0
1421 1.4 oster RF_ETIMER_STOP(raidPtr->recon_tracerecs[ctrl->col].recon_timer);
1422 1.4 oster RF_ETIMER_EVAL(raidPtr->recon_tracerecs[ctrl->col].recon_timer);
1423 1.4 oster raidPtr->recon_tracerecs[ctrl->col].specific.recon.recon_fetch_to_return_us =
1424 1.4 oster RF_ETIMER_VAL_US(raidPtr->recon_tracerecs[ctrl->col].recon_timer);
1425 1.4 oster RF_ETIMER_START(raidPtr->recon_tracerecs[ctrl->col].recon_timer);
1426 1.67 oster #endif
1427 1.57 oster rf_CauseReconEvent(raidPtr, ctrl->col, NULL, RF_REVENT_READDONE);
1428 1.4 oster return (0);
1429 1.1 oster }
1430 1.1 oster /* this gets called upon the completion of a reconstruction write operation.
1431 1.1 oster * the arg is a pointer to the rbuf that was just written
1432 1.1 oster *
1433 1.1 oster * called at interrupt context in the kernel, so don't do anything illegal here.
1434 1.1 oster */
1435 1.4 oster static int
1436 1.60 oster ReconWriteDoneProc(void *arg, int status)
1437 1.4 oster {
1438 1.4 oster RF_ReconBuffer_t *rbuf = (RF_ReconBuffer_t *) arg;
1439 1.4 oster
1440 1.81.2.1 yamt /* Detect that reconControl is no longer valid, and if that
1441 1.81.2.1 yamt is the case, bail without calling rf_CauseReconEvent().
1442 1.81.2.1 yamt There won't be anyone listening for this event anyway */
1443 1.81.2.1 yamt
1444 1.81.2.1 yamt if (rbuf->raidPtr->reconControl == NULL)
1445 1.81.2.1 yamt return(0);
1446 1.81.2.1 yamt
1447 1.4 oster Dprintf2("Reconstruction completed on psid %ld ru %d\n", rbuf->parityStripeID, rbuf->which_ru);
1448 1.4 oster if (status) {
1449 1.70 oster printf("raid%d: Recon write failed!\n", rbuf->raidPtr->raidid);
1450 1.71 oster rf_CauseReconEvent(rbuf->raidPtr, rbuf->col, arg, RF_REVENT_WRITE_FAILED);
1451 1.70 oster return(0);
1452 1.4 oster }
1453 1.71 oster rf_CauseReconEvent(rbuf->raidPtr, rbuf->col, arg, RF_REVENT_WRITEDONE);
1454 1.4 oster return (0);
1455 1.1 oster }
1456 1.1 oster
1457 1.1 oster
1458 1.13 oster /*
1459 1.13 oster * computes a new minimum head sep, and wakes up anyone who needs to
1460 1.13 oster * be woken as a result
1461 1.13 oster */
1462 1.4 oster static void
1463 1.60 oster CheckForNewMinHeadSep(RF_Raid_t *raidPtr, RF_HeadSepLimit_t hsCtr)
1464 1.4 oster {
1465 1.57 oster RF_ReconCtrl_t *reconCtrlPtr = raidPtr->reconControl;
1466 1.4 oster RF_HeadSepLimit_t new_min;
1467 1.4 oster RF_RowCol_t i;
1468 1.4 oster RF_CallbackDesc_t *p;
1469 1.4 oster RF_ASSERT(hsCtr >= reconCtrlPtr->minHeadSepCounter); /* from the definition
1470 1.4 oster * of a minimum */
1471 1.4 oster
1472 1.4 oster
1473 1.4 oster RF_LOCK_MUTEX(reconCtrlPtr->rb_mutex);
1474 1.76 oster while(reconCtrlPtr->rb_lock) {
1475 1.76 oster ltsleep(&reconCtrlPtr->rb_lock, PRIBIO, "reconctlcnmhs", 0, &reconCtrlPtr->rb_mutex);
1476 1.76 oster }
1477 1.76 oster reconCtrlPtr->rb_lock = 1;
1478 1.76 oster RF_UNLOCK_MUTEX(reconCtrlPtr->rb_mutex);
1479 1.4 oster
1480 1.4 oster new_min = ~(1L << (8 * sizeof(long) - 1)); /* 0x7FFF....FFF */
1481 1.4 oster for (i = 0; i < raidPtr->numCol; i++)
1482 1.4 oster if (i != reconCtrlPtr->fcol) {
1483 1.4 oster if (reconCtrlPtr->perDiskInfo[i].headSepCounter < new_min)
1484 1.4 oster new_min = reconCtrlPtr->perDiskInfo[i].headSepCounter;
1485 1.4 oster }
1486 1.4 oster /* set the new minimum and wake up anyone who can now run again */
1487 1.4 oster if (new_min != reconCtrlPtr->minHeadSepCounter) {
1488 1.4 oster reconCtrlPtr->minHeadSepCounter = new_min;
1489 1.4 oster Dprintf1("RECON: new min head pos counter val is %ld\n", new_min);
1490 1.4 oster while (reconCtrlPtr->headSepCBList) {
1491 1.4 oster if (reconCtrlPtr->headSepCBList->callbackArg.v > new_min)
1492 1.4 oster break;
1493 1.4 oster p = reconCtrlPtr->headSepCBList;
1494 1.4 oster reconCtrlPtr->headSepCBList = p->next;
1495 1.4 oster p->next = NULL;
1496 1.57 oster rf_CauseReconEvent(raidPtr, p->col, NULL, RF_REVENT_HEADSEPCLEAR);
1497 1.4 oster rf_FreeCallbackDesc(p);
1498 1.4 oster }
1499 1.1 oster
1500 1.4 oster }
1501 1.76 oster RF_LOCK_MUTEX(reconCtrlPtr->rb_mutex);
1502 1.76 oster reconCtrlPtr->rb_lock = 0;
1503 1.76 oster wakeup(&reconCtrlPtr->rb_lock);
1504 1.4 oster RF_UNLOCK_MUTEX(reconCtrlPtr->rb_mutex);
1505 1.1 oster }
1506 1.13 oster
1507 1.13 oster /*
1508 1.13 oster * checks to see that the maximum head separation will not be violated
1509 1.13 oster * if we initiate a reconstruction I/O on the indicated disk.
1510 1.13 oster * Limiting the maximum head separation between two disks eliminates
1511 1.13 oster * the nasty buffer-stall conditions that occur when one disk races
1512 1.13 oster * ahead of the others and consumes all of the floating recon buffers.
1513 1.13 oster * This code is complex and unpleasant but it's necessary to avoid
1514 1.13 oster * some very nasty, albeit fairly rare, reconstruction behavior.
1515 1.1 oster *
1516 1.13 oster * returns non-zero if and only if we have to stop working on the
1517 1.13 oster * indicated disk due to a head-separation delay.
1518 1.1 oster */
1519 1.4 oster static int
1520 1.60 oster CheckHeadSeparation(RF_Raid_t *raidPtr, RF_PerDiskReconCtrl_t *ctrl,
1521 1.60 oster RF_RowCol_t col, RF_HeadSepLimit_t hsCtr,
1522 1.60 oster RF_ReconUnitNum_t which_ru)
1523 1.4 oster {
1524 1.57 oster RF_ReconCtrl_t *reconCtrlPtr = raidPtr->reconControl;
1525 1.4 oster RF_CallbackDesc_t *cb, *p, *pt;
1526 1.10 oster int retval = 0;
1527 1.4 oster
1528 1.4 oster /* if we're too far ahead of the slowest disk, stop working on this
1529 1.4 oster * disk until the slower ones catch up. We do this by scheduling a
1530 1.4 oster * wakeup callback for the time when the slowest disk has caught up.
1531 1.4 oster * We define "caught up" with 20% hysteresis, i.e. the head separation
1532 1.4 oster * must have fallen to at most 80% of the max allowable head
1533 1.4 oster * separation before we'll wake up.
1534 1.4 oster *
1535 1.4 oster */
1536 1.4 oster RF_LOCK_MUTEX(reconCtrlPtr->rb_mutex);
1537 1.76 oster while(reconCtrlPtr->rb_lock) {
1538 1.76 oster ltsleep(&reconCtrlPtr->rb_lock, PRIBIO, "reconctlchs", 0, &reconCtrlPtr->rb_mutex);
1539 1.76 oster }
1540 1.76 oster reconCtrlPtr->rb_lock = 1;
1541 1.76 oster RF_UNLOCK_MUTEX(reconCtrlPtr->rb_mutex);
1542 1.4 oster if ((raidPtr->headSepLimit >= 0) &&
1543 1.4 oster ((ctrl->headSepCounter - reconCtrlPtr->minHeadSepCounter) > raidPtr->headSepLimit)) {
1544 1.57 oster Dprintf5("raid%d: RECON: head sep stall: col %d hsCtr %ld minHSCtr %ld limit %ld\n",
1545 1.57 oster raidPtr->raidid, col, ctrl->headSepCounter,
1546 1.10 oster reconCtrlPtr->minHeadSepCounter,
1547 1.10 oster raidPtr->headSepLimit);
1548 1.4 oster cb = rf_AllocCallbackDesc();
1549 1.4 oster /* the minHeadSepCounter value we have to get to before we'll
1550 1.4 oster * wake up. build in 20% hysteresis. */
1551 1.4 oster cb->callbackArg.v = (ctrl->headSepCounter - raidPtr->headSepLimit + raidPtr->headSepLimit / 5);
1552 1.4 oster cb->col = col;
1553 1.4 oster cb->next = NULL;
1554 1.4 oster
1555 1.4 oster /* insert this callback descriptor into the sorted list of
1556 1.4 oster * pending head-sep callbacks */
1557 1.4 oster p = reconCtrlPtr->headSepCBList;
1558 1.4 oster if (!p)
1559 1.4 oster reconCtrlPtr->headSepCBList = cb;
1560 1.4 oster else
1561 1.4 oster if (cb->callbackArg.v < p->callbackArg.v) {
1562 1.4 oster cb->next = reconCtrlPtr->headSepCBList;
1563 1.4 oster reconCtrlPtr->headSepCBList = cb;
1564 1.4 oster } else {
1565 1.4 oster for (pt = p, p = p->next; p && (p->callbackArg.v < cb->callbackArg.v); pt = p, p = p->next);
1566 1.4 oster cb->next = p;
1567 1.4 oster pt->next = cb;
1568 1.4 oster }
1569 1.4 oster retval = 1;
1570 1.1 oster #if RF_RECON_STATS > 0
1571 1.4 oster ctrl->reconCtrl->reconDesc->hsStallCount++;
1572 1.4 oster #endif /* RF_RECON_STATS > 0 */
1573 1.4 oster }
1574 1.76 oster RF_LOCK_MUTEX(reconCtrlPtr->rb_mutex);
1575 1.76 oster reconCtrlPtr->rb_lock = 0;
1576 1.76 oster wakeup(&reconCtrlPtr->rb_lock);
1577 1.4 oster RF_UNLOCK_MUTEX(reconCtrlPtr->rb_mutex);
1578 1.1 oster
1579 1.4 oster return (retval);
1580 1.1 oster }
1581 1.13 oster /*
1582 1.13 oster * checks to see if reconstruction has been either forced or blocked
1583 1.13 oster * by a user operation. if forced, we skip this RU entirely. else if
1584 1.13 oster * blocked, put ourselves on the wait list. else return 0.
1585 1.1 oster *
1586 1.13 oster * ASSUMES THE PSS MUTEX IS LOCKED UPON ENTRY
1587 1.1 oster */
1588 1.4 oster static int
1589 1.60 oster CheckForcedOrBlockedReconstruction(RF_Raid_t *raidPtr,
1590 1.60 oster RF_ReconParityStripeStatus_t *pssPtr,
1591 1.60 oster RF_PerDiskReconCtrl_t *ctrl,
1592 1.60 oster RF_RowCol_t col, RF_StripeNum_t psid,
1593 1.60 oster RF_ReconUnitNum_t which_ru)
1594 1.4 oster {
1595 1.4 oster RF_CallbackDesc_t *cb;
1596 1.4 oster int retcode = 0;
1597 1.4 oster
1598 1.4 oster if ((pssPtr->flags & RF_PSS_FORCED_ON_READ) || (pssPtr->flags & RF_PSS_FORCED_ON_WRITE))
1599 1.4 oster retcode = RF_PSS_FORCED_ON_WRITE;
1600 1.4 oster else
1601 1.4 oster if (pssPtr->flags & RF_PSS_RECON_BLOCKED) {
1602 1.57 oster Dprintf3("RECON: col %d blocked at psid %ld ru %d\n", col, psid, which_ru);
1603 1.4 oster cb = rf_AllocCallbackDesc(); /* append ourselves to
1604 1.4 oster * the blockage-wait
1605 1.4 oster * list */
1606 1.4 oster cb->col = col;
1607 1.4 oster cb->next = pssPtr->blockWaitList;
1608 1.4 oster pssPtr->blockWaitList = cb;
1609 1.4 oster retcode = RF_PSS_RECON_BLOCKED;
1610 1.4 oster }
1611 1.4 oster if (!retcode)
1612 1.4 oster pssPtr->flags |= RF_PSS_UNDER_RECON; /* mark this RU as under
1613 1.4 oster * reconstruction */
1614 1.4 oster
1615 1.4 oster return (retcode);
1616 1.1 oster }
1617 1.13 oster /*
1618 1.13 oster * if reconstruction is currently ongoing for the indicated stripeID,
1619 1.13 oster * reconstruction is forced to completion and we return non-zero to
1620 1.13 oster * indicate that the caller must wait. If not, then reconstruction is
1621 1.13 oster * blocked on the indicated stripe and the routine returns zero. If
1622 1.13 oster * and only if we return non-zero, we'll cause the cbFunc to get
1623 1.13 oster * invoked with the cbArg when the reconstruction has completed.
1624 1.1 oster */
1625 1.4 oster int
1626 1.60 oster rf_ForceOrBlockRecon(RF_Raid_t *raidPtr, RF_AccessStripeMap_t *asmap,
1627 1.60 oster void (*cbFunc)(RF_Raid_t *, void *), void *cbArg)
1628 1.4 oster {
1629 1.4 oster RF_StripeNum_t stripeID = asmap->stripeID; /* the stripe ID we're
1630 1.4 oster * forcing recon on */
1631 1.4 oster RF_SectorCount_t sectorsPerRU = raidPtr->Layout.sectorsPerStripeUnit * raidPtr->Layout.SUsPerRU; /* num sects in one RU */
1632 1.68 oster RF_ReconParityStripeStatus_t *pssPtr, *newpssPtr; /* a pointer to the parity
1633 1.4 oster * stripe status structure */
1634 1.4 oster RF_StripeNum_t psid; /* parity stripe id */
1635 1.4 oster RF_SectorNum_t offset, fd_offset; /* disk offset, failed-disk
1636 1.4 oster * offset */
1637 1.4 oster RF_RowCol_t *diskids;
1638 1.4 oster RF_ReconUnitNum_t which_ru; /* RU within parity stripe */
1639 1.4 oster RF_RowCol_t fcol, diskno, i;
1640 1.4 oster RF_ReconBuffer_t *new_rbuf; /* ptr to newly allocated rbufs */
1641 1.4 oster RF_DiskQueueData_t *req;/* disk I/O req to be enqueued */
1642 1.4 oster RF_CallbackDesc_t *cb;
1643 1.68 oster int nPromoted;
1644 1.4 oster
1645 1.4 oster psid = rf_MapStripeIDToParityStripeID(&raidPtr->Layout, stripeID, &which_ru);
1646 1.4 oster
1647 1.68 oster /* allocate a new PSS in case we need it */
1648 1.68 oster newpssPtr = rf_AllocPSStatus(raidPtr);
1649 1.68 oster
1650 1.57 oster RF_LOCK_PSS_MUTEX(raidPtr, psid);
1651 1.4 oster
1652 1.68 oster pssPtr = rf_LookupRUStatus(raidPtr, raidPtr->reconControl->pssTable, psid, which_ru, RF_PSS_CREATE | RF_PSS_RECON_BLOCKED, newpssPtr);
1653 1.68 oster
1654 1.68 oster if (pssPtr != newpssPtr) {
1655 1.68 oster rf_FreePSStatus(raidPtr, newpssPtr);
1656 1.68 oster }
1657 1.4 oster
1658 1.4 oster /* if recon is not ongoing on this PS, just return */
1659 1.4 oster if (!(pssPtr->flags & RF_PSS_UNDER_RECON)) {
1660 1.57 oster RF_UNLOCK_PSS_MUTEX(raidPtr, psid);
1661 1.4 oster return (0);
1662 1.4 oster }
1663 1.4 oster /* otherwise, we have to wait for reconstruction to complete on this
1664 1.4 oster * RU. */
1665 1.4 oster /* In order to avoid waiting for a potentially large number of
1666 1.4 oster * low-priority accesses to complete, we force a normal-priority (i.e.
1667 1.4 oster * not low-priority) reconstruction on this RU. */
1668 1.4 oster if (!(pssPtr->flags & RF_PSS_FORCED_ON_WRITE) && !(pssPtr->flags & RF_PSS_FORCED_ON_READ)) {
1669 1.4 oster DDprintf1("Forcing recon on psid %ld\n", psid);
1670 1.4 oster pssPtr->flags |= RF_PSS_FORCED_ON_WRITE; /* mark this RU as under
1671 1.4 oster * forced recon */
1672 1.4 oster pssPtr->flags &= ~RF_PSS_RECON_BLOCKED; /* clear the blockage
1673 1.4 oster * that we just set */
1674 1.57 oster fcol = raidPtr->reconControl->fcol;
1675 1.4 oster
1676 1.4 oster /* get a listing of the disks comprising the indicated stripe */
1677 1.57 oster (raidPtr->Layout.map->IdentifyStripe) (raidPtr, asmap->raidAddress, &diskids);
1678 1.4 oster
1679 1.4 oster /* For previously issued reads, elevate them to normal
1680 1.4 oster * priority. If the I/O has already completed, it won't be
1681 1.4 oster * found in the queue, and hence this will be a no-op. For
1682 1.4 oster * unissued reads, allocate buffers and issue new reads. The
1683 1.4 oster * fact that we've set the FORCED bit means that the regular
1684 1.4 oster * recon procs will not re-issue these reqs */
1685 1.4 oster for (i = 0; i < raidPtr->Layout.numDataCol + raidPtr->Layout.numParityCol; i++)
1686 1.4 oster if ((diskno = diskids[i]) != fcol) {
1687 1.4 oster if (pssPtr->issued[diskno]) {
1688 1.57 oster nPromoted = rf_DiskIOPromote(&raidPtr->Queues[diskno], psid, which_ru);
1689 1.4 oster if (rf_reconDebug && nPromoted)
1690 1.57 oster printf("raid%d: promoted read from col %d\n", raidPtr->raidid, diskno);
1691 1.4 oster } else {
1692 1.57 oster new_rbuf = rf_MakeReconBuffer(raidPtr, diskno, RF_RBUF_TYPE_FORCED); /* create new buf */
1693 1.57 oster ComputePSDiskOffsets(raidPtr, psid, diskno, &offset, &fd_offset,
1694 1.57 oster &new_rbuf->spCol, &new_rbuf->spOffset); /* find offsets & spare
1695 1.4 oster * location */
1696 1.4 oster new_rbuf->parityStripeID = psid; /* fill in the buffer */
1697 1.4 oster new_rbuf->which_ru = which_ru;
1698 1.4 oster new_rbuf->failedDiskSectorOffset = fd_offset;
1699 1.4 oster new_rbuf->priority = RF_IO_NORMAL_PRIORITY;
1700 1.4 oster
1701 1.4 oster /* use NULL b_proc b/c all addrs
1702 1.4 oster * should be in kernel space */
1703 1.4 oster req = rf_CreateDiskQueueData(RF_IO_TYPE_READ, offset + which_ru * sectorsPerRU, sectorsPerRU, new_rbuf->buffer,
1704 1.81.2.1 yamt psid, which_ru, (int (*) (void *, int)) ForceReconReadDoneProc, (void *) new_rbuf,
1705 1.81.2.1 yamt NULL, (void *) raidPtr, 0, NULL, PR_WAITOK);
1706 1.4 oster
1707 1.4 oster new_rbuf->arg = req;
1708 1.57 oster rf_DiskIOEnqueue(&raidPtr->Queues[diskno], req, RF_IO_NORMAL_PRIORITY); /* enqueue the I/O */
1709 1.57 oster Dprintf2("raid%d: Issued new read req on col %d\n", raidPtr->raidid, diskno);
1710 1.4 oster }
1711 1.4 oster }
1712 1.4 oster /* if the write is sitting in the disk queue, elevate its
1713 1.4 oster * priority */
1714 1.57 oster if (rf_DiskIOPromote(&raidPtr->Queues[fcol], psid, which_ru))
1715 1.57 oster printf("raid%d: promoted write to col %d\n",
1716 1.57 oster raidPtr->raidid, fcol);
1717 1.4 oster }
1718 1.4 oster /* install a callback descriptor to be invoked when recon completes on
1719 1.4 oster * this parity stripe. */
1720 1.4 oster cb = rf_AllocCallbackDesc();
1721 1.4 oster /* XXX the following is bogus.. These functions don't really match!!
1722 1.4 oster * GO */
1723 1.4 oster cb->callbackFunc = (void (*) (RF_CBParam_t)) cbFunc;
1724 1.4 oster cb->callbackArg.p = (void *) cbArg;
1725 1.4 oster cb->next = pssPtr->procWaitList;
1726 1.4 oster pssPtr->procWaitList = cb;
1727 1.10 oster DDprintf2("raid%d: Waiting for forced recon on psid %ld\n",
1728 1.10 oster raidPtr->raidid, psid);
1729 1.4 oster
1730 1.57 oster RF_UNLOCK_PSS_MUTEX(raidPtr, psid);
1731 1.4 oster return (1);
1732 1.1 oster }
1733 1.1 oster /* called upon the completion of a forced reconstruction read.
1734 1.1 oster * all we do is schedule the FORCEDREADONE event.
1735 1.1 oster * called at interrupt context in the kernel, so don't do anything illegal here.
1736 1.1 oster */
1737 1.4 oster static void
1738 1.60 oster ForceReconReadDoneProc(void *arg, int status)
1739 1.4 oster {
1740 1.4 oster RF_ReconBuffer_t *rbuf = arg;
1741 1.4 oster
1742 1.81.2.1 yamt /* Detect that reconControl is no longer valid, and if that
1743 1.81.2.1 yamt is the case, bail without calling rf_CauseReconEvent().
1744 1.81.2.1 yamt There won't be anyone listening for this event anyway */
1745 1.81.2.1 yamt
1746 1.81.2.1 yamt if (rbuf->raidPtr->reconControl == NULL)
1747 1.81.2.1 yamt return;
1748 1.81.2.1 yamt
1749 1.4 oster if (status) {
1750 1.70 oster printf("raid%d: Forced recon read failed!\n", rbuf->raidPtr->raidid);
1751 1.71 oster rf_CauseReconEvent(rbuf->raidPtr, rbuf->col, (void *) rbuf, RF_REVENT_FORCEDREAD_FAILED);
1752 1.79 oster return;
1753 1.4 oster }
1754 1.71 oster rf_CauseReconEvent(rbuf->raidPtr, rbuf->col, (void *) rbuf, RF_REVENT_FORCEDREADDONE);
1755 1.1 oster }
1756 1.1 oster /* releases a block on the reconstruction of the indicated stripe */
1757 1.4 oster int
1758 1.60 oster rf_UnblockRecon(RF_Raid_t *raidPtr, RF_AccessStripeMap_t *asmap)
1759 1.4 oster {
1760 1.4 oster RF_StripeNum_t stripeID = asmap->stripeID;
1761 1.4 oster RF_ReconParityStripeStatus_t *pssPtr;
1762 1.4 oster RF_ReconUnitNum_t which_ru;
1763 1.4 oster RF_StripeNum_t psid;
1764 1.4 oster RF_CallbackDesc_t *cb;
1765 1.4 oster
1766 1.4 oster psid = rf_MapStripeIDToParityStripeID(&raidPtr->Layout, stripeID, &which_ru);
1767 1.57 oster RF_LOCK_PSS_MUTEX(raidPtr, psid);
1768 1.68 oster pssPtr = rf_LookupRUStatus(raidPtr, raidPtr->reconControl->pssTable, psid, which_ru, RF_PSS_NONE, NULL);
1769 1.4 oster
1770 1.4 oster /* When recon is forced, the pss desc can get deleted before we get
1771 1.4 oster * back to unblock recon. But, this can _only_ happen when recon is
1772 1.4 oster * forced. It would be good to put some kind of sanity check here, but
1773 1.4 oster * how to decide if recon was just forced or not? */
1774 1.4 oster if (!pssPtr) {
1775 1.4 oster /* printf("Warning: no pss descriptor upon unblock on psid %ld
1776 1.4 oster * RU %d\n",psid,which_ru); */
1777 1.43 oster #if (RF_DEBUG_RECON > 0) || (RF_DEBUG_PSS > 0)
1778 1.4 oster if (rf_reconDebug || rf_pssDebug)
1779 1.4 oster printf("Warning: no pss descriptor upon unblock on psid %ld RU %d\n", (long) psid, which_ru);
1780 1.43 oster #endif
1781 1.4 oster goto out;
1782 1.4 oster }
1783 1.4 oster pssPtr->blockCount--;
1784 1.10 oster Dprintf3("raid%d: unblocking recon on psid %ld: blockcount is %d\n",
1785 1.10 oster raidPtr->raidid, psid, pssPtr->blockCount);
1786 1.4 oster if (pssPtr->blockCount == 0) { /* if recon blockage has been released */
1787 1.4 oster
1788 1.4 oster /* unblock recon before calling CauseReconEvent in case
1789 1.4 oster * CauseReconEvent causes us to try to issue a new read before
1790 1.4 oster * returning here. */
1791 1.4 oster pssPtr->flags &= ~RF_PSS_RECON_BLOCKED;
1792 1.4 oster
1793 1.4 oster
1794 1.13 oster while (pssPtr->blockWaitList) {
1795 1.13 oster /* spin through the block-wait list and
1796 1.13 oster release all the waiters */
1797 1.4 oster cb = pssPtr->blockWaitList;
1798 1.4 oster pssPtr->blockWaitList = cb->next;
1799 1.4 oster cb->next = NULL;
1800 1.57 oster rf_CauseReconEvent(raidPtr, cb->col, NULL, RF_REVENT_BLOCKCLEAR);
1801 1.4 oster rf_FreeCallbackDesc(cb);
1802 1.4 oster }
1803 1.13 oster if (!(pssPtr->flags & RF_PSS_UNDER_RECON)) {
1804 1.13 oster /* if no recon was requested while recon was blocked */
1805 1.57 oster rf_PSStatusDelete(raidPtr, raidPtr->reconControl->pssTable, pssPtr);
1806 1.4 oster }
1807 1.4 oster }
1808 1.1 oster out:
1809 1.57 oster RF_UNLOCK_PSS_MUTEX(raidPtr, psid);
1810 1.4 oster return (0);
1811 1.1 oster }
1812