]>
Commit | Line | Data |
---|---|---|
1 | /* | |
2 | * mdadm - manage Linux "md" devices aka RAID arrays. | |
3 | * | |
4 | * Copyright (C) 2001-2012 Neil Brown <neilb@suse.de> | |
5 | * | |
6 | * | |
7 | * This program is free software; you can redistribute it and/or modify | |
8 | * it under the terms of the GNU General Public License as published by | |
9 | * the Free Software Foundation; either version 2 of the License, or | |
10 | * (at your option) any later version. | |
11 | * | |
12 | * This program is distributed in the hope that it will be useful, | |
13 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
14 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
15 | * GNU General Public License for more details. | |
16 | * | |
17 | * You should have received a copy of the GNU General Public License | |
18 | * along with this program; if not, write to the Free Software | |
19 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA | |
20 | * | |
21 | * Author: Neil Brown | |
22 | * Email: <neilb@suse.de> | |
23 | */ | |
24 | #include "mdadm.h" | |
25 | #include "dlink.h" | |
26 | #include <sys/mman.h> | |
27 | #include <stdint.h> | |
28 | ||
29 | #if ! defined(__BIG_ENDIAN) && ! defined(__LITTLE_ENDIAN) | |
30 | #error no endian defined | |
31 | #endif | |
32 | #include "md_u.h" | |
33 | #include "md_p.h" | |
34 | ||
35 | #ifndef offsetof | |
36 | #define offsetof(t,f) ((size_t)&(((t*)0)->f)) | |
37 | #endif | |
38 | ||
39 | int restore_backup(struct supertype *st, | |
40 | struct mdinfo *content, | |
41 | int working_disks, | |
42 | int next_spare, | |
43 | char *backup_file, | |
44 | int verbose) | |
45 | { | |
46 | int i; | |
47 | int *fdlist; | |
48 | struct mdinfo *dev; | |
49 | int err; | |
50 | int disk_count = next_spare + working_disks; | |
51 | ||
52 | dprintf("Called restore_backup()\n"); | |
53 | fdlist = xmalloc(sizeof(int) * disk_count); | |
54 | ||
55 | enable_fds(next_spare); | |
56 | for (i = 0; i < next_spare; i++) | |
57 | fdlist[i] = -1; | |
58 | for (dev = content->devs; dev; dev = dev->next) { | |
59 | char buf[22]; | |
60 | int fd; | |
61 | sprintf(buf, "%d:%d", | |
62 | dev->disk.major, | |
63 | dev->disk.minor); | |
64 | fd = dev_open(buf, O_RDWR); | |
65 | ||
66 | if (dev->disk.raid_disk >= 0) | |
67 | fdlist[dev->disk.raid_disk] = fd; | |
68 | else | |
69 | fdlist[next_spare++] = fd; | |
70 | } | |
71 | ||
72 | if (st->ss->external && st->ss->recover_backup) | |
73 | err = st->ss->recover_backup(st, content); | |
74 | else | |
75 | err = Grow_restart(st, content, fdlist, next_spare, | |
76 | backup_file, verbose > 0); | |
77 | ||
78 | while (next_spare > 0) { | |
79 | next_spare--; | |
80 | if (fdlist[next_spare] >= 0) | |
81 | close(fdlist[next_spare]); | |
82 | } | |
83 | free(fdlist); | |
84 | if (err) { | |
85 | pr_err("Failed to restore critical" | |
86 | " section for reshape - sorry.\n"); | |
87 | if (!backup_file) | |
88 | pr_err("Possibly you need" | |
89 | " to specify a --backup-file\n"); | |
90 | return 1; | |
91 | } | |
92 | ||
93 | dprintf("restore_backup() returns status OK.\n"); | |
94 | return 0; | |
95 | } | |
96 | ||
97 | int Grow_Add_device(char *devname, int fd, char *newdev) | |
98 | { | |
99 | /* Add a device to an active array. | |
100 | * Currently, just extend a linear array. | |
101 | * This requires writing a new superblock on the | |
102 | * new device, calling the kernel to add the device, | |
103 | * and if that succeeds, update the superblock on | |
104 | * all other devices. | |
105 | * This means that we need to *find* all other devices. | |
106 | */ | |
107 | struct mdinfo info; | |
108 | ||
109 | struct stat stb; | |
110 | int nfd, fd2; | |
111 | int d, nd; | |
112 | struct supertype *st = NULL; | |
113 | char *subarray = NULL; | |
114 | ||
115 | if (ioctl(fd, GET_ARRAY_INFO, &info.array) < 0) { | |
116 | pr_err("cannot get array info for %s\n", devname); | |
117 | return 1; | |
118 | } | |
119 | ||
120 | if (info.array.level != -1) { | |
121 | pr_err("can only add devices to linear arrays\n"); | |
122 | return 1; | |
123 | } | |
124 | ||
125 | st = super_by_fd(fd, &subarray); | |
126 | if (!st) { | |
127 | pr_err("cannot handle arrays with superblock version %d\n", | |
128 | info.array.major_version); | |
129 | return 1; | |
130 | } | |
131 | ||
132 | if (subarray) { | |
133 | pr_err("Cannot grow linear sub-arrays yet\n"); | |
134 | free(subarray); | |
135 | free(st); | |
136 | return 1; | |
137 | } | |
138 | ||
139 | nfd = open(newdev, O_RDWR|O_EXCL|O_DIRECT); | |
140 | if (nfd < 0) { | |
141 | pr_err("cannot open %s\n", newdev); | |
142 | free(st); | |
143 | return 1; | |
144 | } | |
145 | fstat(nfd, &stb); | |
146 | if ((stb.st_mode & S_IFMT) != S_IFBLK) { | |
147 | pr_err("%s is not a block device!\n", newdev); | |
148 | close(nfd); | |
149 | free(st); | |
150 | return 1; | |
151 | } | |
152 | /* now check out all the devices and make sure we can read the | |
153 | * superblock */ | |
154 | for (d=0 ; d < info.array.raid_disks ; d++) { | |
155 | mdu_disk_info_t disk; | |
156 | char *dv; | |
157 | ||
158 | st->ss->free_super(st); | |
159 | ||
160 | disk.number = d; | |
161 | if (ioctl(fd, GET_DISK_INFO, &disk) < 0) { | |
162 | pr_err("cannot get device detail for device %d\n", | |
163 | d); | |
164 | close(nfd); | |
165 | free(st); | |
166 | return 1; | |
167 | } | |
168 | dv = map_dev(disk.major, disk.minor, 1); | |
169 | if (!dv) { | |
170 | pr_err("cannot find device file for device %d\n", | |
171 | d); | |
172 | close(nfd); | |
173 | free(st); | |
174 | return 1; | |
175 | } | |
176 | fd2 = dev_open(dv, O_RDWR); | |
177 | if (fd2 < 0) { | |
178 | pr_err("cannot open device file %s\n", dv); | |
179 | close(nfd); | |
180 | free(st); | |
181 | return 1; | |
182 | } | |
183 | ||
184 | if (st->ss->load_super(st, fd2, NULL)) { | |
185 | pr_err("cannot find super block on %s\n", dv); | |
186 | close(nfd); | |
187 | close(fd2); | |
188 | free(st); | |
189 | return 1; | |
190 | } | |
191 | close(fd2); | |
192 | } | |
193 | /* Ok, looks good. Lets update the superblock and write it out to | |
194 | * newdev. | |
195 | */ | |
196 | ||
197 | info.disk.number = d; | |
198 | info.disk.major = major(stb.st_rdev); | |
199 | info.disk.minor = minor(stb.st_rdev); | |
200 | info.disk.raid_disk = d; | |
201 | info.disk.state = (1 << MD_DISK_SYNC) | (1 << MD_DISK_ACTIVE); | |
202 | st->ss->update_super(st, &info, "linear-grow-new", newdev, | |
203 | 0, 0, NULL); | |
204 | ||
205 | if (st->ss->store_super(st, nfd)) { | |
206 | pr_err("Cannot store new superblock on %s\n", | |
207 | newdev); | |
208 | close(nfd); | |
209 | return 1; | |
210 | } | |
211 | close(nfd); | |
212 | ||
213 | if (ioctl(fd, ADD_NEW_DISK, &info.disk) != 0) { | |
214 | pr_err("Cannot add new disk to this array\n"); | |
215 | return 1; | |
216 | } | |
217 | /* Well, that seems to have worked. | |
218 | * Now go through and update all superblocks | |
219 | */ | |
220 | ||
221 | if (ioctl(fd, GET_ARRAY_INFO, &info.array) < 0) { | |
222 | pr_err("cannot get array info for %s\n", devname); | |
223 | return 1; | |
224 | } | |
225 | ||
226 | nd = d; | |
227 | for (d=0 ; d < info.array.raid_disks ; d++) { | |
228 | mdu_disk_info_t disk; | |
229 | char *dv; | |
230 | ||
231 | disk.number = d; | |
232 | if (ioctl(fd, GET_DISK_INFO, &disk) < 0) { | |
233 | pr_err("cannot get device detail for device %d\n", | |
234 | d); | |
235 | return 1; | |
236 | } | |
237 | dv = map_dev(disk.major, disk.minor, 1); | |
238 | if (!dv) { | |
239 | pr_err("cannot find device file for device %d\n", | |
240 | d); | |
241 | return 1; | |
242 | } | |
243 | fd2 = dev_open(dv, O_RDWR); | |
244 | if (fd2 < 0) { | |
245 | pr_err("cannot open device file %s\n", dv); | |
246 | return 1; | |
247 | } | |
248 | if (st->ss->load_super(st, fd2, NULL)) { | |
249 | pr_err("cannot find super block on %s\n", dv); | |
250 | close(fd); | |
251 | return 1; | |
252 | } | |
253 | info.array.raid_disks = nd+1; | |
254 | info.array.nr_disks = nd+1; | |
255 | info.array.active_disks = nd+1; | |
256 | info.array.working_disks = nd+1; | |
257 | ||
258 | st->ss->update_super(st, &info, "linear-grow-update", dv, | |
259 | 0, 0, NULL); | |
260 | ||
261 | if (st->ss->store_super(st, fd2)) { | |
262 | pr_err("Cannot store new superblock on %s\n", dv); | |
263 | close(fd2); | |
264 | return 1; | |
265 | } | |
266 | close(fd2); | |
267 | } | |
268 | ||
269 | return 0; | |
270 | } | |
271 | ||
272 | int Grow_addbitmap(char *devname, int fd, struct context *c, struct shape *s) | |
273 | { | |
274 | /* | |
275 | * First check that array doesn't have a bitmap | |
276 | * Then create the bitmap | |
277 | * Then add it | |
278 | * | |
279 | * For internal bitmaps, we need to check the version, | |
280 | * find all the active devices, and write the bitmap block | |
281 | * to all devices | |
282 | */ | |
283 | mdu_bitmap_file_t bmf; | |
284 | mdu_array_info_t array; | |
285 | struct supertype *st; | |
286 | char *subarray = NULL; | |
287 | int major = BITMAP_MAJOR_HI; | |
288 | int vers = md_get_version(fd); | |
289 | unsigned long long bitmapsize, array_size; | |
290 | ||
291 | if (vers < 9003) { | |
292 | major = BITMAP_MAJOR_HOSTENDIAN; | |
293 | pr_err("Warning - bitmaps created on this kernel" | |
294 | " are not portable\n" | |
295 | " between different architectures. Consider upgrading" | |
296 | " the Linux kernel.\n"); | |
297 | } | |
298 | ||
299 | if (ioctl(fd, GET_BITMAP_FILE, &bmf) != 0) { | |
300 | if (errno == ENOMEM) | |
301 | pr_err("Memory allocation failure.\n"); | |
302 | else | |
303 | pr_err("bitmaps not supported by this kernel.\n"); | |
304 | return 1; | |
305 | } | |
306 | if (bmf.pathname[0]) { | |
307 | if (strcmp(s->bitmap_file,"none")==0) { | |
308 | if (ioctl(fd, SET_BITMAP_FILE, -1)!= 0) { | |
309 | pr_err("failed to remove bitmap %s\n", | |
310 | bmf.pathname); | |
311 | return 1; | |
312 | } | |
313 | return 0; | |
314 | } | |
315 | pr_err("%s already has a bitmap (%s)\n", | |
316 | devname, bmf.pathname); | |
317 | return 1; | |
318 | } | |
319 | if (ioctl(fd, GET_ARRAY_INFO, &array) != 0) { | |
320 | pr_err("cannot get array status for %s\n", devname); | |
321 | return 1; | |
322 | } | |
323 | if (array.state & (1<<MD_SB_BITMAP_PRESENT)) { | |
324 | if (strcmp(s->bitmap_file, "none")==0) { | |
325 | array.state &= ~(1<<MD_SB_BITMAP_PRESENT); | |
326 | if (ioctl(fd, SET_ARRAY_INFO, &array)!= 0) { | |
327 | pr_err("failed to remove internal bitmap.\n"); | |
328 | return 1; | |
329 | } | |
330 | return 0; | |
331 | } | |
332 | pr_err("Internal bitmap already present on %s\n", | |
333 | devname); | |
334 | return 1; | |
335 | } | |
336 | ||
337 | if (strcmp(s->bitmap_file, "none") == 0) { | |
338 | pr_err("no bitmap found on %s\n", devname); | |
339 | return 1; | |
340 | } | |
341 | if (array.level <= 0) { | |
342 | pr_err("Bitmaps not meaningful with level %s\n", | |
343 | map_num(pers, array.level)?:"of this array"); | |
344 | return 1; | |
345 | } | |
346 | bitmapsize = array.size; | |
347 | bitmapsize <<= 1; | |
348 | if (get_dev_size(fd, NULL, &array_size) && | |
349 | array_size > (0x7fffffffULL<<9)) { | |
350 | /* Array is big enough that we cannot trust array.size | |
351 | * try other approaches | |
352 | */ | |
353 | bitmapsize = get_component_size(fd); | |
354 | } | |
355 | if (bitmapsize == 0) { | |
356 | pr_err("Cannot reliably determine size of array to create bitmap - sorry.\n"); | |
357 | return 1; | |
358 | } | |
359 | ||
360 | if (array.level == 10) { | |
361 | int ncopies = (array.layout&255)*((array.layout>>8)&255); | |
362 | bitmapsize = bitmapsize * array.raid_disks / ncopies; | |
363 | } | |
364 | ||
365 | st = super_by_fd(fd, &subarray); | |
366 | if (!st) { | |
367 | pr_err("Cannot understand version %d.%d\n", | |
368 | array.major_version, array.minor_version); | |
369 | return 1; | |
370 | } | |
371 | if (subarray) { | |
372 | pr_err("Cannot add bitmaps to sub-arrays yet\n"); | |
373 | free(subarray); | |
374 | free(st); | |
375 | return 1; | |
376 | } | |
377 | if (strcmp(s->bitmap_file, "internal") == 0) { | |
378 | int rv; | |
379 | int d; | |
380 | int offset_setable = 0; | |
381 | struct mdinfo *mdi; | |
382 | if (st->ss->add_internal_bitmap == NULL) { | |
383 | pr_err("Internal bitmaps not supported " | |
384 | "with %s metadata\n", st->ss->name); | |
385 | return 1; | |
386 | } | |
387 | mdi = sysfs_read(fd, NULL, GET_BITMAP_LOCATION); | |
388 | if (mdi) | |
389 | offset_setable = 1; | |
390 | for (d=0; d< st->max_devs; d++) { | |
391 | mdu_disk_info_t disk; | |
392 | char *dv; | |
393 | disk.number = d; | |
394 | if (ioctl(fd, GET_DISK_INFO, &disk) < 0) | |
395 | continue; | |
396 | if (disk.major == 0 && | |
397 | disk.minor == 0) | |
398 | continue; | |
399 | if ((disk.state & (1<<MD_DISK_SYNC))==0) | |
400 | continue; | |
401 | dv = map_dev(disk.major, disk.minor, 1); | |
402 | if (dv) { | |
403 | int fd2 = dev_open(dv, O_RDWR); | |
404 | if (fd2 < 0) | |
405 | continue; | |
406 | if (st->ss->load_super(st, fd2, NULL)==0) { | |
407 | if (st->ss->add_internal_bitmap( | |
408 | st, | |
409 | &s->bitmap_chunk, c->delay, s->write_behind, | |
410 | bitmapsize, offset_setable, | |
411 | major) | |
412 | ) | |
413 | st->ss->write_bitmap(st, fd2); | |
414 | else { | |
415 | pr_err("failed to create internal bitmap" | |
416 | " - chunksize problem.\n"); | |
417 | close(fd2); | |
418 | return 1; | |
419 | } | |
420 | } | |
421 | close(fd2); | |
422 | } | |
423 | } | |
424 | if (offset_setable) { | |
425 | st->ss->getinfo_super(st, mdi, NULL); | |
426 | sysfs_init(mdi, fd, NULL); | |
427 | rv = sysfs_set_num_signed(mdi, NULL, "bitmap/location", | |
428 | mdi->bitmap_offset); | |
429 | } else { | |
430 | array.state |= (1<<MD_SB_BITMAP_PRESENT); | |
431 | rv = ioctl(fd, SET_ARRAY_INFO, &array); | |
432 | } | |
433 | if (rv < 0) { | |
434 | if (errno == EBUSY) | |
435 | pr_err("Cannot add bitmap while array is" | |
436 | " resyncing or reshaping etc.\n"); | |
437 | pr_err("failed to set internal bitmap.\n"); | |
438 | return 1; | |
439 | } | |
440 | } else { | |
441 | int uuid[4]; | |
442 | int bitmap_fd; | |
443 | int d; | |
444 | int max_devs = st->max_devs; | |
445 | ||
446 | /* try to load a superblock */ | |
447 | for (d = 0; d < max_devs; d++) { | |
448 | mdu_disk_info_t disk; | |
449 | char *dv; | |
450 | int fd2; | |
451 | disk.number = d; | |
452 | if (ioctl(fd, GET_DISK_INFO, &disk) < 0) | |
453 | continue; | |
454 | if ((disk.major==0 && disk.minor==0) || | |
455 | (disk.state & (1<<MD_DISK_REMOVED))) | |
456 | continue; | |
457 | dv = map_dev(disk.major, disk.minor, 1); | |
458 | if (!dv) | |
459 | continue; | |
460 | fd2 = dev_open(dv, O_RDONLY); | |
461 | if (fd2 >= 0) { | |
462 | if (st->ss->load_super(st, fd2, NULL) == 0) { | |
463 | close(fd2); | |
464 | st->ss->uuid_from_super(st, uuid); | |
465 | break; | |
466 | } | |
467 | close(fd2); | |
468 | } | |
469 | } | |
470 | if (d == max_devs) { | |
471 | pr_err("cannot find UUID for array!\n"); | |
472 | return 1; | |
473 | } | |
474 | if (CreateBitmap(s->bitmap_file, c->force, (char*)uuid, s->bitmap_chunk, | |
475 | c->delay, s->write_behind, bitmapsize, major)) { | |
476 | return 1; | |
477 | } | |
478 | bitmap_fd = open(s->bitmap_file, O_RDWR); | |
479 | if (bitmap_fd < 0) { | |
480 | pr_err("weird: %s cannot be opened\n", | |
481 | s->bitmap_file); | |
482 | return 1; | |
483 | } | |
484 | if (ioctl(fd, SET_BITMAP_FILE, bitmap_fd) < 0) { | |
485 | int err = errno; | |
486 | if (errno == EBUSY) | |
487 | pr_err("Cannot add bitmap while array is" | |
488 | " resyncing or reshaping etc.\n"); | |
489 | pr_err("Cannot set bitmap file for %s: %s\n", | |
490 | devname, strerror(err)); | |
491 | return 1; | |
492 | } | |
493 | } | |
494 | ||
495 | return 0; | |
496 | } | |
497 | ||
498 | /* | |
499 | * When reshaping an array we might need to backup some data. | |
500 | * This is written to all spares with a 'super_block' describing it. | |
501 | * The superblock goes 4K from the end of the used space on the | |
502 | * device. | |
503 | * It if written after the backup is complete. | |
504 | * It has the following structure. | |
505 | */ | |
506 | ||
507 | static struct mdp_backup_super { | |
508 | char magic[16]; /* md_backup_data-1 or -2 */ | |
509 | __u8 set_uuid[16]; | |
510 | __u64 mtime; | |
511 | /* start/sizes in 512byte sectors */ | |
512 | __u64 devstart; /* address on backup device/file of data */ | |
513 | __u64 arraystart; | |
514 | __u64 length; | |
515 | __u32 sb_csum; /* csum of preceeding bytes. */ | |
516 | __u32 pad1; | |
517 | __u64 devstart2; /* offset in to data of second section */ | |
518 | __u64 arraystart2; | |
519 | __u64 length2; | |
520 | __u32 sb_csum2; /* csum of preceeding bytes. */ | |
521 | __u8 pad[512-68-32]; | |
522 | } __attribute__((aligned(512))) bsb, bsb2; | |
523 | ||
524 | static __u32 bsb_csum(char *buf, int len) | |
525 | { | |
526 | int i; | |
527 | int csum = 0; | |
528 | for (i = 0; i < len; i++) | |
529 | csum = (csum<<3) + buf[0]; | |
530 | return __cpu_to_le32(csum); | |
531 | } | |
532 | ||
533 | static int check_idle(struct supertype *st) | |
534 | { | |
535 | /* Check that all member arrays for this container, or the | |
536 | * container of this array, are idle | |
537 | */ | |
538 | char *container = (st->container_devnm[0] | |
539 | ? st->container_devnm : st->devnm); | |
540 | struct mdstat_ent *ent, *e; | |
541 | int is_idle = 1; | |
542 | ||
543 | ent = mdstat_read(0, 0); | |
544 | for (e = ent ; e; e = e->next) { | |
545 | if (!is_container_member(e, container)) | |
546 | continue; | |
547 | if (e->percent >= 0) { | |
548 | is_idle = 0; | |
549 | break; | |
550 | } | |
551 | } | |
552 | free_mdstat(ent); | |
553 | return is_idle; | |
554 | } | |
555 | ||
556 | static int freeze_container(struct supertype *st) | |
557 | { | |
558 | char *container = (st->container_devnm[0] | |
559 | ? st->container_devnm : st->devnm); | |
560 | ||
561 | if (!check_idle(st)) | |
562 | return -1; | |
563 | ||
564 | if (block_monitor(container, 1)) { | |
565 | pr_err("failed to freeze container\n"); | |
566 | return -2; | |
567 | } | |
568 | ||
569 | return 1; | |
570 | } | |
571 | ||
572 | static void unfreeze_container(struct supertype *st) | |
573 | { | |
574 | char *container = (st->container_devnm[0] | |
575 | ? st->container_devnm : st->devnm); | |
576 | ||
577 | unblock_monitor(container, 1); | |
578 | } | |
579 | ||
580 | static int freeze(struct supertype *st) | |
581 | { | |
582 | /* Try to freeze resync/rebuild on this array/container. | |
583 | * Return -1 if the array is busy, | |
584 | * return -2 container cannot be frozen, | |
585 | * return 0 if this kernel doesn't support 'frozen' | |
586 | * return 1 if it worked. | |
587 | */ | |
588 | if (st->ss->external) | |
589 | return freeze_container(st); | |
590 | else { | |
591 | struct mdinfo *sra = sysfs_read(-1, st->devnm, GET_VERSION); | |
592 | int err; | |
593 | char buf[20]; | |
594 | ||
595 | if (!sra) | |
596 | return -1; | |
597 | /* Need to clear any 'read-auto' status */ | |
598 | if (sysfs_get_str(sra, NULL, "array_state", buf, 20) > 0 && | |
599 | strncmp(buf, "read-auto", 9) == 0) | |
600 | sysfs_set_str(sra, NULL, "array_state", "clean"); | |
601 | ||
602 | err = sysfs_freeze_array(sra); | |
603 | sysfs_free(sra); | |
604 | return err; | |
605 | } | |
606 | } | |
607 | ||
608 | static void unfreeze(struct supertype *st) | |
609 | { | |
610 | if (st->ss->external) | |
611 | return unfreeze_container(st); | |
612 | else { | |
613 | struct mdinfo *sra = sysfs_read(-1, st->devnm, GET_VERSION); | |
614 | ||
615 | if (sra) | |
616 | sysfs_set_str(sra, NULL, "sync_action", "idle"); | |
617 | sysfs_free(sra); | |
618 | } | |
619 | } | |
620 | ||
621 | static void wait_reshape(struct mdinfo *sra) | |
622 | { | |
623 | int fd = sysfs_get_fd(sra, NULL, "sync_action"); | |
624 | char action[20]; | |
625 | ||
626 | if (fd < 0) | |
627 | return; | |
628 | ||
629 | while (sysfs_fd_get_str(fd, action, 20) > 0 && | |
630 | strncmp(action, "reshape", 7) == 0) { | |
631 | fd_set rfds; | |
632 | FD_ZERO(&rfds); | |
633 | FD_SET(fd, &rfds); | |
634 | select(fd+1, NULL, NULL, &rfds, NULL); | |
635 | } | |
636 | close(fd); | |
637 | } | |
638 | ||
639 | static int reshape_super(struct supertype *st, unsigned long long size, | |
640 | int level, int layout, int chunksize, int raid_disks, | |
641 | int delta_disks, char *backup_file, char *dev, | |
642 | int direction, int verbose) | |
643 | { | |
644 | /* nothing extra to check in the native case */ | |
645 | if (!st->ss->external) | |
646 | return 0; | |
647 | if (!st->ss->reshape_super || | |
648 | !st->ss->manage_reshape) { | |
649 | pr_err("%s metadata does not support reshape\n", | |
650 | st->ss->name); | |
651 | return 1; | |
652 | } | |
653 | ||
654 | return st->ss->reshape_super(st, size, level, layout, chunksize, | |
655 | raid_disks, delta_disks, backup_file, dev, | |
656 | direction, verbose); | |
657 | } | |
658 | ||
659 | static void sync_metadata(struct supertype *st) | |
660 | { | |
661 | if (st->ss->external) { | |
662 | if (st->update_tail) { | |
663 | flush_metadata_updates(st); | |
664 | st->update_tail = &st->updates; | |
665 | } else | |
666 | st->ss->sync_metadata(st); | |
667 | } | |
668 | } | |
669 | ||
670 | static int subarray_set_num(char *container, struct mdinfo *sra, char *name, int n) | |
671 | { | |
672 | /* when dealing with external metadata subarrays we need to be | |
673 | * prepared to handle EAGAIN. The kernel may need to wait for | |
674 | * mdmon to mark the array active so the kernel can handle | |
675 | * allocations/writeback when preparing the reshape action | |
676 | * (md_allow_write()). We temporarily disable safe_mode_delay | |
677 | * to close a race with the array_state going clean before the | |
678 | * next write to raid_disks / stripe_cache_size | |
679 | */ | |
680 | char safe[50]; | |
681 | int rc; | |
682 | ||
683 | /* only 'raid_disks' and 'stripe_cache_size' trigger md_allow_write */ | |
684 | if (!container || | |
685 | (strcmp(name, "raid_disks") != 0 && | |
686 | strcmp(name, "stripe_cache_size") != 0)) | |
687 | return sysfs_set_num(sra, NULL, name, n); | |
688 | ||
689 | rc = sysfs_get_str(sra, NULL, "safe_mode_delay", safe, sizeof(safe)); | |
690 | if (rc <= 0) | |
691 | return -1; | |
692 | sysfs_set_num(sra, NULL, "safe_mode_delay", 0); | |
693 | rc = sysfs_set_num(sra, NULL, name, n); | |
694 | if (rc < 0 && errno == EAGAIN) { | |
695 | ping_monitor(container); | |
696 | /* if we get EAGAIN here then the monitor is not active | |
697 | * so stop trying | |
698 | */ | |
699 | rc = sysfs_set_num(sra, NULL, name, n); | |
700 | } | |
701 | sysfs_set_str(sra, NULL, "safe_mode_delay", safe); | |
702 | return rc; | |
703 | } | |
704 | ||
705 | int start_reshape(struct mdinfo *sra, int already_running, | |
706 | int before_data_disks, int data_disks) | |
707 | { | |
708 | int err; | |
709 | unsigned long long sync_max_to_set; | |
710 | ||
711 | sysfs_set_num(sra, NULL, "suspend_lo", 0x7FFFFFFFFFFFFFFFULL); | |
712 | err = sysfs_set_num(sra, NULL, "suspend_hi", sra->reshape_progress); | |
713 | err = err ?: sysfs_set_num(sra, NULL, "suspend_lo", | |
714 | sra->reshape_progress); | |
715 | if (before_data_disks <= data_disks) | |
716 | sync_max_to_set = sra->reshape_progress / data_disks; | |
717 | else | |
718 | sync_max_to_set = (sra->component_size * data_disks | |
719 | - sra->reshape_progress) / data_disks; | |
720 | if (!already_running) | |
721 | sysfs_set_num(sra, NULL, "sync_min", sync_max_to_set); | |
722 | err = err ?: sysfs_set_num(sra, NULL, "sync_max", sync_max_to_set); | |
723 | if (!already_running) | |
724 | err = err ?: sysfs_set_str(sra, NULL, "sync_action", "reshape"); | |
725 | ||
726 | return err; | |
727 | } | |
728 | ||
729 | void abort_reshape(struct mdinfo *sra) | |
730 | { | |
731 | sysfs_set_str(sra, NULL, "sync_action", "idle"); | |
732 | sysfs_set_num(sra, NULL, "suspend_lo", 0x7FFFFFFFFFFFFFFFULL); | |
733 | sysfs_set_num(sra, NULL, "suspend_hi", 0); | |
734 | sysfs_set_num(sra, NULL, "suspend_lo", 0); | |
735 | sysfs_set_num(sra, NULL, "sync_min", 0); | |
736 | sysfs_set_str(sra, NULL, "sync_max", "max"); | |
737 | } | |
738 | ||
739 | int remove_disks_for_takeover(struct supertype *st, | |
740 | struct mdinfo *sra, | |
741 | int layout) | |
742 | { | |
743 | int nr_of_copies; | |
744 | struct mdinfo *remaining; | |
745 | int slot; | |
746 | ||
747 | if (sra->array.level == 10) | |
748 | nr_of_copies = layout & 0xff; | |
749 | else if (sra->array.level == 1) | |
750 | nr_of_copies = sra->array.raid_disks; | |
751 | else | |
752 | return 1; | |
753 | ||
754 | remaining = sra->devs; | |
755 | sra->devs = NULL; | |
756 | /* for each 'copy', select one device and remove from the list. */ | |
757 | for (slot = 0; slot < sra->array.raid_disks; slot += nr_of_copies) { | |
758 | struct mdinfo **diskp; | |
759 | int found = 0; | |
760 | ||
761 | /* Find a working device to keep */ | |
762 | for (diskp = &remaining; *diskp ; diskp = &(*diskp)->next) { | |
763 | struct mdinfo *disk = *diskp; | |
764 | ||
765 | if (disk->disk.raid_disk < slot) | |
766 | continue; | |
767 | if (disk->disk.raid_disk >= slot + nr_of_copies) | |
768 | continue; | |
769 | if (disk->disk.state & (1<<MD_DISK_REMOVED)) | |
770 | continue; | |
771 | if (disk->disk.state & (1<<MD_DISK_FAULTY)) | |
772 | continue; | |
773 | if (!(disk->disk.state & (1<<MD_DISK_SYNC))) | |
774 | continue; | |
775 | ||
776 | /* We have found a good disk to use! */ | |
777 | *diskp = disk->next; | |
778 | disk->next = sra->devs; | |
779 | sra->devs = disk; | |
780 | found = 1; | |
781 | break; | |
782 | } | |
783 | if (!found) | |
784 | break; | |
785 | } | |
786 | ||
787 | if (slot < sra->array.raid_disks) { | |
788 | /* didn't find all slots */ | |
789 | struct mdinfo **e; | |
790 | e = &remaining; | |
791 | while (*e) | |
792 | e = &(*e)->next; | |
793 | *e = sra->devs; | |
794 | sra->devs = remaining; | |
795 | return 1; | |
796 | } | |
797 | ||
798 | /* Remove all 'remaining' devices from the array */ | |
799 | while (remaining) { | |
800 | struct mdinfo *sd = remaining; | |
801 | remaining = sd->next; | |
802 | ||
803 | sysfs_set_str(sra, sd, "state", "faulty"); | |
804 | sysfs_set_str(sra, sd, "slot", "none"); | |
805 | /* for external metadata disks should be removed in mdmon */ | |
806 | if (!st->ss->external) | |
807 | sysfs_set_str(sra, sd, "state", "remove"); | |
808 | sd->disk.state |= (1<<MD_DISK_REMOVED); | |
809 | sd->disk.state &= ~(1<<MD_DISK_SYNC); | |
810 | sd->next = sra->devs; | |
811 | sra->devs = sd; | |
812 | } | |
813 | return 0; | |
814 | } | |
815 | ||
816 | void reshape_free_fdlist(int *fdlist, | |
817 | unsigned long long *offsets, | |
818 | int size) | |
819 | { | |
820 | int i; | |
821 | ||
822 | for (i = 0; i < size; i++) | |
823 | if (fdlist[i] >= 0) | |
824 | close(fdlist[i]); | |
825 | ||
826 | free(fdlist); | |
827 | free(offsets); | |
828 | } | |
829 | ||
830 | int reshape_prepare_fdlist(char *devname, | |
831 | struct mdinfo *sra, | |
832 | int raid_disks, | |
833 | int nrdisks, | |
834 | unsigned long blocks, | |
835 | char *backup_file, | |
836 | int *fdlist, | |
837 | unsigned long long *offsets) | |
838 | { | |
839 | int d = 0; | |
840 | struct mdinfo *sd; | |
841 | ||
842 | enable_fds(nrdisks); | |
843 | for (d = 0; d <= nrdisks; d++) | |
844 | fdlist[d] = -1; | |
845 | d = raid_disks; | |
846 | for (sd = sra->devs; sd; sd = sd->next) { | |
847 | if (sd->disk.state & (1<<MD_DISK_FAULTY)) | |
848 | continue; | |
849 | if (sd->disk.state & (1<<MD_DISK_SYNC)) { | |
850 | char *dn = map_dev(sd->disk.major, | |
851 | sd->disk.minor, 1); | |
852 | fdlist[sd->disk.raid_disk] | |
853 | = dev_open(dn, O_RDONLY); | |
854 | offsets[sd->disk.raid_disk] = sd->data_offset*512; | |
855 | if (fdlist[sd->disk.raid_disk] < 0) { | |
856 | pr_err("%s: cannot open component %s\n", | |
857 | devname, dn ? dn : "-unknown-"); | |
858 | d = -1; | |
859 | goto release; | |
860 | } | |
861 | } else if (backup_file == NULL) { | |
862 | /* spare */ | |
863 | char *dn = map_dev(sd->disk.major, | |
864 | sd->disk.minor, 1); | |
865 | fdlist[d] = dev_open(dn, O_RDWR); | |
866 | offsets[d] = (sd->data_offset + sra->component_size - blocks - 8)*512; | |
867 | if (fdlist[d] < 0) { | |
868 | pr_err("%s: cannot open component %s\n", | |
869 | devname, dn ? dn : "-unknown-"); | |
870 | d = -1; | |
871 | goto release; | |
872 | } | |
873 | d++; | |
874 | } | |
875 | } | |
876 | release: | |
877 | return d; | |
878 | } | |
879 | ||
880 | int reshape_open_backup_file(char *backup_file, | |
881 | int fd, | |
882 | char *devname, | |
883 | long blocks, | |
884 | int *fdlist, | |
885 | unsigned long long *offsets, | |
886 | int restart) | |
887 | { | |
888 | /* Return 1 on success, 0 on any form of failure */ | |
889 | /* need to check backup file is large enough */ | |
890 | char buf[512]; | |
891 | struct stat stb; | |
892 | unsigned int dev; | |
893 | int i; | |
894 | ||
895 | *fdlist = open(backup_file, O_RDWR|O_CREAT|(restart ? O_TRUNC : O_EXCL), | |
896 | S_IRUSR | S_IWUSR); | |
897 | *offsets = 8 * 512; | |
898 | if (*fdlist < 0) { | |
899 | pr_err("%s: cannot create backup file %s: %s\n", | |
900 | devname, backup_file, strerror(errno)); | |
901 | return 0; | |
902 | } | |
903 | /* Guard against backup file being on array device. | |
904 | * If array is partitioned or if LVM etc is in the | |
905 | * way this will not notice, but it is better than | |
906 | * nothing. | |
907 | */ | |
908 | fstat(*fdlist, &stb); | |
909 | dev = stb.st_dev; | |
910 | fstat(fd, &stb); | |
911 | if (stb.st_rdev == dev) { | |
912 | pr_err("backup file must NOT be" | |
913 | " on the array being reshaped.\n"); | |
914 | close(*fdlist); | |
915 | return 0; | |
916 | } | |
917 | ||
918 | memset(buf, 0, 512); | |
919 | for (i=0; i < blocks + 8 ; i++) { | |
920 | if (write(*fdlist, buf, 512) != 512) { | |
921 | pr_err("%s: cannot create" | |
922 | " backup file %s: %s\n", | |
923 | devname, backup_file, strerror(errno)); | |
924 | return 0; | |
925 | } | |
926 | } | |
927 | if (fsync(*fdlist) != 0) { | |
928 | pr_err("%s: cannot create backup file %s: %s\n", | |
929 | devname, backup_file, strerror(errno)); | |
930 | return 0; | |
931 | } | |
932 | ||
933 | return 1; | |
934 | } | |
935 | ||
936 | unsigned long GCD(unsigned long a, unsigned long b) | |
937 | { | |
938 | while (a != b) { | |
939 | if (a < b) | |
940 | b -= a; | |
941 | if (b < a) | |
942 | a -= b; | |
943 | } | |
944 | return a; | |
945 | } | |
946 | ||
947 | unsigned long compute_backup_blocks(int nchunk, int ochunk, | |
948 | unsigned int ndata, unsigned int odata) | |
949 | { | |
950 | unsigned long a, b, blocks; | |
951 | /* So how much do we need to backup. | |
952 | * We need an amount of data which is both a whole number of | |
953 | * old stripes and a whole number of new stripes. | |
954 | * So LCM for (chunksize*datadisks). | |
955 | */ | |
956 | a = (ochunk/512) * odata; | |
957 | b = (nchunk/512) * ndata; | |
958 | /* Find GCD */ | |
959 | a = GCD(a, b); | |
960 | /* LCM == product / GCD */ | |
961 | blocks = (ochunk/512) * (nchunk/512) * odata * ndata / a; | |
962 | ||
963 | return blocks; | |
964 | } | |
965 | ||
966 | char *analyse_change(struct mdinfo *info, struct reshape *re) | |
967 | { | |
968 | /* Based on the current array state in info->array and | |
969 | * the changes in info->new_* etc, determine: | |
970 | * - whether the change is possible | |
971 | * - Intermediate level/raid_disks/layout | |
972 | * - whether a restriping reshape is needed | |
973 | * - number of sectors in minimum change unit. This | |
974 | * will cover a whole number of stripes in 'before' and | |
975 | * 'after'. | |
976 | * | |
977 | * Return message if the change should be rejected | |
978 | * NULL if the change can be achieved | |
979 | * | |
980 | * This can be called as part of starting a reshape, or | |
981 | * when assembling an array that is undergoing reshape. | |
982 | */ | |
983 | int near, far, offset, copies; | |
984 | int new_disks; | |
985 | int old_chunk, new_chunk; | |
986 | /* delta_parity records change in number of devices | |
987 | * caused by level change | |
988 | */ | |
989 | int delta_parity = 0; | |
990 | ||
991 | memset(re, 0, sizeof(*re)); | |
992 | ||
993 | /* If a new level not explicitly given, we assume no-change */ | |
994 | if (info->new_level == UnSet) | |
995 | info->new_level = info->array.level; | |
996 | ||
997 | if (info->new_chunk) | |
998 | switch (info->new_level) { | |
999 | case 0: | |
1000 | case 4: | |
1001 | case 5: | |
1002 | case 6: | |
1003 | case 10: | |
1004 | /* chunk size is meaningful, must divide component_size | |
1005 | * evenly | |
1006 | */ | |
1007 | if (info->component_size % (info->new_chunk/512)) | |
1008 | return "New chunk size does not" | |
1009 | " divide component size"; | |
1010 | break; | |
1011 | default: | |
1012 | return "chunk size not meaningful for this level"; | |
1013 | } | |
1014 | else | |
1015 | info->new_chunk = info->array.chunk_size; | |
1016 | ||
1017 | switch (info->array.level) { | |
1018 | default: | |
1019 | return "Cannot understand this RAID level"; | |
1020 | case 1: | |
1021 | /* RAID1 can convert to RAID1 with different disks, or | |
1022 | * raid5 with 2 disks, or | |
1023 | * raid0 with 1 disk | |
1024 | */ | |
1025 | if (info->new_level > 1 && | |
1026 | (info->component_size & 7)) | |
1027 | return "Cannot convert RAID1 of this size - " | |
1028 | "reduce size to multiple of 4K first."; | |
1029 | if (info->new_level == 0) { | |
1030 | if (info->delta_disks != UnSet && | |
1031 | info->delta_disks != 0) | |
1032 | return "Cannot change number of disks " | |
1033 | "with RAID1->RAID0 conversion"; | |
1034 | re->level = 0; | |
1035 | re->before.data_disks = 1; | |
1036 | re->after.data_disks = 1; | |
1037 | return NULL; | |
1038 | } | |
1039 | if (info->new_level == 1) { | |
1040 | if (info->delta_disks == UnSet) | |
1041 | /* Don't know what to do */ | |
1042 | return "no change requested for Growing RAID1"; | |
1043 | re->level = 1; | |
1044 | return NULL; | |
1045 | } | |
1046 | if (info->array.raid_disks == 2 && | |
1047 | info->new_level == 5) { | |
1048 | ||
1049 | re->level = 5; | |
1050 | re->before.data_disks = 1; | |
1051 | if (info->delta_disks != UnSet && | |
1052 | info->delta_disks != 0) | |
1053 | re->after.data_disks = 1 + info->delta_disks; | |
1054 | else | |
1055 | re->after.data_disks = 1; | |
1056 | if (re->after.data_disks < 1) | |
1057 | return "Number of disks too small for RAID5"; | |
1058 | ||
1059 | re->before.layout = ALGORITHM_LEFT_SYMMETRIC; | |
1060 | info->array.chunk_size = 65536; | |
1061 | break; | |
1062 | } | |
1063 | /* Could do some multi-stage conversions, but leave that to | |
1064 | * later. | |
1065 | */ | |
1066 | return "Impossibly level change request for RAID1"; | |
1067 | ||
1068 | case 10: | |
1069 | /* RAID10 can be converted from near mode to | |
1070 | * RAID0 by removing some devices. | |
1071 | * It can also be reshaped if the kernel supports | |
1072 | * new_data_offset. | |
1073 | */ | |
1074 | switch (info->new_level) { | |
1075 | case 0: | |
1076 | if ((info->array.layout & ~0xff) != 0x100) | |
1077 | return "Cannot Grow RAID10 with far/offset layout"; | |
1078 | /* number of devices must be multiple of number of copies */ | |
1079 | if (info->array.raid_disks % (info->array.layout & 0xff)) | |
1080 | return "RAID10 layout too complex for Grow operation"; | |
1081 | ||
1082 | new_disks = (info->array.raid_disks | |
1083 | / (info->array.layout & 0xff)); | |
1084 | if (info->delta_disks == UnSet) | |
1085 | info->delta_disks = (new_disks | |
1086 | - info->array.raid_disks); | |
1087 | ||
1088 | if (info->delta_disks != new_disks - info->array.raid_disks) | |
1089 | return "New number of raid-devices impossible for RAID10"; | |
1090 | if (info->new_chunk && | |
1091 | info->new_chunk != info->array.chunk_size) | |
1092 | return "Cannot change chunk-size with RAID10 Grow"; | |
1093 | ||
1094 | /* looks good */ | |
1095 | re->level = 0; | |
1096 | re->before.data_disks = new_disks; | |
1097 | re->after.data_disks = re->before.data_disks; | |
1098 | return NULL; | |
1099 | ||
1100 | case 10: | |
1101 | near = info->array.layout & 0xff; | |
1102 | far = (info->array.layout >> 8) & 0xff; | |
1103 | offset = info->array.layout & 0x10000; | |
1104 | if (far > 1 && !offset) | |
1105 | return "Cannot reshape RAID10 in far-mode"; | |
1106 | copies = near * far; | |
1107 | ||
1108 | old_chunk = info->array.chunk_size * far; | |
1109 | ||
1110 | if (info->new_layout == UnSet) | |
1111 | info->new_layout = info->array.layout; | |
1112 | else { | |
1113 | near = info->new_layout & 0xff; | |
1114 | far = (info->new_layout >> 8) & 0xff; | |
1115 | offset = info->new_layout & 0x10000; | |
1116 | if (far > 1 && !offset) | |
1117 | return "Cannot reshape RAID10 to far-mode"; | |
1118 | if (near * far != copies) | |
1119 | return "Cannot change number of copies" | |
1120 | " when reshaping RAID10"; | |
1121 | } | |
1122 | if (info->delta_disks == UnSet) | |
1123 | info->delta_disks = 0; | |
1124 | new_disks = (info->array.raid_disks + | |
1125 | info->delta_disks); | |
1126 | ||
1127 | new_chunk = info->new_chunk * far; | |
1128 | ||
1129 | re->level = 10; | |
1130 | re->before.layout = info->array.layout; | |
1131 | re->before.data_disks = info->array.raid_disks; | |
1132 | re->after.layout = info->new_layout; | |
1133 | re->after.data_disks = new_disks; | |
1134 | /* For RAID10 we don't do backup but do allow reshape, | |
1135 | * so set backup_blocks to INVALID_SECTORS rather than | |
1136 | * zero. | |
1137 | * And there is no need to synchronise stripes on both | |
1138 | * 'old' and 'new'. So the important | |
1139 | * number is the minimum data_offset difference | |
1140 | * which is the larger of (offset copies * chunk). | |
1141 | */ | |
1142 | re->backup_blocks = INVALID_SECTORS; | |
1143 | re->min_offset_change = max(old_chunk, new_chunk) / 512; | |
1144 | if (new_disks < re->before.data_disks && | |
1145 | info->space_after < re->min_offset_change) | |
1146 | /* Reduce component size by one chunk */ | |
1147 | re->new_size = (info->component_size - | |
1148 | re->min_offset_change); | |
1149 | else | |
1150 | re->new_size = info->component_size; | |
1151 | re->new_size = re->new_size * new_disks / copies; | |
1152 | return NULL; | |
1153 | ||
1154 | default: | |
1155 | return "RAID10 can only be changed to RAID0"; | |
1156 | } | |
1157 | case 0: | |
1158 | /* RAID0 can be converted to RAID10, or to RAID456 */ | |
1159 | if (info->new_level == 10) { | |
1160 | if (info->new_layout == UnSet && info->delta_disks == UnSet) { | |
1161 | /* Assume near=2 layout */ | |
1162 | info->new_layout = 0x102; | |
1163 | info->delta_disks = info->array.raid_disks; | |
1164 | } | |
1165 | if (info->new_layout == UnSet) { | |
1166 | int copies = 1 + (info->delta_disks | |
1167 | / info->array.raid_disks); | |
1168 | if (info->array.raid_disks * (copies-1) | |
1169 | != info->delta_disks) | |
1170 | return "Impossible number of devices" | |
1171 | " for RAID0->RAID10"; | |
1172 | info->new_layout = 0x100 + copies; | |
1173 | } | |
1174 | if (info->delta_disks == UnSet) { | |
1175 | int copies = info->new_layout & 0xff; | |
1176 | if (info->new_layout != 0x100 + copies) | |
1177 | return "New layout impossible" | |
1178 | " for RAID0->RAID10";; | |
1179 | info->delta_disks = (copies - 1) * | |
1180 | info->array.raid_disks; | |
1181 | } | |
1182 | if (info->new_chunk && | |
1183 | info->new_chunk != info->array.chunk_size) | |
1184 | return "Cannot change chunk-size with RAID0->RAID10"; | |
1185 | /* looks good */ | |
1186 | re->level = 10; | |
1187 | re->before.data_disks = (info->array.raid_disks + | |
1188 | info->delta_disks); | |
1189 | re->after.data_disks = re->before.data_disks; | |
1190 | re->before.layout = info->new_layout; | |
1191 | return NULL; | |
1192 | } | |
1193 | ||
1194 | /* RAID0 can also covert to RAID0/4/5/6 by first converting to | |
1195 | * a raid4 style layout of the final level. | |
1196 | */ | |
1197 | switch (info->new_level) { | |
1198 | case 4: | |
1199 | delta_parity = 1; | |
1200 | case 0: | |
1201 | re->level = 4; | |
1202 | re->before.layout = 0; | |
1203 | break; | |
1204 | case 5: | |
1205 | delta_parity = 1; | |
1206 | re->level = 5; | |
1207 | re->before.layout = ALGORITHM_PARITY_N; | |
1208 | if (info->new_layout == UnSet) | |
1209 | info->new_layout = map_name(r5layout, "default"); | |
1210 | break; | |
1211 | case 6: | |
1212 | delta_parity = 2; | |
1213 | re->level = 6; | |
1214 | re->before.layout = ALGORITHM_PARITY_N; | |
1215 | if (info->new_layout == UnSet) | |
1216 | info->new_layout = map_name(r6layout, "default"); | |
1217 | break; | |
1218 | default: | |
1219 | return "Impossible level change requested"; | |
1220 | } | |
1221 | re->before.data_disks = info->array.raid_disks; | |
1222 | /* determining 'after' layout happens outside this 'switch' */ | |
1223 | break; | |
1224 | ||
1225 | case 4: | |
1226 | info->array.layout = ALGORITHM_PARITY_N; | |
1227 | case 5: | |
1228 | switch (info->new_level) { | |
1229 | case 0: | |
1230 | delta_parity = -1; | |
1231 | case 4: | |
1232 | re->level = info->array.level; | |
1233 | re->before.data_disks = info->array.raid_disks - 1; | |
1234 | re->before.layout = info->array.layout; | |
1235 | break; | |
1236 | case 5: | |
1237 | re->level = 5; | |
1238 | re->before.data_disks = info->array.raid_disks - 1; | |
1239 | re->before.layout = info->array.layout; | |
1240 | break; | |
1241 | case 6: | |
1242 | delta_parity = 1; | |
1243 | re->level = 6; | |
1244 | re->before.data_disks = info->array.raid_disks - 1; | |
1245 | switch (info->array.layout) { | |
1246 | case ALGORITHM_LEFT_ASYMMETRIC: | |
1247 | re->before.layout = ALGORITHM_LEFT_ASYMMETRIC_6; | |
1248 | break; | |
1249 | case ALGORITHM_RIGHT_ASYMMETRIC: | |
1250 | re->before.layout = ALGORITHM_RIGHT_ASYMMETRIC_6; | |
1251 | break; | |
1252 | case ALGORITHM_LEFT_SYMMETRIC: | |
1253 | re->before.layout = ALGORITHM_LEFT_SYMMETRIC_6; | |
1254 | break; | |
1255 | case ALGORITHM_RIGHT_SYMMETRIC: | |
1256 | re->before.layout = ALGORITHM_RIGHT_SYMMETRIC_6; | |
1257 | break; | |
1258 | case ALGORITHM_PARITY_0: | |
1259 | re->before.layout = ALGORITHM_PARITY_0_6; | |
1260 | break; | |
1261 | case ALGORITHM_PARITY_N: | |
1262 | re->before.layout = ALGORITHM_PARITY_N_6; | |
1263 | break; | |
1264 | default: | |
1265 | return "Cannot convert an array with this layout"; | |
1266 | } | |
1267 | break; | |
1268 | case 1: | |
1269 | if (info->array.raid_disks != 2) | |
1270 | return "Can only convert a 2-device array to RAID1"; | |
1271 | if (info->delta_disks != UnSet && | |
1272 | info->delta_disks != 0) | |
1273 | return "Cannot set raid_disk when " | |
1274 | "converting RAID5->RAID1"; | |
1275 | re->level = 1; | |
1276 | info->new_chunk = 0; | |
1277 | return NULL; | |
1278 | default: | |
1279 | return "Impossible level change requested"; | |
1280 | } | |
1281 | break; | |
1282 | case 6: | |
1283 | switch (info->new_level) { | |
1284 | case 4: | |
1285 | case 5: | |
1286 | delta_parity = -1; | |
1287 | case 6: | |
1288 | re->level = 6; | |
1289 | re->before.data_disks = info->array.raid_disks - 2; | |
1290 | re->before.layout = info->array.layout; | |
1291 | break; | |
1292 | default: | |
1293 | return "Impossible level change requested"; | |
1294 | } | |
1295 | break; | |
1296 | } | |
1297 | ||
1298 | /* If we reached here then it looks like a re-stripe is | |
1299 | * happening. We have determined the intermediate level | |
1300 | * and initial raid_disks/layout and stored these in 're'. | |
1301 | * | |
1302 | * We need to deduce the final layout that can be atomically | |
1303 | * converted to the end state. | |
1304 | */ | |
1305 | switch (info->new_level) { | |
1306 | case 0: | |
1307 | /* We can only get to RAID0 from RAID4 or RAID5 | |
1308 | * with appropriate layout and one extra device | |
1309 | */ | |
1310 | if (re->level != 4 && re->level != 5) | |
1311 | return "Cannot covert to RAID0 from this level"; | |
1312 | ||
1313 | switch (re->level) { | |
1314 | case 4: | |
1315 | re->before.layout = 0; | |
1316 | re->after.layout = 0; | |
1317 | break; | |
1318 | case 5: | |
1319 | re->after.layout = ALGORITHM_PARITY_N; | |
1320 | break; | |
1321 | } | |
1322 | break; | |
1323 | ||
1324 | case 4: | |
1325 | /* We can only get to RAID4 from RAID5 */ | |
1326 | if (re->level != 4 && re->level != 5) | |
1327 | return "Cannot convert to RAID4 from this level"; | |
1328 | ||
1329 | switch (re->level) { | |
1330 | case 4: | |
1331 | re->before.layout = 0; | |
1332 | re->after.layout = 0; | |
1333 | break; | |
1334 | case 5: | |
1335 | re->after.layout = ALGORITHM_PARITY_N; | |
1336 | break; | |
1337 | } | |
1338 | break; | |
1339 | ||
1340 | case 5: | |
1341 | /* We get to RAID5 from RAID5 or RAID6 */ | |
1342 | if (re->level != 5 && re->level != 6) | |
1343 | return "Cannot convert to RAID5 from this level"; | |
1344 | ||
1345 | switch (re->level) { | |
1346 | case 5: | |
1347 | if (info->new_layout == UnSet) | |
1348 | re->after.layout = re->before.layout; | |
1349 | else | |
1350 | re->after.layout = info->new_layout; | |
1351 | break; | |
1352 | case 6: | |
1353 | if (info->new_layout == UnSet) | |
1354 | info->new_layout = re->before.layout; | |
1355 | ||
1356 | /* after.layout needs to be raid6 version of new_layout */ | |
1357 | if (info->new_layout == ALGORITHM_PARITY_N) | |
1358 | re->after.layout = ALGORITHM_PARITY_N; | |
1359 | else { | |
1360 | char layout[40]; | |
1361 | char *ls = map_num(r5layout, info->new_layout); | |
1362 | int l; | |
1363 | if (ls) { | |
1364 | /* Current RAID6 layout has a RAID5 | |
1365 | * equivalent - good | |
1366 | */ | |
1367 | strcat(strcpy(layout, ls), "-6"); | |
1368 | l = map_name(r6layout, layout); | |
1369 | if (l == UnSet) | |
1370 | return "Cannot find RAID6 layout" | |
1371 | " to convert to"; | |
1372 | } else { | |
1373 | /* Current RAID6 has no equivalent. | |
1374 | * If it is already a '-6' layout we | |
1375 | * can leave it unchanged, else we must | |
1376 | * fail | |
1377 | */ | |
1378 | ls = map_num(r6layout, info->new_layout); | |
1379 | if (!ls || | |
1380 | strcmp(ls+strlen(ls)-2, "-6") != 0) | |
1381 | return "Please specify new layout"; | |
1382 | l = info->new_layout; | |
1383 | } | |
1384 | re->after.layout = l; | |
1385 | } | |
1386 | } | |
1387 | break; | |
1388 | ||
1389 | case 6: | |
1390 | /* We must already be at level 6 */ | |
1391 | if (re->level != 6) | |
1392 | return "Impossible level change"; | |
1393 | if (info->new_layout == UnSet) | |
1394 | re->after.layout = info->array.layout; | |
1395 | else | |
1396 | re->after.layout = info->new_layout; | |
1397 | break; | |
1398 | default: | |
1399 | return "Impossible level change requested"; | |
1400 | } | |
1401 | if (info->delta_disks == UnSet) | |
1402 | info->delta_disks = delta_parity; | |
1403 | ||
1404 | re->after.data_disks = (re->before.data_disks | |
1405 | + info->delta_disks | |
1406 | - delta_parity); | |
1407 | switch (re->level) { | |
1408 | case 6: re->parity = 2; | |
1409 | break; | |
1410 | case 4: | |
1411 | case 5: re->parity = 1; | |
1412 | break; | |
1413 | default: re->parity = 0; | |
1414 | break; | |
1415 | } | |
1416 | /* So we have a restripe operation, we need to calculate the number | |
1417 | * of blocks per reshape operation. | |
1418 | */ | |
1419 | re->new_size = info->component_size * re->before.data_disks; | |
1420 | if (info->new_chunk == 0) | |
1421 | info->new_chunk = info->array.chunk_size; | |
1422 | if (re->after.data_disks == re->before.data_disks && | |
1423 | re->after.layout == re->before.layout && | |
1424 | info->new_chunk == info->array.chunk_size) { | |
1425 | /* Nothing to change, can change level immediately. */ | |
1426 | re->level = info->new_level; | |
1427 | re->backup_blocks = 0; | |
1428 | return NULL; | |
1429 | } | |
1430 | if (re->after.data_disks == 1 && re->before.data_disks == 1) { | |
1431 | /* chunk and layout changes make no difference */ | |
1432 | re->level = info->new_level; | |
1433 | re->backup_blocks = 0; | |
1434 | return NULL; | |
1435 | } | |
1436 | ||
1437 | if (re->after.data_disks == re->before.data_disks && | |
1438 | get_linux_version() < 2006032) | |
1439 | return "in-place reshape is not safe before 2.6.32 - sorry."; | |
1440 | ||
1441 | if (re->after.data_disks < re->before.data_disks && | |
1442 | get_linux_version() < 2006030) | |
1443 | return "reshape to fewer devices is not supported before 2.6.30 - sorry."; | |
1444 | ||
1445 | re->backup_blocks = compute_backup_blocks( | |
1446 | info->new_chunk, info->array.chunk_size, | |
1447 | re->after.data_disks, | |
1448 | re->before.data_disks); | |
1449 | re->min_offset_change = re->backup_blocks / re->before.data_disks; | |
1450 | ||
1451 | re->new_size = info->component_size * re->after.data_disks; | |
1452 | return NULL; | |
1453 | } | |
1454 | ||
1455 | static int set_array_size(struct supertype *st, struct mdinfo *sra, | |
1456 | char *text_version) | |
1457 | { | |
1458 | struct mdinfo *info; | |
1459 | char *subarray; | |
1460 | int ret_val = -1; | |
1461 | ||
1462 | if ((st == NULL) || (sra == NULL)) | |
1463 | return ret_val; | |
1464 | ||
1465 | if (text_version == NULL) | |
1466 | text_version = sra->text_version; | |
1467 | subarray = strchr(text_version+1, '/')+1; | |
1468 | info = st->ss->container_content(st, subarray); | |
1469 | if (info) { | |
1470 | unsigned long long current_size = 0; | |
1471 | unsigned long long new_size = | |
1472 | info->custom_array_size/2; | |
1473 | ||
1474 | if (sysfs_get_ll(sra, NULL, "array_size", ¤t_size) == 0 && | |
1475 | new_size > current_size) { | |
1476 | if (sysfs_set_num(sra, NULL, "array_size", new_size) | |
1477 | < 0) | |
1478 | dprintf("Error: Cannot set array size"); | |
1479 | else { | |
1480 | ret_val = 0; | |
1481 | dprintf("Array size changed"); | |
1482 | } | |
1483 | dprintf(" from %llu to %llu.\n", | |
1484 | current_size, new_size); | |
1485 | } | |
1486 | sysfs_free(info); | |
1487 | } else | |
1488 | dprintf("Error: set_array_size(): info pointer in NULL\n"); | |
1489 | ||
1490 | return ret_val; | |
1491 | } | |
1492 | ||
1493 | static int reshape_array(char *container, int fd, char *devname, | |
1494 | struct supertype *st, struct mdinfo *info, | |
1495 | int force, struct mddev_dev *devlist, | |
1496 | unsigned long long data_offset, | |
1497 | char *backup_file, int verbose, int forked, | |
1498 | int restart, int freeze_reshape); | |
1499 | static int reshape_container(char *container, char *devname, | |
1500 | int mdfd, | |
1501 | struct supertype *st, | |
1502 | struct mdinfo *info, | |
1503 | int force, | |
1504 | char *backup_file, | |
1505 | int verbose, int restart, int freeze_reshape); | |
1506 | ||
1507 | int Grow_reshape(char *devname, int fd, | |
1508 | struct mddev_dev *devlist, | |
1509 | unsigned long long data_offset, | |
1510 | struct context *c, struct shape *s) | |
1511 | { | |
1512 | /* Make some changes in the shape of an array. | |
1513 | * The kernel must support the change. | |
1514 | * | |
1515 | * There are three different changes. Each can trigger | |
1516 | * a resync or recovery so we freeze that until we have | |
1517 | * requested everything (if kernel supports freezing - 2.6.30). | |
1518 | * The steps are: | |
1519 | * - change size (i.e. component_size) | |
1520 | * - change level | |
1521 | * - change layout/chunksize/ndisks | |
1522 | * | |
1523 | * The last can require a reshape. It is different on different | |
1524 | * levels so we need to check the level before actioning it. | |
1525 | * Some times the level change needs to be requested after the | |
1526 | * reshape (e.g. raid6->raid5, raid5->raid0) | |
1527 | * | |
1528 | */ | |
1529 | struct mdu_array_info_s array; | |
1530 | int rv = 0; | |
1531 | struct supertype *st; | |
1532 | char *subarray = NULL; | |
1533 | ||
1534 | int frozen; | |
1535 | int changed = 0; | |
1536 | char *container = NULL; | |
1537 | int cfd = -1; | |
1538 | ||
1539 | struct mddev_dev *dv; | |
1540 | int added_disks; | |
1541 | ||
1542 | struct mdinfo info; | |
1543 | struct mdinfo *sra; | |
1544 | ||
1545 | if (ioctl(fd, GET_ARRAY_INFO, &array) < 0) { | |
1546 | pr_err("%s is not an active md array - aborting\n", | |
1547 | devname); | |
1548 | return 1; | |
1549 | } | |
1550 | if (data_offset != INVALID_SECTORS && array.level != 10 | |
1551 | && (array.level < 4 || array.level > 6)) { | |
1552 | pr_err("--grow --data-offset not yet supported\n"); | |
1553 | return 1; | |
1554 | } | |
1555 | ||
1556 | if (s->size > 0 && | |
1557 | (s->chunk || s->level!= UnSet || s->layout_str || s->raiddisks)) { | |
1558 | pr_err("cannot change component size at the same time " | |
1559 | "as other changes.\n" | |
1560 | " Change size first, then check data is intact before " | |
1561 | "making other changes.\n"); | |
1562 | return 1; | |
1563 | } | |
1564 | ||
1565 | if (s->raiddisks && s->raiddisks < array.raid_disks && array.level > 1 && | |
1566 | get_linux_version() < 2006032 && | |
1567 | !check_env("MDADM_FORCE_FEWER")) { | |
1568 | pr_err("reducing the number of devices is not safe before Linux 2.6.32\n" | |
1569 | " Please use a newer kernel\n"); | |
1570 | return 1; | |
1571 | } | |
1572 | ||
1573 | st = super_by_fd(fd, &subarray); | |
1574 | if (!st) { | |
1575 | pr_err("Unable to determine metadata format for %s\n", devname); | |
1576 | return 1; | |
1577 | } | |
1578 | if (s->raiddisks > st->max_devs) { | |
1579 | pr_err("Cannot increase raid-disks on this array" | |
1580 | " beyond %d\n", st->max_devs); | |
1581 | return 1; | |
1582 | } | |
1583 | ||
1584 | /* in the external case we need to check that the requested reshape is | |
1585 | * supported, and perform an initial check that the container holds the | |
1586 | * pre-requisite spare devices (mdmon owns final validation) | |
1587 | */ | |
1588 | if (st->ss->external) { | |
1589 | int rv; | |
1590 | ||
1591 | if (subarray) { | |
1592 | container = st->container_devnm; | |
1593 | cfd = open_dev_excl(st->container_devnm); | |
1594 | } else { | |
1595 | container = st->devnm; | |
1596 | close(fd); | |
1597 | cfd = open_dev_excl(st->devnm); | |
1598 | fd = cfd; | |
1599 | } | |
1600 | if (cfd < 0) { | |
1601 | pr_err("Unable to open container for %s\n", | |
1602 | devname); | |
1603 | free(subarray); | |
1604 | return 1; | |
1605 | } | |
1606 | ||
1607 | rv = st->ss->load_container(st, cfd, NULL); | |
1608 | ||
1609 | if (rv) { | |
1610 | pr_err("Cannot read superblock for %s\n", | |
1611 | devname); | |
1612 | free(subarray); | |
1613 | return 1; | |
1614 | } | |
1615 | ||
1616 | /* check if operation is supported for metadata handler */ | |
1617 | if (st->ss->container_content) { | |
1618 | struct mdinfo *cc = NULL; | |
1619 | struct mdinfo *content = NULL; | |
1620 | ||
1621 | cc = st->ss->container_content(st, subarray); | |
1622 | for (content = cc; content ; content = content->next) { | |
1623 | int allow_reshape = 1; | |
1624 | ||
1625 | /* check if reshape is allowed based on metadata | |
1626 | * indications stored in content.array.status | |
1627 | */ | |
1628 | if (content->array.state & (1<<MD_SB_BLOCK_VOLUME)) | |
1629 | allow_reshape = 0; | |
1630 | if (content->array.state | |
1631 | & (1<<MD_SB_BLOCK_CONTAINER_RESHAPE)) | |
1632 | allow_reshape = 0; | |
1633 | if (!allow_reshape) { | |
1634 | pr_err("cannot reshape arrays in" | |
1635 | " container with unsupported" | |
1636 | " metadata: %s(%s)\n", | |
1637 | devname, container); | |
1638 | sysfs_free(cc); | |
1639 | free(subarray); | |
1640 | return 1; | |
1641 | } | |
1642 | } | |
1643 | sysfs_free(cc); | |
1644 | } | |
1645 | if (mdmon_running(container)) | |
1646 | st->update_tail = &st->updates; | |
1647 | } | |
1648 | ||
1649 | added_disks = 0; | |
1650 | for (dv = devlist; dv; dv = dv->next) | |
1651 | added_disks++; | |
1652 | if (s->raiddisks > array.raid_disks && | |
1653 | array.spare_disks +added_disks < (s->raiddisks - array.raid_disks) && | |
1654 | !c->force) { | |
1655 | pr_err("Need %d spare%s to avoid degraded array," | |
1656 | " and only have %d.\n" | |
1657 | " Use --force to over-ride this check.\n", | |
1658 | s->raiddisks - array.raid_disks, | |
1659 | s->raiddisks - array.raid_disks == 1 ? "" : "s", | |
1660 | array.spare_disks + added_disks); | |
1661 | return 1; | |
1662 | } | |
1663 | ||
1664 | sra = sysfs_read(fd, NULL, GET_LEVEL | GET_DISKS | GET_DEVS | |
1665 | | GET_STATE | GET_VERSION); | |
1666 | if (sra) { | |
1667 | if (st->ss->external && subarray == NULL) { | |
1668 | array.level = LEVEL_CONTAINER; | |
1669 | sra->array.level = LEVEL_CONTAINER; | |
1670 | } | |
1671 | } else { | |
1672 | pr_err("failed to read sysfs parameters for %s\n", | |
1673 | devname); | |
1674 | return 1; | |
1675 | } | |
1676 | frozen = freeze(st); | |
1677 | if (frozen < -1) { | |
1678 | /* freeze() already spewed the reason */ | |
1679 | sysfs_free(sra); | |
1680 | return 1; | |
1681 | } else if (frozen < 0) { | |
1682 | pr_err("%s is performing resync/recovery and cannot" | |
1683 | " be reshaped\n", devname); | |
1684 | sysfs_free(sra); | |
1685 | return 1; | |
1686 | } | |
1687 | ||
1688 | /* ========= set size =============== */ | |
1689 | if (s->size > 0 && (s->size == MAX_SIZE || s->size != (unsigned)array.size)) { | |
1690 | unsigned long long orig_size = get_component_size(fd)/2; | |
1691 | unsigned long long min_csize; | |
1692 | struct mdinfo *mdi; | |
1693 | int raid0_takeover = 0; | |
1694 | ||
1695 | if (orig_size == 0) | |
1696 | orig_size = (unsigned) array.size; | |
1697 | ||
1698 | if (orig_size == 0) { | |
1699 | pr_err("Cannot set device size in this type of array.\n"); | |
1700 | rv = 1; | |
1701 | goto release; | |
1702 | } | |
1703 | ||
1704 | if (reshape_super(st, s->size, UnSet, UnSet, 0, 0, UnSet, NULL, | |
1705 | devname, APPLY_METADATA_CHANGES, c->verbose > 0)) { | |
1706 | rv = 1; | |
1707 | goto release; | |
1708 | } | |
1709 | sync_metadata(st); | |
1710 | if (st->ss->external) { | |
1711 | /* metadata can have size limitation | |
1712 | * update size value according to metadata information | |
1713 | */ | |
1714 | struct mdinfo *sizeinfo = | |
1715 | st->ss->container_content(st, subarray); | |
1716 | if (sizeinfo) { | |
1717 | unsigned long long new_size = | |
1718 | sizeinfo->custom_array_size/2; | |
1719 | int data_disks = get_data_disks( | |
1720 | sizeinfo->array.level, | |
1721 | sizeinfo->array.layout, | |
1722 | sizeinfo->array.raid_disks); | |
1723 | new_size /= data_disks; | |
1724 | dprintf("Metadata size correction from %llu to " | |
1725 | "%llu (%llu)\n", orig_size, new_size, | |
1726 | new_size * data_disks); | |
1727 | s->size = new_size; | |
1728 | sysfs_free(sizeinfo); | |
1729 | } | |
1730 | } | |
1731 | ||
1732 | /* Update the size of each member device in case | |
1733 | * they have been resized. This will never reduce | |
1734 | * below the current used-size. The "size" attribute | |
1735 | * understands '0' to mean 'max'. | |
1736 | */ | |
1737 | min_csize = 0; | |
1738 | rv = 0; | |
1739 | for (mdi = sra->devs; mdi; mdi = mdi->next) { | |
1740 | if (sysfs_set_num(sra, mdi, "size", | |
1741 | s->size == MAX_SIZE ? 0 : s->size) < 0) { | |
1742 | /* Probably kernel refusing to let us | |
1743 | * reduce the size - not an error. | |
1744 | */ | |
1745 | break; | |
1746 | } | |
1747 | if (array.not_persistent == 0 && | |
1748 | array.major_version == 0 && | |
1749 | get_linux_version() < 3001000) { | |
1750 | /* Dangerous to allow size to exceed 2TB */ | |
1751 | unsigned long long csize; | |
1752 | if (sysfs_get_ll(sra, mdi, "size", &csize) == 0) { | |
1753 | if (csize >= 2ULL*1024*1024*1024) | |
1754 | csize = 2ULL*1024*1024*1024; | |
1755 | if ((min_csize == 0 || (min_csize | |
1756 | > csize))) | |
1757 | min_csize = csize; | |
1758 | } | |
1759 | } | |
1760 | } | |
1761 | if (rv) { | |
1762 | pr_err("Cannot set size on " | |
1763 | "array members.\n"); | |
1764 | goto size_change_error; | |
1765 | } | |
1766 | if (min_csize && s->size > min_csize) { | |
1767 | pr_err("Cannot safely make this array " | |
1768 | "use more than 2TB per device on this kernel.\n"); | |
1769 | rv = 1; | |
1770 | goto size_change_error; | |
1771 | } | |
1772 | if (min_csize && s->size == MAX_SIZE) { | |
1773 | /* Don't let the kernel choose a size - it will get | |
1774 | * it wrong | |
1775 | */ | |
1776 | pr_err("Limited v0.90 array to " | |
1777 | "2TB per device\n"); | |
1778 | s->size = min_csize; | |
1779 | } | |
1780 | if (st->ss->external) { | |
1781 | if (sra->array.level == 0) { | |
1782 | rv = sysfs_set_str(sra, NULL, "level", | |
1783 | "raid5"); | |
1784 | if (!rv) { | |
1785 | raid0_takeover = 1; | |
1786 | /* get array parametes after takeover | |
1787 | * to chane one parameter at time only | |
1788 | */ | |
1789 | rv = ioctl(fd, GET_ARRAY_INFO, &array); | |
1790 | } | |
1791 | } | |
1792 | /* make sure mdmon is | |
1793 | * aware of the new level */ | |
1794 | if (!mdmon_running(st->container_devnm)) | |
1795 | start_mdmon(st->container_devnm); | |
1796 | ping_monitor(container); | |
1797 | if (mdmon_running(st->container_devnm) && | |
1798 | st->update_tail == NULL) | |
1799 | st->update_tail = &st->updates; | |
1800 | } | |
1801 | ||
1802 | if (s->size == MAX_SIZE) | |
1803 | s->size = 0; | |
1804 | array.size = s->size; | |
1805 | if ((unsigned)array.size != s->size) { | |
1806 | /* got truncated to 32bit, write to | |
1807 | * component_size instead | |
1808 | */ | |
1809 | if (sra) | |
1810 | rv = sysfs_set_num(sra, NULL, | |
1811 | "component_size", s->size); | |
1812 | else | |
1813 | rv = -1; | |
1814 | } else { | |
1815 | rv = ioctl(fd, SET_ARRAY_INFO, &array); | |
1816 | ||
1817 | /* manage array size when it is managed externally | |
1818 | */ | |
1819 | if ((rv == 0) && st->ss->external) | |
1820 | rv = set_array_size(st, sra, sra->text_version); | |
1821 | } | |
1822 | ||
1823 | if (raid0_takeover) { | |
1824 | /* do not recync non-existing parity, | |
1825 | * we will drop it anyway | |
1826 | */ | |
1827 | sysfs_set_str(sra, NULL, "sync_action", "frozen"); | |
1828 | /* go back to raid0, drop parity disk | |
1829 | */ | |
1830 | sysfs_set_str(sra, NULL, "level", "raid0"); | |
1831 | ioctl(fd, GET_ARRAY_INFO, &array); | |
1832 | } | |
1833 | ||
1834 | size_change_error: | |
1835 | if (rv != 0) { | |
1836 | int err = errno; | |
1837 | ||
1838 | /* restore metadata */ | |
1839 | if (reshape_super(st, orig_size, UnSet, UnSet, 0, 0, | |
1840 | UnSet, NULL, devname, | |
1841 | ROLLBACK_METADATA_CHANGES, | |
1842 | c->verbose) == 0) | |
1843 | sync_metadata(st); | |
1844 | pr_err("Cannot set device size for %s: %s\n", | |
1845 | devname, strerror(err)); | |
1846 | if (err == EBUSY && | |
1847 | (array.state & (1<<MD_SB_BITMAP_PRESENT))) | |
1848 | cont_err("Bitmap must be removed before size can be changed\n"); | |
1849 | rv = 1; | |
1850 | goto release; | |
1851 | } | |
1852 | if (s->assume_clean) { | |
1853 | /* This will fail on kernels older than 3.0 unless | |
1854 | * a backport has been arranged. | |
1855 | */ | |
1856 | if (sra == NULL || | |
1857 | sysfs_set_str(sra, NULL, "resync_start", "none") < 0) | |
1858 | pr_err("--assume-clean not supported with --grow on this kernel\n"); | |
1859 | } | |
1860 | ioctl(fd, GET_ARRAY_INFO, &array); | |
1861 | s->size = get_component_size(fd)/2; | |
1862 | if (s->size == 0) | |
1863 | s->size = array.size; | |
1864 | if (c->verbose >= 0) { | |
1865 | if (s->size == orig_size) | |
1866 | pr_err("component size of %s " | |
1867 | "unchanged at %lluK\n", | |
1868 | devname, s->size); | |
1869 | else | |
1870 | pr_err("component size of %s " | |
1871 | "has been set to %lluK\n", | |
1872 | devname, s->size); | |
1873 | } | |
1874 | changed = 1; | |
1875 | } else if (array.level != LEVEL_CONTAINER) { | |
1876 | s->size = get_component_size(fd)/2; | |
1877 | if (s->size == 0) | |
1878 | s->size = array.size; | |
1879 | } | |
1880 | ||
1881 | /* See if there is anything else to do */ | |
1882 | if ((s->level == UnSet || s->level == array.level) && | |
1883 | (s->layout_str == NULL) && | |
1884 | (s->chunk == 0 || s->chunk == array.chunk_size) && | |
1885 | data_offset == INVALID_SECTORS && | |
1886 | (s->raiddisks == 0 || s->raiddisks == array.raid_disks)) { | |
1887 | /* Nothing more to do */ | |
1888 | if (!changed && c->verbose >= 0) | |
1889 | pr_err("%s: no change requested\n", | |
1890 | devname); | |
1891 | goto release; | |
1892 | } | |
1893 | ||
1894 | /* ========= check for Raid10/Raid1 -> Raid0 conversion =============== | |
1895 | * current implementation assumes that following conditions must be met: | |
1896 | * - RAID10: | |
1897 | * - far_copies == 1 | |
1898 | * - near_copies == 2 | |
1899 | */ | |
1900 | if ((s->level == 0 && array.level == 10 && sra && | |
1901 | array.layout == ((1 << 8) + 2) && !(array.raid_disks & 1)) || | |
1902 | (s->level == 0 && array.level == 1 && sra)) { | |
1903 | int err; | |
1904 | err = remove_disks_for_takeover(st, sra, array.layout); | |
1905 | if (err) { | |
1906 | dprintf(Name": Array cannot be reshaped\n"); | |
1907 | if (cfd > -1) | |
1908 | close(cfd); | |
1909 | rv = 1; | |
1910 | goto release; | |
1911 | } | |
1912 | /* Make sure mdmon has seen the device removal | |
1913 | * and updated metadata before we continue with | |
1914 | * level change | |
1915 | */ | |
1916 | if (container) | |
1917 | ping_monitor(container); | |
1918 | } | |
1919 | ||
1920 | memset(&info, 0, sizeof(info)); | |
1921 | info.array = array; | |
1922 | sysfs_init(&info, fd, NULL); | |
1923 | strcpy(info.text_version, sra->text_version); | |
1924 | info.component_size = s->size*2; | |
1925 | info.new_level = s->level; | |
1926 | info.new_chunk = s->chunk * 1024; | |
1927 | if (info.array.level == LEVEL_CONTAINER) { | |
1928 | info.delta_disks = UnSet; | |
1929 | info.array.raid_disks = s->raiddisks; | |
1930 | } else if (s->raiddisks) | |
1931 | info.delta_disks = s->raiddisks - info.array.raid_disks; | |
1932 | else | |
1933 | info.delta_disks = UnSet; | |
1934 | if (s->layout_str == NULL) { | |
1935 | info.new_layout = UnSet; | |
1936 | if (info.array.level == 6 && | |
1937 | (info.new_level == 6 || info.new_level == UnSet) && | |
1938 | info.array.layout >= 16) { | |
1939 | pr_err("%s has a non-standard layout. If you" | |
1940 | " wish to preserve this\n", devname); | |
1941 | cont_err("during the reshape, please specify" | |
1942 | " --layout=preserve\n"); | |
1943 | cont_err("If you want to change it, specify a" | |
1944 | " layout or use --layout=normalise\n"); | |
1945 | rv = 1; | |
1946 | goto release; | |
1947 | } | |
1948 | } else if (strcmp(s->layout_str, "normalise") == 0 || | |
1949 | strcmp(s->layout_str, "normalize") == 0) { | |
1950 | /* If we have a -6 RAID6 layout, remove the '-6'. */ | |
1951 | info.new_layout = UnSet; | |
1952 | if (info.array.level == 6 && info.new_level == UnSet) { | |
1953 | char l[40], *h; | |
1954 | strcpy(l, map_num(r6layout, info.array.layout)); | |
1955 | h = strrchr(l, '-'); | |
1956 | if (h && strcmp(h, "-6") == 0) { | |
1957 | *h = 0; | |
1958 | info.new_layout = map_name(r6layout, l); | |
1959 | } | |
1960 | } else { | |
1961 | pr_err("%s is only meaningful when reshaping" | |
1962 | " a RAID6 array.\n", s->layout_str); | |
1963 | rv = 1; | |
1964 | goto release; | |
1965 | } | |
1966 | } else if (strcmp(s->layout_str, "preserve") == 0) { | |
1967 | /* This means that a non-standard RAID6 layout | |
1968 | * is OK. | |
1969 | * In particular: | |
1970 | * - When reshape a RAID6 (e.g. adding a device) | |
1971 | * which is in a non-standard layout, it is OK | |
1972 | * to preserve that layout. | |
1973 | * - When converting a RAID5 to RAID6, leave it in | |
1974 | * the XXX-6 layout, don't re-layout. | |
1975 | */ | |
1976 | if (info.array.level == 6 && info.new_level == UnSet) | |
1977 | info.new_layout = info.array.layout; | |
1978 | else if (info.array.level == 5 && info.new_level == 6) { | |
1979 | char l[40]; | |
1980 | strcpy(l, map_num(r5layout, info.array.layout)); | |
1981 | strcat(l, "-6"); | |
1982 | info.new_layout = map_name(r6layout, l); | |
1983 | } else { | |
1984 | pr_err("%s in only meaningful when reshaping" | |
1985 | " to RAID6\n", s->layout_str); | |
1986 | rv = 1; | |
1987 | goto release; | |
1988 | } | |
1989 | } else { | |
1990 | int l = info.new_level; | |
1991 | if (l == UnSet) | |
1992 | l = info.array.level; | |
1993 | switch (l) { | |
1994 | case 5: | |
1995 | info.new_layout = map_name(r5layout, s->layout_str); | |
1996 | break; | |
1997 | case 6: | |
1998 | info.new_layout = map_name(r6layout, s->layout_str); | |
1999 | break; | |
2000 | case 10: | |
2001 | info.new_layout = parse_layout_10(s->layout_str); | |
2002 | break; | |
2003 | case LEVEL_FAULTY: | |
2004 | info.new_layout = parse_layout_faulty(s->layout_str); | |
2005 | break; | |
2006 | default: | |
2007 | pr_err("layout not meaningful" | |
2008 | " with this level\n"); | |
2009 | rv = 1; | |
2010 | goto release; | |
2011 | } | |
2012 | if (info.new_layout == UnSet) { | |
2013 | pr_err("layout %s not understood" | |
2014 | " for this level\n", | |
2015 | s->layout_str); | |
2016 | rv = 1; | |
2017 | goto release; | |
2018 | } | |
2019 | } | |
2020 | ||
2021 | if (array.level == LEVEL_FAULTY) { | |
2022 | if (s->level != UnSet && s->level != array.level) { | |
2023 | pr_err("cannot change level of Faulty device\n"); | |
2024 | rv =1 ; | |
2025 | } | |
2026 | if (s->chunk) { | |
2027 | pr_err("cannot set chunksize of Faulty device\n"); | |
2028 | rv =1 ; | |
2029 | } | |
2030 | if (s->raiddisks && s->raiddisks != 1) { | |
2031 | pr_err("cannot set raid_disks of Faulty device\n"); | |
2032 | rv =1 ; | |
2033 | } | |
2034 | if (s->layout_str) { | |
2035 | if (ioctl(fd, GET_ARRAY_INFO, &array) != 0) { | |
2036 | dprintf("Cannot get array information.\n"); | |
2037 | goto release; | |
2038 | } | |
2039 | array.layout = info.new_layout; | |
2040 | if (ioctl(fd, SET_ARRAY_INFO, &array) != 0) { | |
2041 | pr_err("failed to set new layout\n"); | |
2042 | rv = 1; | |
2043 | } else if (c->verbose >= 0) | |
2044 | printf("layout for %s set to %d\n", | |
2045 | devname, array.layout); | |
2046 | } | |
2047 | } else if (array.level == LEVEL_CONTAINER) { | |
2048 | /* This change is to be applied to every array in the | |
2049 | * container. This is only needed when the metadata imposes | |
2050 | * restraints of the various arrays in the container. | |
2051 | * Currently we only know that IMSM requires all arrays | |
2052 | * to have the same number of devices so changing the | |
2053 | * number of devices (On-Line Capacity Expansion) must be | |
2054 | * performed at the level of the container | |
2055 | */ | |
2056 | rv = reshape_container(container, devname, -1, st, &info, | |
2057 | c->force, c->backup_file, c->verbose, 0, 0); | |
2058 | frozen = 0; | |
2059 | } else { | |
2060 | /* get spare devices from external metadata | |
2061 | */ | |
2062 | if (st->ss->external) { | |
2063 | struct mdinfo *info2; | |
2064 | ||
2065 | info2 = st->ss->container_content(st, subarray); | |
2066 | if (info2) { | |
2067 | info.array.spare_disks = | |
2068 | info2->array.spare_disks; | |
2069 | sysfs_free(info2); | |
2070 | } | |
2071 | } | |
2072 | ||
2073 | /* Impose these changes on a single array. First | |
2074 | * check that the metadata is OK with the change. */ | |
2075 | ||
2076 | if (reshape_super(st, 0, info.new_level, | |
2077 | info.new_layout, info.new_chunk, | |
2078 | info.array.raid_disks, info.delta_disks, | |
2079 | c->backup_file, devname, APPLY_METADATA_CHANGES, | |
2080 | c->verbose)) { | |
2081 | rv = 1; | |
2082 | goto release; | |
2083 | } | |
2084 | sync_metadata(st); | |
2085 | rv = reshape_array(container, fd, devname, st, &info, c->force, | |
2086 | devlist, data_offset, c->backup_file, c->verbose, | |
2087 | 0, 0, 0); | |
2088 | frozen = 0; | |
2089 | } | |
2090 | release: | |
2091 | sysfs_free(sra); | |
2092 | if (frozen > 0) | |
2093 | unfreeze(st); | |
2094 | return rv; | |
2095 | } | |
2096 | ||
2097 | /* verify_reshape_position() | |
2098 | * Function checks if reshape position in metadata is not farther | |
2099 | * than position in md. | |
2100 | * Return value: | |
2101 | * 0 : not valid sysfs entry | |
2102 | * it can be caused by not started reshape, it should be started | |
2103 | * by reshape array or raid0 array is before takeover | |
2104 | * -1 : error, reshape position is obviously wrong | |
2105 | * 1 : success, reshape progress correct or updated | |
2106 | */ | |
2107 | static int verify_reshape_position(struct mdinfo *info, int level) | |
2108 | { | |
2109 | int ret_val = 0; | |
2110 | char buf[40]; | |
2111 | int rv; | |
2112 | ||
2113 | /* read sync_max, failure can mean raid0 array */ | |
2114 | rv = sysfs_get_str(info, NULL, "sync_max", buf, 40); | |
2115 | ||
2116 | if (rv > 0) { | |
2117 | char *ep; | |
2118 | unsigned long long position = strtoull(buf, &ep, 0); | |
2119 | ||
2120 | dprintf(Name": Read sync_max sysfs entry is: %s\n", buf); | |
2121 | if (!(ep == buf || (*ep != 0 && *ep != '\n' && *ep != ' '))) { | |
2122 | position *= get_data_disks(level, | |
2123 | info->new_layout, | |
2124 | info->array.raid_disks); | |
2125 | if (info->reshape_progress < position) { | |
2126 | dprintf("Corrected reshape progress (%llu) to " | |
2127 | "md position (%llu)\n", | |
2128 | info->reshape_progress, position); | |
2129 | info->reshape_progress = position; | |
2130 | ret_val = 1; | |
2131 | } else if (info->reshape_progress > position) { | |
2132 | pr_err("Fatal error: array " | |
2133 | "reshape was not properly frozen " | |
2134 | "(expected reshape position is %llu, " | |
2135 | "but reshape progress is %llu.\n", | |
2136 | position, info->reshape_progress); | |
2137 | ret_val = -1; | |
2138 | } else { | |
2139 | dprintf("Reshape position in md and metadata " | |
2140 | "are the same;"); | |
2141 | ret_val = 1; | |
2142 | } | |
2143 | } | |
2144 | } else if (rv == 0) { | |
2145 | /* for valid sysfs entry, 0-length content | |
2146 | * should be indicated as error | |
2147 | */ | |
2148 | ret_val = -1; | |
2149 | } | |
2150 | ||
2151 | return ret_val; | |
2152 | } | |
2153 | ||
2154 | static unsigned long long choose_offset(unsigned long long lo, | |
2155 | unsigned long long hi, | |
2156 | unsigned long long min, | |
2157 | unsigned long long max) | |
2158 | { | |
2159 | /* Choose a new offset between hi and lo. | |
2160 | * It must be between min and max, but | |
2161 | * we would prefer something near the middle of hi/lo, and also | |
2162 | * prefer to be aligned to a big power of 2. | |
2163 | * | |
2164 | * So we start with the middle, then for each bit, | |
2165 | * starting at '1' and increasing, if it is set, we either | |
2166 | * add it or subtract it if possible, preferring the option | |
2167 | * which is furthest from the boundary. | |
2168 | * | |
2169 | * We stop once we get a 1MB alignment. As units are in sectors, | |
2170 | * 1MB = 2*1024 sectors. | |
2171 | */ | |
2172 | unsigned long long choice = (lo + hi) / 2; | |
2173 | unsigned long long bit = 1; | |
2174 | ||
2175 | for (bit = 1; bit < 2*1024; bit = bit << 1) { | |
2176 | unsigned long long bigger, smaller; | |
2177 | if (! (bit & choice)) | |
2178 | continue; | |
2179 | bigger = choice + bit; | |
2180 | smaller = choice - bit; | |
2181 | if (bigger > max && smaller < min) | |
2182 | break; | |
2183 | if (bigger > max) | |
2184 | choice = smaller; | |
2185 | else if (smaller < min) | |
2186 | choice = bigger; | |
2187 | else if (hi - bigger > smaller - lo) | |
2188 | choice = bigger; | |
2189 | else | |
2190 | choice = smaller; | |
2191 | } | |
2192 | return choice; | |
2193 | } | |
2194 | ||
2195 | static int set_new_data_offset(struct mdinfo *sra, struct supertype *st, | |
2196 | char *devname, int delta_disks, | |
2197 | unsigned long long data_offset, | |
2198 | unsigned long long min) | |
2199 | { | |
2200 | struct mdinfo *sd; | |
2201 | int dir = 0; | |
2202 | int err = 0; | |
2203 | unsigned long long before, after; | |
2204 | ||
2205 | /* Need to find min space before and after so same is used | |
2206 | * on all devices | |
2207 | */ | |
2208 | before = UINT64_MAX; | |
2209 | after = UINT64_MAX; | |
2210 | for (sd = sra->devs; sd; sd = sd->next) { | |
2211 | char *dn; | |
2212 | int dfd; | |
2213 | int rv; | |
2214 | struct supertype *st2; | |
2215 | struct mdinfo info2; | |
2216 | ||
2217 | if (sd->disk.state & (1<<MD_DISK_FAULTY)) | |
2218 | continue; | |
2219 | dn = map_dev(sd->disk.major, sd->disk.minor, 0); | |
2220 | dfd = dev_open(dn, O_RDONLY); | |
2221 | if (dfd < 0) { | |
2222 | pr_err("%s: cannot open component %s\n", | |
2223 | devname, dn ? dn : "-unknown-"); | |
2224 | goto release; | |
2225 | } | |
2226 | st2 = dup_super(st); | |
2227 | rv = st2->ss->load_super(st2,dfd, NULL); | |
2228 | close(dfd); | |
2229 | if (rv) { | |
2230 | free(st2); | |
2231 | pr_err("%s: cannot get superblock from %s\n", | |
2232 | devname, dn); | |
2233 | goto release; | |
2234 | } | |
2235 | st2->ss->getinfo_super(st2, &info2, NULL); | |
2236 | st2->ss->free_super(st2); | |
2237 | free(st2); | |
2238 | if (info2.space_before == 0 && | |
2239 | info2.space_after == 0) { | |
2240 | /* Metadata doesn't support data_offset changes */ | |
2241 | return 1; | |
2242 | } | |
2243 | if (before > info2.space_before) | |
2244 | before = info2.space_before; | |
2245 | if (after > info2.space_after) | |
2246 | after = info2.space_after; | |
2247 | ||
2248 | if (data_offset != INVALID_SECTORS) { | |
2249 | if (dir == 0) { | |
2250 | if (info2.data_offset == data_offset) { | |
2251 | pr_err("%s: already has that data_offset\n", | |
2252 | dn); | |
2253 | goto release; | |
2254 | } | |
2255 | if (data_offset < info2.data_offset) | |
2256 | dir = -1; | |
2257 | else | |
2258 | dir = 1; | |
2259 | } else if ((data_offset <= info2.data_offset && dir == 1) || | |
2260 | (data_offset >= info2.data_offset && dir == -1)) { | |
2261 | pr_err("%s: differing data offsets on devices make this --data-offset setting impossible\n", | |
2262 | dn); | |
2263 | goto release; | |
2264 | } | |
2265 | } | |
2266 | } | |
2267 | if (before == UINT64_MAX) | |
2268 | /* impossible really, there must be no devices */ | |
2269 | return 1; | |
2270 | ||
2271 | for (sd = sra->devs; sd; sd = sd->next) { | |
2272 | char *dn = map_dev(sd->disk.major, sd->disk.minor, 0); | |
2273 | unsigned long long new_data_offset; | |
2274 | ||
2275 | if (sd->disk.state & (1<<MD_DISK_FAULTY)) | |
2276 | continue; | |
2277 | if (delta_disks < 0) { | |
2278 | /* Don't need any space as array is shrinking | |
2279 | * just move data_offset up by min | |
2280 | */ | |
2281 | if (data_offset == INVALID_SECTORS) | |
2282 | new_data_offset = sd->data_offset + min; | |
2283 | else { | |
2284 | if (data_offset < sd->data_offset + min) { | |
2285 | pr_err("--data-offset too small for %s\n", | |
2286 | dn); | |
2287 | goto release; | |
2288 | } | |
2289 | new_data_offset = data_offset; | |
2290 | } | |
2291 | } else if (delta_disks > 0) { | |
2292 | /* need space before */ | |
2293 | if (before < min) { | |
2294 | pr_err("Insufficient head-space for reshape on %s\n", | |
2295 | dn); | |
2296 | goto release; | |
2297 | } | |
2298 | if (data_offset == INVALID_SECTORS) | |
2299 | new_data_offset = sd->data_offset - min; | |
2300 | else { | |
2301 | if (data_offset > sd->data_offset - min) { | |
2302 | pr_err("--data-offset too large for %s\n", | |
2303 | dn); | |
2304 | goto release; | |
2305 | } | |
2306 | new_data_offset = data_offset; | |
2307 | } | |
2308 | } else { | |
2309 | if (dir == 0) { | |
2310 | /* can move up or down. If 'data_offset' | |
2311 | * was set we would have already decided, | |
2312 | * so just choose direction with most space. | |
2313 | */ | |
2314 | if (before > after) | |
2315 | dir = -1; | |
2316 | else | |
2317 | dir = 1; | |
2318 | } | |
2319 | sysfs_set_str(sra, NULL, "reshape_direction", | |
2320 | dir == 1 ? "backwards" : "forwards"); | |
2321 | if (dir > 0) { | |
2322 | /* Increase data offset */ | |
2323 | if (after < min) { | |
2324 | pr_err("Insufficient tail-space for reshape on %s\n", | |
2325 | dn); | |
2326 | goto release; | |
2327 | } | |
2328 | if (data_offset != INVALID_SECTORS && | |
2329 | data_offset < sd->data_offset + min) { | |
2330 | pr_err("--data-offset too small on %s\n", | |
2331 | dn); | |
2332 | goto release; | |
2333 | } | |
2334 | if (data_offset != INVALID_SECTORS) | |
2335 | new_data_offset = data_offset; | |
2336 | else | |
2337 | new_data_offset = choose_offset(sd->data_offset, | |
2338 | sd->data_offset + after, | |
2339 | sd->data_offset + min, | |
2340 | sd->data_offset + after); | |
2341 | } else { | |
2342 | /* Decrease data offset */ | |
2343 | if (before < min) { | |
2344 | pr_err("insufficient head-room on %s\n", | |
2345 | dn); | |
2346 | goto release; | |
2347 | } | |
2348 | if (data_offset != INVALID_SECTORS && | |
2349 | data_offset < sd->data_offset - min) { | |
2350 | pr_err("--data-offset too small on %s\n", | |
2351 | dn); | |
2352 | goto release; | |
2353 | } | |
2354 | if (data_offset != INVALID_SECTORS) | |
2355 | new_data_offset = data_offset; | |
2356 | else | |
2357 | new_data_offset = choose_offset(sd->data_offset - before, | |
2358 | sd->data_offset, | |
2359 | sd->data_offset - before, | |
2360 | sd->data_offset - min); | |
2361 | } | |
2362 | } | |
2363 | err = sysfs_set_num(sra, sd, "new_offset", new_data_offset); | |
2364 | if (err < 0 && errno == E2BIG) { | |
2365 | /* try again after increasing data size to max */ | |
2366 | err = sysfs_set_num(sra, sd, "size", 0); | |
2367 | if (err < 0 && errno == EINVAL && | |
2368 | !(sd->disk.state & (1<<MD_DISK_SYNC))) { | |
2369 | /* some kernels have a bug where you cannot | |
2370 | * use '0' on spare devices. */ | |
2371 | sysfs_set_num(sra, sd, "size", | |
2372 | (sra->component_size + after)/2); | |
2373 | } | |
2374 | err = sysfs_set_num(sra, sd, "new_offset", | |
2375 | new_data_offset); | |
2376 | } | |
2377 | if (err < 0) { | |
2378 | if (errno == E2BIG && data_offset != INVALID_SECTORS) { | |
2379 | pr_err("data-offset is too big for %s\n", | |
2380 | dn); | |
2381 | goto release; | |
2382 | } | |
2383 | if (sd == sra->devs && | |
2384 | (errno == ENOENT || errno == E2BIG)) | |
2385 | /* Early kernel, no 'new_offset' file, | |
2386 | * or kernel doesn't like us. | |
2387 | * For RAID5/6 this is not fatal | |
2388 | */ | |
2389 | return 1; | |
2390 | pr_err("Cannot set new_offset for %s\n", | |
2391 | dn); | |
2392 | break; | |
2393 | } | |
2394 | } | |
2395 | return err; | |
2396 | release: | |
2397 | return -1; | |
2398 | } | |
2399 | ||
2400 | static int raid10_reshape(char *container, int fd, char *devname, | |
2401 | struct supertype *st, struct mdinfo *info, | |
2402 | struct reshape *reshape, | |
2403 | unsigned long long data_offset, | |
2404 | int force, int verbose) | |
2405 | { | |
2406 | /* Changing raid_disks, layout, chunksize or possibly | |
2407 | * just data_offset for a RAID10. | |
2408 | * We must always change data_offset. We change by at least | |
2409 | * ->min_offset_change which is the largest of the old and new | |
2410 | * chunk sizes. | |
2411 | * If raid_disks is increasing, then data_offset must decrease | |
2412 | * by at least this copy size. | |
2413 | * If raid_disks is unchanged, data_offset must increase or | |
2414 | * decrease by at least min_offset_change but preferably by much more. | |
2415 | * We choose half of the available space. | |
2416 | * If raid_disks is decreasing, data_offset must increase by | |
2417 | * at least min_offset_change. To allow of this, component_size | |
2418 | * must be decreased by the same amount. | |
2419 | * | |
2420 | * So we calculate the required minimum and direction, possibly | |
2421 | * reduce the component_size, then iterate through the devices | |
2422 | * and set the new_data_offset. | |
2423 | * If that all works, we set chunk_size, layout, raid_disks, and start | |
2424 | * 'reshape' | |
2425 | */ | |
2426 | struct mdinfo *sra; | |
2427 | unsigned long long min; | |
2428 | int err = 0; | |
2429 | ||
2430 | sra = sysfs_read(fd, NULL, | |
2431 | GET_COMPONENT|GET_DEVS|GET_OFFSET|GET_STATE|GET_CHUNK | |
2432 | ); | |
2433 | if (!sra) { | |
2434 | pr_err("%s: Cannot get array details from sysfs\n", | |
2435 | devname); | |
2436 | goto release; | |
2437 | } | |
2438 | min = reshape->min_offset_change; | |
2439 | ||
2440 | if (info->delta_disks) | |
2441 | sysfs_set_str(sra, NULL, "reshape_direction", | |
2442 | info->delta_disks < 0 ? "backwards" : "forwards"); | |
2443 | if (info->delta_disks < 0 && | |
2444 | info->space_after < min) { | |
2445 | int rv = sysfs_set_num(sra, NULL, "component_size", | |
2446 | (sra->component_size - | |
2447 | min)/2); | |
2448 | if (rv) { | |
2449 | pr_err("cannot reduce component size\n"); | |
2450 | goto release; | |
2451 | } | |
2452 | } | |
2453 | err = set_new_data_offset(sra, st, devname, info->delta_disks, data_offset, | |
2454 | min); | |
2455 | if (err == 1) { | |
2456 | pr_err("Cannot set new_data_offset: RAID10 reshape not\n"); | |
2457 | cont_err("supported on this kernel\n"); | |
2458 | err = -1; | |
2459 | } | |
2460 | if (err < 0) | |
2461 | goto release; | |
2462 | ||
2463 | if (!err && sysfs_set_num(sra, NULL, "chunk_size", info->new_chunk) < 0) | |
2464 | err = errno; | |
2465 | if (!err && sysfs_set_num(sra, NULL, "layout", reshape->after.layout) < 0) | |
2466 | err = errno; | |
2467 | if (!err && sysfs_set_num(sra, NULL, "raid_disks", | |
2468 | info->array.raid_disks + info->delta_disks) < 0) | |
2469 | err = errno; | |
2470 | if (!err && sysfs_set_str(sra, NULL, "sync_action", "reshape") < 0) | |
2471 | err = errno; | |
2472 | if (err) { | |
2473 | pr_err("Cannot set array shape for %s\n", | |
2474 | devname); | |
2475 | if (err == EBUSY && | |
2476 | (info->array.state & (1<<MD_SB_BITMAP_PRESENT))) | |
2477 | cont_err(" Bitmap must be removed before" | |
2478 | " shape can be changed\n"); | |
2479 | goto release; | |
2480 | } | |
2481 | sysfs_free(sra); | |
2482 | return 0; | |
2483 | release: | |
2484 | sysfs_free(sra); | |
2485 | return 1; | |
2486 | } | |
2487 | ||
2488 | static void get_space_after(int fd, struct supertype *st, struct mdinfo *info) | |
2489 | { | |
2490 | struct mdinfo *sra, *sd; | |
2491 | /* Initialisation to silence compiler warning */ | |
2492 | unsigned long long min_space_before = 0, min_space_after = 0; | |
2493 | int first = 1; | |
2494 | ||
2495 | sra = sysfs_read(fd, NULL, GET_DEVS); | |
2496 | if (!sra) | |
2497 | return; | |
2498 | for (sd = sra->devs; sd; sd = sd->next) { | |
2499 | char *dn; | |
2500 | int dfd; | |
2501 | struct supertype *st2; | |
2502 | struct mdinfo info2; | |
2503 | ||
2504 | if (sd->disk.state & (1<<MD_DISK_FAULTY)) | |
2505 | continue; | |
2506 | dn = map_dev(sd->disk.major, sd->disk.minor, 0); | |
2507 | dfd = dev_open(dn, O_RDONLY); | |
2508 | if (dfd < 0) | |
2509 | break; | |
2510 | st2 = dup_super(st); | |
2511 | if (st2->ss->load_super(st2,dfd, NULL)) { | |
2512 | close(dfd); | |
2513 | free(st2); | |
2514 | break; | |
2515 | } | |
2516 | close(dfd); | |
2517 | st2->ss->getinfo_super(st2, &info2, NULL); | |
2518 | st2->ss->free_super(st2); | |
2519 | free(st2); | |
2520 | if (first || | |
2521 | min_space_before > info2.space_before) | |
2522 | min_space_before = info2.space_before; | |
2523 | if (first || | |
2524 | min_space_after > info2.space_after) | |
2525 | min_space_after = info2.space_after; | |
2526 | first = 0; | |
2527 | } | |
2528 | if (sd == NULL && !first) { | |
2529 | info->space_after = min_space_after; | |
2530 | info->space_before = min_space_before; | |
2531 | } | |
2532 | sysfs_free(sra); | |
2533 | } | |
2534 | ||
2535 | static void update_cache_size(char *container, struct mdinfo *sra, | |
2536 | struct mdinfo *info, | |
2537 | int disks, unsigned long long blocks) | |
2538 | { | |
2539 | /* Check that the internal stripe cache is | |
2540 | * large enough, or it won't work. | |
2541 | * It must hold at least 4 stripes of the larger | |
2542 | * chunk size | |
2543 | */ | |
2544 | unsigned long cache; | |
2545 | cache = max(info->array.chunk_size, info->new_chunk); | |
2546 | cache *= 4; /* 4 stripes minimum */ | |
2547 | cache /= 512; /* convert to sectors */ | |
2548 | /* make sure there is room for 'blocks' with a bit to spare */ | |
2549 | if (cache < 16 + blocks / disks) | |
2550 | cache = 16 + blocks / disks; | |
2551 | cache /= (4096/512); /* Covert from sectors to pages */ | |
2552 | ||
2553 | if (sra->cache_size < cache) | |
2554 | subarray_set_num(container, sra, "stripe_cache_size", | |
2555 | cache+1); | |
2556 | } | |
2557 | ||
2558 | static int impose_reshape(struct mdinfo *sra, | |
2559 | struct mdinfo *info, | |
2560 | struct supertype *st, | |
2561 | int fd, | |
2562 | int restart, | |
2563 | char *devname, char *container, | |
2564 | struct reshape *reshape) | |
2565 | { | |
2566 | struct mdu_array_info_s array; | |
2567 | ||
2568 | sra->new_chunk = info->new_chunk; | |
2569 | ||
2570 | if (restart) { | |
2571 | /* for external metadata checkpoint saved by mdmon can be lost | |
2572 | * or missed /due to e.g. crash/. Check if md is not during | |
2573 | * restart farther than metadata points to. | |
2574 | * If so, this means metadata information is obsolete. | |
2575 | */ | |
2576 | if (st->ss->external) | |
2577 | verify_reshape_position(info, reshape->level); | |
2578 | sra->reshape_progress = info->reshape_progress; | |
2579 | } else { | |
2580 | sra->reshape_progress = 0; | |
2581 | if (reshape->after.data_disks < reshape->before.data_disks) | |
2582 | /* start from the end of the new array */ | |
2583 | sra->reshape_progress = (sra->component_size | |
2584 | * reshape->after.data_disks); | |
2585 | } | |
2586 | ||
2587 | ioctl(fd, GET_ARRAY_INFO, &array); | |
2588 | if (info->array.chunk_size == info->new_chunk && | |
2589 | reshape->before.layout == reshape->after.layout && | |
2590 | st->ss->external == 0) { | |
2591 | /* use SET_ARRAY_INFO but only if reshape hasn't started */ | |
2592 | array.raid_disks = reshape->after.data_disks + reshape->parity; | |
2593 | if (!restart && | |
2594 | ioctl(fd, SET_ARRAY_INFO, &array) != 0) { | |
2595 | int err = errno; | |
2596 | ||
2597 | pr_err("Cannot set device shape for %s: %s\n", | |
2598 | devname, strerror(errno)); | |
2599 | ||
2600 | if (err == EBUSY && | |
2601 | (array.state & (1<<MD_SB_BITMAP_PRESENT))) | |
2602 | cont_err("Bitmap must be removed before" | |
2603 | " shape can be changed\n"); | |
2604 | ||
2605 | goto release; | |
2606 | } | |
2607 | } else if (!restart) { | |
2608 | /* set them all just in case some old 'new_*' value | |
2609 | * persists from some earlier problem. | |
2610 | */ | |
2611 | int err = 0; | |
2612 | if (sysfs_set_num(sra, NULL, "chunk_size", info->new_chunk) < 0) | |
2613 | err = errno; | |
2614 | if (!err && sysfs_set_num(sra, NULL, "layout", | |
2615 | reshape->after.layout) < 0) | |
2616 | err = errno; | |
2617 | if (!err && subarray_set_num(container, sra, "raid_disks", | |
2618 | reshape->after.data_disks + | |
2619 | reshape->parity) < 0) | |
2620 | err = errno; | |
2621 | if (err) { | |
2622 | pr_err("Cannot set device shape for %s\n", | |
2623 | devname); | |
2624 | ||
2625 | if (err == EBUSY && | |
2626 | (array.state & (1<<MD_SB_BITMAP_PRESENT))) | |
2627 | cont_err("Bitmap must be removed before" | |
2628 | " shape can be changed\n"); | |
2629 | goto release; | |
2630 | } | |
2631 | } | |
2632 | return 0; | |
2633 | release: | |
2634 | return -1; | |
2635 | } | |
2636 | ||
2637 | static int impose_level(int fd, int level, char *devname, int verbose) | |
2638 | { | |
2639 | char *c; | |
2640 | struct mdu_array_info_s array; | |
2641 | struct mdinfo info; | |
2642 | sysfs_init(&info, fd, NULL); | |
2643 | ||
2644 | ioctl(fd, GET_ARRAY_INFO, &array); | |
2645 | if (level == 0 && | |
2646 | (array.level >= 4 && array.level <= 6)) { | |
2647 | /* To convert to RAID0 we need to fail and | |
2648 | * remove any non-data devices. */ | |
2649 | int found = 0; | |
2650 | int d; | |
2651 | int data_disks = array.raid_disks - 1; | |
2652 | if (array.level == 6) | |
2653 | data_disks -= 1; | |
2654 | if (array.level == 5 && | |
2655 | array.layout != ALGORITHM_PARITY_N) | |
2656 | return -1; | |
2657 | if (array.level == 6 && | |
2658 | array.layout != ALGORITHM_PARITY_N_6) | |
2659 | return -1; | |
2660 | sysfs_set_str(&info, NULL,"sync_action", "idle"); | |
2661 | /* First remove any spares so no recovery starts */ | |
2662 | for (d = 0, found = 0; | |
2663 | d < MAX_DISKS && found < array.nr_disks; | |
2664 | d++) { | |
2665 | mdu_disk_info_t disk; | |
2666 | disk.number = d; | |
2667 | if (ioctl(fd, GET_DISK_INFO, &disk) < 0) | |
2668 | continue; | |
2669 | if (disk.major == 0 && disk.minor == 0) | |
2670 | continue; | |
2671 | found++; | |
2672 | if ((disk.state & (1 << MD_DISK_ACTIVE)) | |
2673 | && disk.raid_disk < data_disks) | |
2674 | /* keep this */ | |
2675 | continue; | |
2676 | ioctl(fd, HOT_REMOVE_DISK, | |
2677 | makedev(disk.major, disk.minor)); | |
2678 | } | |
2679 | /* Now fail anything left */ | |
2680 | ioctl(fd, GET_ARRAY_INFO, &array); | |
2681 | for (d = 0, found = 0; | |
2682 | d < MAX_DISKS && found < array.nr_disks; | |
2683 | d++) { | |
2684 | int cnt; | |
2685 | mdu_disk_info_t disk; | |
2686 | disk.number = d; | |
2687 | if (ioctl(fd, GET_DISK_INFO, &disk) < 0) | |
2688 | continue; | |
2689 | if (disk.major == 0 && disk.minor == 0) | |
2690 | continue; | |
2691 | found++; | |
2692 | if ((disk.state & (1 << MD_DISK_ACTIVE)) | |
2693 | && disk.raid_disk < data_disks) | |
2694 | /* keep this */ | |
2695 | continue; | |
2696 | ioctl(fd, SET_DISK_FAULTY, | |
2697 | makedev(disk.major, disk.minor)); | |
2698 | cnt = 5; | |
2699 | while (ioctl(fd, HOT_REMOVE_DISK, | |
2700 | makedev(disk.major, disk.minor)) < 0 | |
2701 | && errno == EBUSY | |
2702 | && cnt--) { | |
2703 | usleep(10000); | |
2704 | } | |
2705 | } | |
2706 | } | |
2707 | c = map_num(pers, level); | |
2708 | if (c) { | |
2709 | int err = sysfs_set_str(&info, NULL, "level", c); | |
2710 | if (err) { | |
2711 | err = errno; | |
2712 | pr_err("%s: could not set level to %s\n", | |
2713 | devname, c); | |
2714 | if (err == EBUSY && | |
2715 | (array.state & (1<<MD_SB_BITMAP_PRESENT))) | |
2716 | cont_err("Bitmap must be removed" | |
2717 | " before level can be changed\n"); | |
2718 | return err; | |
2719 | } | |
2720 | if (verbose >= 0) | |
2721 | pr_err("level of %s changed to %s\n", | |
2722 | devname, c); | |
2723 | } | |
2724 | return 0; | |
2725 | } | |
2726 | ||
2727 | static int reshape_array(char *container, int fd, char *devname, | |
2728 | struct supertype *st, struct mdinfo *info, | |
2729 | int force, struct mddev_dev *devlist, | |
2730 | unsigned long long data_offset, | |
2731 | char *backup_file, int verbose, int forked, | |
2732 | int restart, int freeze_reshape) | |
2733 | { | |
2734 | struct reshape reshape; | |
2735 | int spares_needed; | |
2736 | char *msg; | |
2737 | int orig_level = UnSet; | |
2738 | int odisks; | |
2739 | int delayed; | |
2740 | ||
2741 | struct mdu_array_info_s array; | |
2742 | char *c; | |
2743 | ||
2744 | struct mddev_dev *dv; | |
2745 | int added_disks; | |
2746 | ||
2747 | int *fdlist = NULL; | |
2748 | unsigned long long *offsets = NULL; | |
2749 | int d; | |
2750 | int nrdisks; | |
2751 | int err; | |
2752 | unsigned long blocks; | |
2753 | unsigned long long array_size; | |
2754 | int done; | |
2755 | struct mdinfo *sra = NULL; | |
2756 | ||
2757 | /* when reshaping a RAID0, the component_size might be zero. | |
2758 | * So try to fix that up. | |
2759 | */ | |
2760 | if (ioctl(fd, GET_ARRAY_INFO, &array) != 0) { | |
2761 | dprintf("Cannot get array information.\n"); | |
2762 | goto release; | |
2763 | } | |
2764 | if (array.level == 0 && info->component_size == 0) { | |
2765 | get_dev_size(fd, NULL, &array_size); | |
2766 | info->component_size = array_size / array.raid_disks; | |
2767 | } | |
2768 | ||
2769 | if (array.level == 10) | |
2770 | /* Need space_after info */ | |
2771 | get_space_after(fd, st, info); | |
2772 | ||
2773 | if (info->reshape_active) { | |
2774 | int new_level = info->new_level; | |
2775 | info->new_level = UnSet; | |
2776 | if (info->delta_disks > 0) | |
2777 | info->array.raid_disks -= info->delta_disks; | |
2778 | msg = analyse_change(info, &reshape); | |
2779 | info->new_level = new_level; | |
2780 | if (info->delta_disks > 0) | |
2781 | info->array.raid_disks += info->delta_disks; | |
2782 | if (!restart) | |
2783 | /* Make sure the array isn't read-only */ | |
2784 | ioctl(fd, RESTART_ARRAY_RW, 0); | |
2785 | } else | |
2786 | msg = analyse_change(info, &reshape); | |
2787 | if (msg) { | |
2788 | pr_err("%s\n", msg); | |
2789 | goto release; | |
2790 | } | |
2791 | if (restart && | |
2792 | (reshape.level != info->array.level || | |
2793 | reshape.before.layout != info->array.layout || | |
2794 | reshape.before.data_disks + reshape.parity | |
2795 | != info->array.raid_disks - max(0, info->delta_disks))) { | |
2796 | pr_err("reshape info is not in native format -" | |
2797 | " cannot continue.\n"); | |
2798 | goto release; | |
2799 | } | |
2800 | ||
2801 | if (st->ss->external && restart && (info->reshape_progress == 0)) { | |
2802 | /* When reshape is restarted from '0', very begin of array | |
2803 | * it is possible that for external metadata reshape and array | |
2804 | * configuration doesn't happen. | |
2805 | * Check if md has the same opinion, and reshape is restarted | |
2806 | * from 0. If so, this is regular reshape start after reshape | |
2807 | * switch in metadata to next array only. | |
2808 | */ | |
2809 | if ((verify_reshape_position(info, reshape.level) >= 0) && | |
2810 | (info->reshape_progress == 0)) | |
2811 | restart = 0; | |
2812 | } | |
2813 | if (restart) { | |
2814 | /* reshape already started. just skip to monitoring the reshape */ | |
2815 | if (reshape.backup_blocks == 0) | |
2816 | return 0; | |
2817 | if (restart & RESHAPE_NO_BACKUP) | |
2818 | return 0; | |
2819 | goto started; | |
2820 | } | |
2821 | /* The container is frozen but the array may not be. | |
2822 | * So freeze the array so spares don't get put to the wrong use | |
2823 | * FIXME there should probably be a cleaner separation between | |
2824 | * freeze_array and freeze_container. | |
2825 | */ | |
2826 | sysfs_freeze_array(info); | |
2827 | /* Check we have enough spares to not be degraded */ | |
2828 | added_disks = 0; | |
2829 | for (dv = devlist; dv ; dv=dv->next) | |
2830 | added_disks++; | |
2831 | spares_needed = max(reshape.before.data_disks, | |
2832 | reshape.after.data_disks) | |
2833 | + reshape.parity - array.raid_disks; | |
2834 | ||
2835 | if (!force && | |
2836 | info->new_level > 1 && info->array.level > 1 && | |
2837 | spares_needed > info->array.spare_disks + added_disks) { | |
2838 | pr_err("Need %d spare%s to avoid degraded array," | |
2839 | " and only have %d.\n" | |
2840 | " Use --force to over-ride this check.\n", | |
2841 | spares_needed, | |
2842 | spares_needed == 1 ? "" : "s", | |
2843 | info->array.spare_disks + added_disks); | |
2844 | goto release; | |
2845 | } | |
2846 | /* Check we have enough spares to not fail */ | |
2847 | spares_needed = max(reshape.before.data_disks, | |
2848 | reshape.after.data_disks) | |
2849 | - array.raid_disks; | |
2850 | if ((info->new_level > 1 || info->new_level == 0) && | |
2851 | spares_needed > info->array.spare_disks +added_disks) { | |
2852 | pr_err("Need %d spare%s to create working array," | |
2853 | " and only have %d.\n", | |
2854 | spares_needed, | |
2855 | spares_needed == 1 ? "" : "s", | |
2856 | info->array.spare_disks + added_disks); | |
2857 | goto release; | |
2858 | } | |
2859 | ||
2860 | if (reshape.level != array.level) { | |
2861 | int err = impose_level(fd, reshape.level, devname, verbose); | |
2862 | if (err) | |
2863 | goto release; | |
2864 | info->new_layout = UnSet; /* after level change, | |
2865 | * layout is meaningless */ | |
2866 | orig_level = array.level; | |
2867 | sysfs_freeze_array(info); | |
2868 | ||
2869 | if (reshape.level > 0 && st->ss->external) { | |
2870 | /* make sure mdmon is aware of the new level */ | |
2871 | if (mdmon_running(container)) | |
2872 | flush_mdmon(container); | |
2873 | ||
2874 | if (!mdmon_running(container)) | |
2875 | start_mdmon(container); | |
2876 | ping_monitor(container); | |
2877 | if (mdmon_running(container) && | |
2878 | st->update_tail == NULL) | |
2879 | st->update_tail = &st->updates; | |
2880 | } | |
2881 | } | |
2882 | /* ->reshape_super might have chosen some spares from the | |
2883 | * container that it wants to be part of the new array. | |
2884 | * We can collect them with ->container_content and give | |
2885 | * them to the kernel. | |
2886 | */ | |
2887 | if (st->ss->reshape_super && st->ss->container_content) { | |
2888 | char *subarray = strchr(info->text_version+1, '/')+1; | |
2889 | struct mdinfo *info2 = | |
2890 | st->ss->container_content(st, subarray); | |
2891 | struct mdinfo *d; | |
2892 | ||
2893 | if (info2) { | |
2894 | sysfs_init(info2, fd, st->devnm); | |
2895 | /* When increasing number of devices, we need to set | |
2896 | * new raid_disks before adding these, or they might | |
2897 | * be rejected. | |
2898 | */ | |
2899 | if (reshape.backup_blocks && | |
2900 | reshape.after.data_disks > reshape.before.data_disks) | |
2901 | subarray_set_num(container, info2, "raid_disks", | |
2902 | reshape.after.data_disks + | |
2903 | reshape.parity); | |
2904 | for (d = info2->devs; d; d = d->next) { | |
2905 | if (d->disk.state == 0 && | |
2906 | d->disk.raid_disk >= 0) { | |
2907 | /* This is a spare that wants to | |
2908 | * be part of the array. | |
2909 | */ | |
2910 | add_disk(fd, st, info2, d); | |
2911 | } | |
2912 | } | |
2913 | sysfs_free(info2); | |
2914 | } | |
2915 | } | |
2916 | /* We might have been given some devices to add to the | |
2917 | * array. Now that the array has been changed to the right | |
2918 | * level and frozen, we can safely add them. | |
2919 | */ | |
2920 | if (devlist) | |
2921 | Manage_subdevs(devname, fd, devlist, verbose, | |
2922 | 0,NULL, 0); | |
2923 | ||
2924 | if (reshape.backup_blocks == 0 && data_offset != INVALID_SECTORS) | |
2925 | reshape.backup_blocks = reshape.before.data_disks * info->array.chunk_size/512; | |
2926 | if (reshape.backup_blocks == 0) { | |
2927 | /* No restriping needed, but we might need to impose | |
2928 | * some more changes: layout, raid_disks, chunk_size | |
2929 | */ | |
2930 | /* read current array info */ | |
2931 | if (ioctl(fd, GET_ARRAY_INFO, &array) != 0) { | |
2932 | dprintf("Cannot get array information.\n"); | |
2933 | goto release; | |
2934 | } | |
2935 | /* compare current array info with new values and if | |
2936 | * it is different update them to new */ | |
2937 | if (info->new_layout != UnSet && | |
2938 | info->new_layout != array.layout) { | |
2939 | array.layout = info->new_layout; | |
2940 | if (ioctl(fd, SET_ARRAY_INFO, &array) != 0) { | |
2941 | pr_err("failed to set new layout\n"); | |
2942 | goto release; | |
2943 | } else if (verbose >= 0) | |
2944 | printf("layout for %s set to %d\n", | |
2945 | devname, array.layout); | |
2946 | } | |
2947 | if (info->delta_disks != UnSet && | |
2948 | info->delta_disks != 0 && | |
2949 | array.raid_disks != (info->array.raid_disks + info->delta_disks)) { | |
2950 | array.raid_disks += info->delta_disks; | |
2951 | if (ioctl(fd, SET_ARRAY_INFO, &array) != 0) { | |
2952 | pr_err("failed to set raid disks\n"); | |
2953 | goto release; | |
2954 | } else if (verbose >= 0) { | |
2955 | printf("raid_disks for %s set to %d\n", | |
2956 | devname, array.raid_disks); | |
2957 | } | |
2958 | } | |
2959 | if (info->new_chunk != 0 && | |
2960 | info->new_chunk != array.chunk_size) { | |
2961 | if (sysfs_set_num(info, NULL, | |
2962 | "chunk_size", info->new_chunk) != 0) { | |
2963 | pr_err("failed to set chunk size\n"); | |
2964 | goto release; | |
2965 | } else if (verbose >= 0) | |
2966 | printf("chunk size for %s set to %d\n", | |
2967 | devname, array.chunk_size); | |
2968 | } | |
2969 | unfreeze(st); | |
2970 | return 0; | |
2971 | } | |
2972 | ||
2973 | /* | |
2974 | * There are three possibilities. | |
2975 | * 1/ The array will shrink. | |
2976 | * We need to ensure the reshape will pause before reaching | |
2977 | * the 'critical section'. We also need to fork and wait for | |
2978 | * that to happen. When it does we | |
2979 | * suspend/backup/complete/unfreeze | |
2980 | * | |
2981 | * 2/ The array will not change size. | |
2982 | * This requires that we keep a backup of a sliding window | |
2983 | * so that we can restore data after a crash. So we need | |
2984 | * to fork and monitor progress. | |
2985 | * In future we will allow the data_offset to change, so | |
2986 | * a sliding backup becomes unnecessary. | |
2987 | * | |
2988 | * 3/ The array will grow. This is relatively easy. | |
2989 | * However the kernel's restripe routines will cheerfully | |
2990 | * overwrite some early data before it is safe. So we | |
2991 | * need to make a backup of the early parts of the array | |
2992 | * and be ready to restore it if rebuild aborts very early. | |
2993 | * For externally managed metadata, we still need a forked | |
2994 | * child to monitor the reshape and suspend IO over the region | |
2995 | * that is being reshaped. | |
2996 | * | |
2997 | * We backup data by writing it to one spare, or to a | |
2998 | * file which was given on command line. | |
2999 | * | |
3000 | * In each case, we first make sure that storage is available | |
3001 | * for the required backup. | |
3002 | * Then we: | |
3003 | * - request the shape change. | |
3004 | * - fork to handle backup etc. | |
3005 | */ | |
3006 | /* Check that we can hold all the data */ | |
3007 | get_dev_size(fd, NULL, &array_size); | |
3008 | if (reshape.new_size < (array_size/512)) { | |
3009 | pr_err("this change will reduce the size of the array.\n" | |
3010 | " use --grow --array-size first to truncate array.\n" | |
3011 | " e.g. mdadm --grow %s --array-size %llu\n", | |
3012 | devname, reshape.new_size/2); | |
3013 | goto release; | |
3014 | } | |
3015 | ||
3016 | if (array.level == 10) { | |
3017 | /* Reshaping RAID10 does not require any data backup by | |
3018 | * user-space. Instead it requires that the data_offset | |
3019 | * is changed to avoid the need for backup. | |
3020 | * So this is handled very separately | |
3021 | */ | |
3022 | if (restart) | |
3023 | /* Nothing to do. */ | |
3024 | return 0; | |
3025 | return raid10_reshape(container, fd, devname, st, info, | |
3026 | &reshape, data_offset, | |
3027 | force, verbose); | |
3028 | } | |
3029 | sra = sysfs_read(fd, NULL, | |
3030 | GET_COMPONENT|GET_DEVS|GET_OFFSET|GET_STATE|GET_CHUNK| | |
3031 | GET_CACHE); | |
3032 | if (!sra) { | |
3033 | pr_err("%s: Cannot get array details from sysfs\n", | |
3034 | devname); | |
3035 | goto release; | |
3036 | } | |
3037 | ||
3038 | if (!backup_file) | |
3039 | switch(set_new_data_offset(sra, st, devname, | |
3040 | reshape.after.data_disks - reshape.before.data_disks, | |
3041 | data_offset, | |
3042 | reshape.min_offset_change)) { | |
3043 | case -1: | |
3044 | goto release; | |
3045 | case 0: | |
3046 | /* Updated data_offset, so it's easy now */ | |
3047 | update_cache_size(container, sra, info, | |
3048 | min(reshape.before.data_disks, | |
3049 | reshape.after.data_disks), | |
3050 | reshape.backup_blocks); | |
3051 | ||
3052 | /* Right, everything seems fine. Let's kick things off. | |
3053 | */ | |
3054 | sync_metadata(st); | |
3055 | ||
3056 | if (impose_reshape(sra, info, st, fd, restart, | |
3057 | devname, container, &reshape) < 0) | |
3058 | goto release; | |
3059 | if (sysfs_set_str(sra, NULL, "sync_action", "reshape") < 0) { | |
3060 | pr_err("Failed to initiate reshape!\n"); | |
3061 | goto release; | |
3062 | } | |
3063 | if (info->new_level == reshape.level) | |
3064 | return 0; | |
3065 | /* need to adjust level when reshape completes */ | |
3066 | switch(fork()) { | |
3067 | case -1: /* ignore error, but don't wait */ | |
3068 | return 0; | |
3069 | default: /* parent */ | |
3070 | return 0; | |
3071 | case 0: | |
3072 | map_fork(); | |
3073 | break; | |
3074 | } | |
3075 | wait_reshape(sra); | |
3076 | impose_level(fd, info->new_level, devname, verbose); | |
3077 | ||
3078 | return 0; | |
3079 | case 1: /* Couldn't set data_offset, try the old way */ | |
3080 | if (data_offset != INVALID_SECTORS) { | |
3081 | pr_err("Cannot update data_offset on this array\n"); | |
3082 | goto release; | |
3083 | } | |
3084 | break; | |
3085 | } | |
3086 | ||
3087 | started: | |
3088 | /* Decide how many blocks (sectors) for a reshape | |
3089 | * unit. The number we have so far is just a minimum | |
3090 | */ | |
3091 | blocks = reshape.backup_blocks; | |
3092 | if (reshape.before.data_disks == | |
3093 | reshape.after.data_disks) { | |
3094 | /* Make 'blocks' bigger for better throughput, but | |
3095 | * not so big that we reject it below. | |
3096 | * Try for 16 megabytes | |
3097 | */ | |
3098 | while (blocks * 32 < sra->component_size && | |
3099 | blocks < 16*1024*2) | |
3100 | blocks *= 2; | |
3101 | } else | |
3102 | pr_err("Need to backup %luK of critical " | |
3103 | "section..\n", blocks/2); | |
3104 | ||
3105 | if (blocks >= sra->component_size/2) { | |
3106 | pr_err("%s: Something wrong" | |
3107 | " - reshape aborted\n", | |
3108 | devname); | |
3109 | goto release; | |
3110 | } | |
3111 | ||
3112 | /* Now we need to open all these devices so we can read/write. | |
3113 | */ | |
3114 | nrdisks = max(reshape.before.data_disks, | |
3115 | reshape.after.data_disks) + reshape.parity | |
3116 | + sra->array.spare_disks; | |
3117 | fdlist = xcalloc((1+nrdisks), sizeof(int)); | |
3118 | offsets = xcalloc((1+nrdisks), sizeof(offsets[0])); | |
3119 | ||
3120 | odisks = reshape.before.data_disks + reshape.parity; | |
3121 | d = reshape_prepare_fdlist(devname, sra, odisks, | |
3122 | nrdisks, blocks, backup_file, | |
3123 | fdlist, offsets); | |
3124 | if (d < 0) { | |
3125 | goto release; | |
3126 | } | |
3127 | if ((st->ss->manage_reshape == NULL) || | |
3128 | (st->ss->recover_backup == NULL)) { | |
3129 | if (backup_file == NULL) { | |
3130 | if (reshape.after.data_disks <= | |
3131 | reshape.before.data_disks) { | |
3132 | pr_err("%s: Cannot grow - " | |
3133 | "need backup-file\n", devname); | |
3134 | goto release; | |
3135 | } else if (sra->array.spare_disks == 0) { | |
3136 | pr_err("%s: Cannot grow - " | |
3137 | "need a spare or backup-file to backup " | |
3138 | "critical section\n", devname); | |
3139 | goto release; | |
3140 | } | |
3141 | } else { | |
3142 | if (!reshape_open_backup_file(backup_file, fd, devname, | |
3143 | (signed)blocks, | |
3144 | fdlist+d, offsets+d, | |
3145 | restart)) { | |
3146 | goto release; | |
3147 | } | |
3148 | d++; | |
3149 | } | |
3150 | } | |
3151 | ||
3152 | update_cache_size(container, sra, info, | |
3153 | min(reshape.before.data_disks, reshape.after.data_disks), | |
3154 | blocks); | |
3155 | ||
3156 | /* Right, everything seems fine. Let's kick things off. | |
3157 | * If only changing raid_disks, use ioctl, else use | |
3158 | * sysfs. | |
3159 | */ | |
3160 | sync_metadata(st); | |
3161 | ||
3162 | if (impose_reshape(sra, info, st, fd, restart, | |
3163 | devname, container, &reshape) < 0) | |
3164 | goto release; | |
3165 | ||
3166 | err = start_reshape(sra, restart, reshape.before.data_disks, | |
3167 | reshape.after.data_disks); | |
3168 | if (err) { | |
3169 | pr_err("Cannot %s reshape for %s\n", | |
3170 | restart ? "continue" : "start", | |
3171 | devname); | |
3172 | goto release; | |
3173 | } | |
3174 | if (restart) | |
3175 | sysfs_set_str(sra, NULL, "array_state", "active"); | |
3176 | if (freeze_reshape) { | |
3177 | free(fdlist); | |
3178 | free(offsets); | |
3179 | sysfs_free(sra); | |
3180 | pr_err("Reshape has to be continued from" | |
3181 | " location %llu when root filesystem has been mounted.\n", | |
3182 | sra->reshape_progress); | |
3183 | return 1; | |
3184 | } | |
3185 | ||
3186 | /* Now we just need to kick off the reshape and watch, while | |
3187 | * handling backups of the data... | |
3188 | * This is all done by a forked background process. | |
3189 | */ | |
3190 | switch(forked ? 0 : fork()) { | |
3191 | case -1: | |
3192 | pr_err("Cannot run child to monitor reshape: %s\n", | |
3193 | strerror(errno)); | |
3194 | abort_reshape(sra); | |
3195 | goto release; | |
3196 | default: | |
3197 | free(fdlist); | |
3198 | free(offsets); | |
3199 | sysfs_free(sra); | |
3200 | return 0; | |
3201 | case 0: | |
3202 | map_fork(); | |
3203 | break; | |
3204 | } | |
3205 | ||
3206 | /* If another array on the same devices is busy, the | |
3207 | * reshape will wait for them. This would mean that | |
3208 | * the first section that we suspend will stay suspended | |
3209 | * for a long time. So check on that possibility | |
3210 | * by looking for "DELAYED" in /proc/mdstat, and if found, | |
3211 | * wait a while | |
3212 | */ | |
3213 | do { | |
3214 | struct mdstat_ent *mds, *m; | |
3215 | delayed = 0; | |
3216 | mds = mdstat_read(0, 0); | |
3217 | for (m = mds; m; m = m->next) | |
3218 | if (strcmp(m->devnm, sra->sys_name) == 0) { | |
3219 | if (m->resync && | |
3220 | m->percent == RESYNC_DELAYED) | |
3221 | delayed = 1; | |
3222 | if (m->resync == 0) | |
3223 | /* Haven't started the reshape thread | |
3224 | * yet, wait a bit | |
3225 | */ | |
3226 | delayed = 2; | |
3227 | break; | |
3228 | } | |
3229 | free_mdstat(mds); | |
3230 | if (delayed == 1 && get_linux_version() < 3007000) { | |
3231 | pr_err("Reshape is delayed, but cannot wait carefully with this kernel.\n" | |
3232 | " You might experience problems until other reshapes complete.\n"); | |
3233 | delayed = 0; | |
3234 | } | |
3235 | if (delayed) | |
3236 | sleep(30 - (delayed-1) * 25); | |
3237 | } while (delayed); | |
3238 | ||
3239 | close(fd); | |
3240 | if (check_env("MDADM_GROW_VERIFY")) | |
3241 | fd = open(devname, O_RDONLY | O_DIRECT); | |
3242 | else | |
3243 | fd = -1; | |
3244 | mlockall(MCL_FUTURE); | |
3245 | ||
3246 | if (st->ss->external) { | |
3247 | /* metadata handler takes it from here */ | |
3248 | done = st->ss->manage_reshape( | |
3249 | fd, sra, &reshape, st, blocks, | |
3250 | fdlist, offsets, | |
3251 | d - odisks, fdlist+odisks, | |
3252 | offsets+odisks); | |
3253 | } else | |
3254 | done = child_monitor( | |
3255 | fd, sra, &reshape, st, blocks, | |
3256 | fdlist, offsets, | |
3257 | d - odisks, fdlist+odisks, | |
3258 | offsets+odisks); | |
3259 | ||
3260 | free(fdlist); | |
3261 | free(offsets); | |
3262 | ||
3263 | if (backup_file && done) | |
3264 | unlink(backup_file); | |
3265 | if (!done) { | |
3266 | abort_reshape(sra); | |
3267 | goto out; | |
3268 | } | |
3269 | ||
3270 | if (!st->ss->external && | |
3271 | !(reshape.before.data_disks != reshape.after.data_disks | |
3272 | && info->custom_array_size) && | |
3273 | info->new_level == reshape.level && | |
3274 | !forked) { | |
3275 | /* no need to wait for the reshape to finish as | |
3276 | * there is nothing more to do. | |
3277 | */ | |
3278 | sysfs_free(sra); | |
3279 | exit(0); | |
3280 | } | |
3281 | wait_reshape(sra); | |
3282 | ||
3283 | if (st->ss->external) { | |
3284 | /* Re-load the metadata as much could have changed */ | |
3285 | int cfd = open_dev(st->container_devnm); | |
3286 | if (cfd >= 0) { | |
3287 | flush_mdmon(container); | |
3288 | st->ss->free_super(st); | |
3289 | st->ss->load_container(st, cfd, container); | |
3290 | close(cfd); | |
3291 | } | |
3292 | } | |
3293 | ||
3294 | /* set new array size if required customer_array_size is used | |
3295 | * by this metadata. | |
3296 | */ | |
3297 | if (reshape.before.data_disks != | |
3298 | reshape.after.data_disks && | |
3299 | info->custom_array_size) | |
3300 | set_array_size(st, info, info->text_version); | |
3301 | ||
3302 | if (info->new_level != reshape.level) { | |
3303 | if (fd < 0) | |
3304 | fd = open(devname, O_RDONLY); | |
3305 | impose_level(fd, info->new_level, devname, verbose); | |
3306 | close(fd); | |
3307 | if (info->new_level == 0) | |
3308 | st->update_tail = NULL; | |
3309 | } | |
3310 | out: | |
3311 | sysfs_free(sra); | |
3312 | if (forked) | |
3313 | return 0; | |
3314 | unfreeze(st); | |
3315 | exit(0); | |
3316 | ||
3317 | release: | |
3318 | free(fdlist); | |
3319 | free(offsets); | |
3320 | if (orig_level != UnSet && sra) { | |
3321 | c = map_num(pers, orig_level); | |
3322 | if (c && sysfs_set_str(sra, NULL, "level", c) == 0) | |
3323 | pr_err("aborting level change\n"); | |
3324 | } | |
3325 | sysfs_free(sra); | |
3326 | if (!forked) | |
3327 | unfreeze(st); | |
3328 | return 1; | |
3329 | } | |
3330 | ||
3331 | /* mdfd handle is passed to be closed in child process (after fork). | |
3332 | */ | |
3333 | int reshape_container(char *container, char *devname, | |
3334 | int mdfd, | |
3335 | struct supertype *st, | |
3336 | struct mdinfo *info, | |
3337 | int force, | |
3338 | char *backup_file, | |
3339 | int verbose, int restart, int freeze_reshape) | |
3340 | { | |
3341 | struct mdinfo *cc = NULL; | |
3342 | int rv = restart; | |
3343 | char last_devnm[32] = ""; | |
3344 | ||
3345 | /* component_size is not meaningful for a container, | |
3346 | * so pass '0' meaning 'no change' | |
3347 | */ | |
3348 | if (!restart && | |
3349 | reshape_super(st, 0, info->new_level, | |
3350 | info->new_layout, info->new_chunk, | |
3351 | info->array.raid_disks, info->delta_disks, | |
3352 | backup_file, devname, APPLY_METADATA_CHANGES, | |
3353 | verbose)) { | |
3354 | unfreeze(st); | |
3355 | return 1; | |
3356 | } | |
3357 | ||
3358 | sync_metadata(st); | |
3359 | ||
3360 | /* ping monitor to be sure that update is on disk | |
3361 | */ | |
3362 | ping_monitor(container); | |
3363 | ||
3364 | switch (fork()) { | |
3365 | case -1: /* error */ | |
3366 | perror("Cannot fork to complete reshape\n"); | |
3367 | unfreeze(st); | |
3368 | return 1; | |
3369 | default: /* parent */ | |
3370 | if (!freeze_reshape) | |
3371 | printf(Name ": multi-array reshape continues" | |
3372 | " in background\n"); | |
3373 | return 0; | |
3374 | case 0: /* child */ | |
3375 | map_fork(); | |
3376 | break; | |
3377 | } | |
3378 | ||
3379 | /* close unused handle in child process | |
3380 | */ | |
3381 | if (mdfd > -1) | |
3382 | close(mdfd); | |
3383 | ||
3384 | while(1) { | |
3385 | /* For each member array with reshape_active, | |
3386 | * we need to perform the reshape. | |
3387 | * We pick the first array that needs reshaping and | |
3388 | * reshape it. reshape_array() will re-read the metadata | |
3389 | * so the next time through a different array should be | |
3390 | * ready for reshape. | |
3391 | * It is possible that the 'different' array will not | |
3392 | * be assembled yet. In that case we simple exit. | |
3393 | * When it is assembled, the mdadm which assembles it | |
3394 | * will take over the reshape. | |
3395 | */ | |
3396 | struct mdinfo *content; | |
3397 | int fd; | |
3398 | struct mdstat_ent *mdstat; | |
3399 | char *adev; | |
3400 | int devid; | |
3401 | ||
3402 | sysfs_free(cc); | |
3403 | ||
3404 | cc = st->ss->container_content(st, NULL); | |
3405 | ||
3406 | for (content = cc; content ; content = content->next) { | |
3407 | char *subarray; | |
3408 | if (!content->reshape_active) | |
3409 | continue; | |
3410 | ||
3411 | subarray = strchr(content->text_version+1, '/')+1; | |
3412 | mdstat = mdstat_by_subdev(subarray, container); | |
3413 | if (!mdstat) | |
3414 | continue; | |
3415 | if (mdstat->active == 0) { | |
3416 | pr_err("Skipping inactive array %s.\n", | |
3417 | mdstat->devnm); | |
3418 | free_mdstat(mdstat); | |
3419 | mdstat = NULL; | |
3420 | continue; | |
3421 | } | |
3422 | break; | |
3423 | } | |
3424 | if (!content) | |
3425 | break; | |
3426 | ||
3427 | devid = devnm2devid(mdstat->devnm); | |
3428 | adev = map_dev(major(devid), minor(devid), 0); | |
3429 | if (!adev) | |
3430 | adev = content->text_version; | |
3431 | ||
3432 | fd = open_dev(mdstat->devnm); | |
3433 | if (fd < 0) { | |
3434 | printf(Name ": Device %s cannot be opened for reshape.", | |
3435 | adev); | |
3436 | break; | |
3437 | } | |
3438 | ||
3439 | if (strcmp(last_devnm, mdstat->devnm) == 0) { | |
3440 | /* Do not allow for multiple reshape_array() calls for | |
3441 | * the same array. | |
3442 | * It can happen when reshape_array() returns without | |
3443 | * error, when reshape is not finished (wrong reshape | |
3444 | * starting/continuation conditions). Mdmon doesn't | |
3445 | * switch to next array in container and reentry | |
3446 | * conditions for the same array occur. | |
3447 | * This is possibly interim until the behaviour of | |
3448 | * reshape_array is resolved(). | |
3449 | */ | |
3450 | printf(Name ": Multiple reshape execution detected for " | |
3451 | "device %s.", adev); | |
3452 | close(fd); | |
3453 | break; | |
3454 | } | |
3455 | strcpy(last_devnm, mdstat->devnm); | |
3456 | ||
3457 | sysfs_init(content, fd, mdstat->devnm); | |
3458 | ||
3459 | if (mdmon_running(container)) | |
3460 | flush_mdmon(container); | |
3461 | ||
3462 | rv = reshape_array(container, fd, adev, st, | |
3463 | content, force, NULL, 0ULL, | |
3464 | backup_file, verbose, 1, restart, | |
3465 | freeze_reshape); | |
3466 | close(fd); | |
3467 | ||
3468 | if (freeze_reshape) { | |
3469 | sysfs_free(cc); | |
3470 | exit(0); | |
3471 | } | |
3472 | ||
3473 | restart = 0; | |
3474 | if (rv) | |
3475 | break; | |
3476 | ||
3477 | if (mdmon_running(container)) | |
3478 | flush_mdmon(container); | |
3479 | } | |
3480 | if (!rv) | |
3481 | unfreeze(st); | |
3482 | sysfs_free(cc); | |
3483 | exit(0); | |
3484 | } | |
3485 | ||
3486 | /* | |
3487 | * We run a child process in the background which performs the following | |
3488 | * steps: | |
3489 | * - wait for resync to reach a certain point | |
3490 | * - suspend io to the following section | |
3491 | * - backup that section | |
3492 | * - allow resync to proceed further | |
3493 | * - resume io | |
3494 | * - discard the backup. | |
3495 | * | |
3496 | * When are combined in slightly different ways in the three cases. | |
3497 | * Grow: | |
3498 | * - suspend/backup/allow/wait/resume/discard | |
3499 | * Shrink: | |
3500 | * - allow/wait/suspend/backup/allow/wait/resume/discard | |
3501 | * same-size: | |
3502 | * - wait/resume/discard/suspend/backup/allow | |
3503 | * | |
3504 | * suspend/backup/allow always come together | |
3505 | * wait/resume/discard do too. | |
3506 | * For the same-size case we have two backups to improve flow. | |
3507 | * | |
3508 | */ | |
3509 | ||
3510 | int progress_reshape(struct mdinfo *info, struct reshape *reshape, | |
3511 | unsigned long long backup_point, | |
3512 | unsigned long long wait_point, | |
3513 | unsigned long long *suspend_point, | |
3514 | unsigned long long *reshape_completed) | |
3515 | { | |
3516 | /* This function is called repeatedly by the reshape manager. | |
3517 | * It determines how much progress can safely be made and allows | |
3518 | * that progress. | |
3519 | * - 'info' identifies the array and particularly records in | |
3520 | * ->reshape_progress the metadata's knowledge of progress | |
3521 | * This is a sector offset from the start of the array | |
3522 | * of the next array block to be relocated. This number | |
3523 | * may increase from 0 or decrease from array_size, depending | |
3524 | * on the type of reshape that is happening. | |
3525 | * Note that in contrast, 'sync_completed' is a block count of the | |
3526 | * reshape so far. It gives the distance between the start point | |
3527 | * (head or tail of device) and the next place that data will be | |
3528 | * written. It always increases. | |
3529 | * - 'reshape' is the structure created by analyse_change | |
3530 | * - 'backup_point' shows how much the metadata manager has backed-up | |
3531 | * data. For reshapes with increasing progress, it is the next address | |
3532 | * to be backed up, previous addresses have been backed-up. For | |
3533 | * decreasing progress, it is the earliest address that has been | |
3534 | * backed up - later address are also backed up. | |
3535 | * So addresses between reshape_progress and backup_point are | |
3536 | * backed up providing those are in the 'correct' order. | |
3537 | * - 'wait_point' is an array address. When reshape_completed | |
3538 | * passes this point, progress_reshape should return. It might | |
3539 | * return earlier if it determines that ->reshape_progress needs | |
3540 | * to be updated or further backup is needed. | |
3541 | * - suspend_point is maintained by progress_reshape and the caller | |
3542 | * should not touch it except to initialise to zero. | |
3543 | * It is an array address and it only increases in 2.6.37 and earlier. | |
3544 | * This makes it difficult to handle reducing reshapes with | |
3545 | * external metadata. | |
3546 | * However: it is similar to backup_point in that it records the | |
3547 | * other end of a suspended region from reshape_progress. | |
3548 | * it is moved to extend the region that is safe to backup and/or | |
3549 | * reshape | |
3550 | * - reshape_completed is read from sysfs and returned. The caller | |
3551 | * should copy this into ->reshape_progress when it has reason to | |
3552 | * believe that the metadata knows this, and any backup outside this | |
3553 | * has been erased. | |
3554 | * | |
3555 | * Return value is: | |
3556 | * 1 if more data from backup_point - but only as far as suspend_point, | |
3557 | * should be backed up | |
3558 | * 0 if things are progressing smoothly | |
3559 | * -1 if the reshape is finished because it is all done, | |
3560 | * -2 if the reshape is finished due to an error. | |
3561 | */ | |
3562 | ||
3563 | int advancing = (reshape->after.data_disks | |
3564 | >= reshape->before.data_disks); | |
3565 | unsigned long long need_backup; /* All data between start of array and | |
3566 | * here will at some point need to | |
3567 | * be backed up. | |
3568 | */ | |
3569 | unsigned long long read_offset, write_offset; | |
3570 | unsigned long long write_range; | |
3571 | unsigned long long max_progress, target, completed; | |
3572 | unsigned long long array_size = (info->component_size | |
3573 | * reshape->before.data_disks); | |
3574 | int fd; | |
3575 | char buf[20]; | |
3576 | ||
3577 | /* First, we unsuspend any region that is now known to be safe. | |
3578 | * If suspend_point is on the 'wrong' side of reshape_progress, then | |
3579 | * we don't have or need suspension at the moment. This is true for | |
3580 | * native metadata when we don't need to back-up. | |
3581 | */ | |
3582 | if (advancing) { | |
3583 | if (info->reshape_progress <= *suspend_point) | |
3584 | sysfs_set_num(info, NULL, "suspend_lo", | |
3585 | info->reshape_progress); | |
3586 | } else { | |
3587 | /* Note: this won't work in 2.6.37 and before. | |
3588 | * Something somewhere should make sure we don't need it! | |
3589 | */ | |
3590 | if (info->reshape_progress >= *suspend_point) | |
3591 | sysfs_set_num(info, NULL, "suspend_hi", | |
3592 | info->reshape_progress); | |
3593 | } | |
3594 | ||
3595 | /* Now work out how far it is safe to progress. | |
3596 | * If the read_offset for ->reshape_progress is less than | |
3597 | * 'blocks' beyond the write_offset, we can only progress as far | |
3598 | * as a backup. | |
3599 | * Otherwise we can progress until the write_offset for the new location | |
3600 | * reaches (within 'blocks' of) the read_offset at the current location. | |
3601 | * However that region must be suspended unless we are using native | |
3602 | * metadata. | |
3603 | * If we need to suspend more, we limit it to 128M per device, which is | |
3604 | * rather arbitrary and should be some time-based calculation. | |
3605 | */ | |
3606 | read_offset = info->reshape_progress / reshape->before.data_disks; | |
3607 | write_offset = info->reshape_progress / reshape->after.data_disks; | |
3608 | write_range = info->new_chunk/512; | |
3609 | if (reshape->before.data_disks == reshape->after.data_disks) | |
3610 | need_backup = array_size; | |
3611 | else | |
3612 | need_backup = reshape->backup_blocks; | |
3613 | if (advancing) { | |
3614 | if (read_offset < write_offset + write_range) | |
3615 | max_progress = backup_point; | |
3616 | else | |
3617 | max_progress = | |
3618 | read_offset * | |
3619 | reshape->after.data_disks; | |
3620 | } else { | |
3621 | if (read_offset > write_offset - write_range) | |
3622 | /* Can only progress as far as has been backed up, | |
3623 | * which must be suspended */ | |
3624 | max_progress = backup_point; | |
3625 | else if (info->reshape_progress <= need_backup) | |
3626 | max_progress = backup_point; | |
3627 | else { | |
3628 | if (info->array.major_version >= 0) | |
3629 | /* Can progress until backup is needed */ | |
3630 | max_progress = need_backup; | |
3631 | else { | |
3632 | /* Can progress until metadata update is required */ | |
3633 | max_progress = | |
3634 | read_offset * | |
3635 | reshape->after.data_disks; | |
3636 | /* but data must be suspended */ | |
3637 | if (max_progress < *suspend_point) | |
3638 | max_progress = *suspend_point; | |
3639 | } | |
3640 | } | |
3641 | } | |
3642 | ||
3643 | /* We know it is safe to progress to 'max_progress' providing | |
3644 | * it is suspended or we are using native metadata. | |
3645 | * Consider extending suspend_point 128M per device if it | |
3646 | * is less than 64M per device beyond reshape_progress. | |
3647 | * But always do a multiple of 'blocks' | |
3648 | * FIXME this is too big - it takes to long to complete | |
3649 | * this much. | |
3650 | */ | |
3651 | target = 64*1024*2 * min(reshape->before.data_disks, | |
3652 | reshape->after.data_disks); | |
3653 | target /= reshape->backup_blocks; | |
3654 | if (target < 2) | |
3655 | target = 2; | |
3656 | target *= reshape->backup_blocks; | |
3657 | ||
3658 | /* For externally managed metadata we always need to suspend IO to | |
3659 | * the area being reshaped so we regularly push suspend_point forward. | |
3660 | * For native metadata we only need the suspend if we are going to do | |
3661 | * a backup. | |
3662 | */ | |
3663 | if (advancing) { | |
3664 | if ((need_backup > info->reshape_progress | |
3665 | || info->array.major_version < 0) && | |
3666 | *suspend_point < info->reshape_progress + target) { | |
3667 | if (need_backup < *suspend_point + 2 * target) | |
3668 | *suspend_point = need_backup; | |
3669 | else if (*suspend_point + 2 * target < array_size) | |
3670 | *suspend_point += 2 * target; | |
3671 | else | |
3672 | *suspend_point = array_size; | |
3673 | sysfs_set_num(info, NULL, "suspend_hi", *suspend_point); | |
3674 | if (max_progress > *suspend_point) | |
3675 | max_progress = *suspend_point; | |
3676 | } | |
3677 | } else { | |
3678 | if (info->array.major_version >= 0) { | |
3679 | /* Only need to suspend when about to backup */ | |
3680 | if (info->reshape_progress < need_backup * 2 && | |
3681 | *suspend_point > 0) { | |
3682 | *suspend_point = 0; | |
3683 | sysfs_set_num(info, NULL, "suspend_lo", 0); | |
3684 | sysfs_set_num(info, NULL, "suspend_hi", need_backup); | |
3685 | } | |
3686 | } else { | |
3687 | /* Need to suspend continually */ | |
3688 | if (info->reshape_progress < *suspend_point) | |
3689 | *suspend_point = info->reshape_progress; | |
3690 | if (*suspend_point + target < info->reshape_progress) | |
3691 | /* No need to move suspend region yet */; | |
3692 | else { | |
3693 | if (*suspend_point >= 2 * target) | |
3694 | *suspend_point -= 2 * target; | |
3695 | else | |
3696 | *suspend_point = 0; | |
3697 | sysfs_set_num(info, NULL, "suspend_lo", | |
3698 | *suspend_point); | |
3699 | } | |
3700 | if (max_progress < *suspend_point) | |
3701 | max_progress = *suspend_point; | |
3702 | } | |
3703 | } | |
3704 | ||
3705 | /* now set sync_max to allow that progress. sync_max, like | |
3706 | * sync_completed is a count of sectors written per device, so | |
3707 | * we find the difference between max_progress and the start point, | |
3708 | * and divide that by after.data_disks to get a sync_max | |
3709 | * number. | |
3710 | * At the same time we convert wait_point to a similar number | |
3711 | * for comparing against sync_completed. | |
3712 | */ | |
3713 | /* scale down max_progress to per_disk */ | |
3714 | max_progress /= reshape->after.data_disks; | |
3715 | /* Round to chunk size as some kernels give an erroneously high number */ | |
3716 | max_progress /= info->new_chunk/512; | |
3717 | max_progress *= info->new_chunk/512; | |
3718 | /* And round to old chunk size as the kernel wants that */ | |
3719 | max_progress /= info->array.chunk_size/512; | |
3720 | max_progress *= info->array.chunk_size/512; | |
3721 | /* Limit progress to the whole device */ | |
3722 | if (max_progress > info->component_size) | |
3723 | max_progress = info->component_size; | |
3724 | wait_point /= reshape->after.data_disks; | |
3725 | if (!advancing) { | |
3726 | /* switch from 'device offset' to 'processed block count' */ | |
3727 | max_progress = info->component_size - max_progress; | |
3728 | wait_point = info->component_size - wait_point; | |
3729 | } | |
3730 | ||
3731 | sysfs_set_num(info, NULL, "sync_max", max_progress); | |
3732 | ||
3733 | /* Now wait. If we have already reached the point that we were | |
3734 | * asked to wait to, don't wait at all, else wait for any change. | |
3735 | * We need to select on 'sync_completed' as that is the place that | |
3736 | * notifications happen, but we are really interested in | |
3737 | * 'reshape_position' | |
3738 | */ | |
3739 | fd = sysfs_get_fd(info, NULL, "sync_completed"); | |
3740 | if (fd < 0) | |
3741 | goto check_progress; | |
3742 | ||
3743 | if (sysfs_fd_get_ll(fd, &completed) < 0) | |
3744 | goto check_progress; | |
3745 | ||
3746 | while (completed < max_progress && completed < wait_point) { | |
3747 | /* Check that sync_action is still 'reshape' to avoid | |
3748 | * waiting forever on a dead array | |
3749 | */ | |
3750 | char action[20]; | |
3751 | fd_set rfds; | |
3752 | if (sysfs_get_str(info, NULL, "sync_action", | |
3753 | action, 20) <= 0 || | |
3754 | strncmp(action, "reshape", 7) != 0) | |
3755 | break; | |
3756 | /* Some kernels reset 'sync_completed' to zero | |
3757 | * before setting 'sync_action' to 'idle'. | |
3758 | * So we need these extra tests. | |
3759 | */ | |
3760 | if (completed == 0 && advancing | |
3761 | && info->reshape_progress > 0) | |
3762 | break; | |
3763 | if (completed == 0 && !advancing | |
3764 | && info->reshape_progress < (info->component_size | |
3765 | * reshape->after.data_disks)) | |
3766 | break; | |
3767 | FD_ZERO(&rfds); | |
3768 | FD_SET(fd, &rfds); | |
3769 | select(fd+1, NULL, NULL, &rfds, NULL); | |
3770 | if (sysfs_fd_get_ll(fd, &completed) < 0) | |
3771 | goto check_progress; | |
3772 | } | |
3773 | /* Some kernels reset 'sync_completed' to zero, | |
3774 | * we need to have real point we are in md | |
3775 | */ | |
3776 | if (completed == 0) | |
3777 | completed = max_progress; | |
3778 | ||
3779 | /* some kernels can give an incorrectly high 'completed' number */ | |
3780 | completed /= (info->new_chunk/512); | |
3781 | completed *= (info->new_chunk/512); | |
3782 | /* Convert 'completed' back in to a 'progress' number */ | |
3783 | completed *= reshape->after.data_disks; | |
3784 | if (!advancing) { | |
3785 | completed = info->component_size * reshape->after.data_disks | |
3786 | - completed; | |
3787 | } | |
3788 | *reshape_completed = completed; | |
3789 | ||
3790 | close(fd); | |
3791 | ||
3792 | /* We return the need_backup flag. Caller will decide | |
3793 | * how much - a multiple of ->backup_blocks up to *suspend_point | |
3794 | */ | |
3795 | if (advancing) | |
3796 | return need_backup > info->reshape_progress; | |
3797 | else | |
3798 | return need_backup >= info->reshape_progress; | |
3799 | ||
3800 | check_progress: | |
3801 | /* if we couldn't read a number from sync_completed, then | |
3802 | * either the reshape did complete, or it aborted. | |
3803 | * We can tell which by checking for 'none' in reshape_position. | |
3804 | * If it did abort, then it might immediately restart if it | |
3805 | * it was just a device failure that leaves us degraded but | |
3806 | * functioning. | |
3807 | */ | |
3808 | strcpy(buf, "hi"); | |
3809 | if (sysfs_get_str(info, NULL, "reshape_position", buf, sizeof(buf)) < 0 | |
3810 | || strncmp(buf, "none", 4) != 0) { | |
3811 | /* The abort might only be temporary. Wait up to 10 | |
3812 | * seconds for fd to contain a valid number again. | |
3813 | */ | |
3814 | struct timeval tv; | |
3815 | int rv = -2; | |
3816 | tv.tv_sec = 10; | |
3817 | tv.tv_usec = 0; | |
3818 | while (fd >= 0 && rv < 0 && tv.tv_sec > 0) { | |
3819 | fd_set rfds; | |
3820 | FD_ZERO(&rfds); | |
3821 | FD_SET(fd, &rfds); | |
3822 | if (select(fd+1, NULL, NULL, &rfds, &tv) != 1) | |
3823 | break; | |
3824 | switch (sysfs_fd_get_ll(fd, &completed)) { | |
3825 | case 0: | |
3826 | /* all good again */ | |
3827 | rv = 1; | |
3828 | break; | |
3829 | case -2: /* read error - abort */ | |
3830 | tv.tv_sec = 0; | |
3831 | break; | |
3832 | } | |
3833 | } | |
3834 | if (fd >= 0) | |
3835 | close(fd); | |
3836 | return rv; /* abort */ | |
3837 | } else { | |
3838 | /* Maybe racing with array shutdown - check state */ | |
3839 | if (fd >= 0) | |
3840 | close(fd); | |
3841 | if (sysfs_get_str(info, NULL, "array_state", buf, sizeof(buf)) < 0 | |
3842 | || strncmp(buf, "inactive", 8) == 0 | |
3843 | || strncmp(buf, "clear",5) == 0) | |
3844 | return -2; /* abort */ | |
3845 | return -1; /* complete */ | |
3846 | } | |
3847 | } | |
3848 | ||
3849 | /* FIXME return status is never checked */ | |
3850 | static int grow_backup(struct mdinfo *sra, | |
3851 | unsigned long long offset, /* per device */ | |
3852 | unsigned long stripes, /* per device, in old chunks */ | |
3853 | int *sources, unsigned long long *offsets, | |
3854 | int disks, int chunk, int level, int layout, | |
3855 | int dests, int *destfd, unsigned long long *destoffsets, | |
3856 | int part, int *degraded, | |
3857 | char *buf) | |
3858 | { | |
3859 | /* Backup 'blocks' sectors at 'offset' on each device of the array, | |
3860 | * to storage 'destfd' (offset 'destoffsets'), after first | |
3861 | * suspending IO. Then allow resync to continue | |
3862 | * over the suspended section. | |
3863 | * Use part 'part' of the backup-super-block. | |
3864 | */ | |
3865 | int odata = disks; | |
3866 | int rv = 0; | |
3867 | int i; | |
3868 | unsigned long long ll; | |
3869 | int new_degraded; | |
3870 | //printf("offset %llu\n", offset); | |
3871 | if (level >= 4) | |
3872 | odata--; | |
3873 | if (level == 6) | |
3874 | odata--; | |
3875 | ||
3876 | /* Check that array hasn't become degraded, else we might backup the wrong data */ | |
3877 | if (sysfs_get_ll(sra, NULL, "degraded", &ll) < 0) | |
3878 | return -1; /* FIXME this error is ignored */ | |
3879 | new_degraded = (int)ll; | |
3880 | if (new_degraded != *degraded) { | |
3881 | /* check each device to ensure it is still working */ | |
3882 | struct mdinfo *sd; | |
3883 | for (sd = sra->devs ; sd ; sd = sd->next) { | |
3884 | if (sd->disk.state & (1<<MD_DISK_FAULTY)) | |
3885 | continue; | |
3886 | if (sd->disk.state & (1<<MD_DISK_SYNC)) { | |
3887 | char sbuf[20]; | |
3888 | if (sysfs_get_str(sra, sd, "state", sbuf, 20) < 0 || | |
3889 | strstr(sbuf, "faulty") || | |
3890 | strstr(sbuf, "in_sync") == NULL) { | |
3891 | /* this device is dead */ | |
3892 | sd->disk.state = (1<<MD_DISK_FAULTY); | |
3893 | if (sd->disk.raid_disk >= 0 && | |
3894 | sources[sd->disk.raid_disk] >= 0) { | |
3895 | close(sources[sd->disk.raid_disk]); | |
3896 | sources[sd->disk.raid_disk] = -1; | |
3897 | } | |
3898 | } | |
3899 | } | |
3900 | } | |
3901 | *degraded = new_degraded; | |
3902 | } | |
3903 | if (part) { | |
3904 | bsb.arraystart2 = __cpu_to_le64(offset * odata); | |
3905 | bsb.length2 = __cpu_to_le64(stripes * (chunk/512) * odata); | |
3906 | } else { | |
3907 | bsb.arraystart = __cpu_to_le64(offset * odata); | |
3908 | bsb.length = __cpu_to_le64(stripes * (chunk/512) * odata); | |
3909 | } | |
3910 | if (part) | |
3911 | bsb.magic[15] = '2'; | |
3912 | for (i = 0; i < dests; i++) | |
3913 | if (part) | |
3914 | lseek64(destfd[i], destoffsets[i] + __le64_to_cpu(bsb.devstart2)*512, 0); | |
3915 | else | |
3916 | lseek64(destfd[i], destoffsets[i], 0); | |
3917 | ||
3918 | rv = save_stripes(sources, offsets, | |
3919 | disks, chunk, level, layout, | |
3920 | dests, destfd, | |
3921 | offset*512*odata, stripes * chunk * odata, | |
3922 | buf); | |
3923 | ||
3924 | if (rv) | |
3925 | return rv; | |
3926 | bsb.mtime = __cpu_to_le64(time(0)); | |
3927 | for (i = 0; i < dests; i++) { | |
3928 | bsb.devstart = __cpu_to_le64(destoffsets[i]/512); | |
3929 | ||
3930 | bsb.sb_csum = bsb_csum((char*)&bsb, ((char*)&bsb.sb_csum)-((char*)&bsb)); | |
3931 | if (memcmp(bsb.magic, "md_backup_data-2", 16) == 0) | |
3932 | bsb.sb_csum2 = bsb_csum((char*)&bsb, | |
3933 | ((char*)&bsb.sb_csum2)-((char*)&bsb)); | |
3934 | ||
3935 | rv = -1; | |
3936 | if ((unsigned long long)lseek64(destfd[i], destoffsets[i] - 4096, 0) | |
3937 | != destoffsets[i] - 4096) | |
3938 | break; | |
3939 | if (write(destfd[i], &bsb, 512) != 512) | |
3940 | break; | |
3941 | if (destoffsets[i] > 4096) { | |
3942 | if ((unsigned long long)lseek64(destfd[i], destoffsets[i]+stripes*chunk*odata, 0) != | |
3943 | destoffsets[i]+stripes*chunk*odata) | |
3944 | break; | |
3945 | if (write(destfd[i], &bsb, 512) != 512) | |
3946 | break; | |
3947 | } | |
3948 | fsync(destfd[i]); | |
3949 | rv = 0; | |
3950 | } | |
3951 | ||
3952 | return rv; | |
3953 | } | |
3954 | ||
3955 | /* in 2.6.30, the value reported by sync_completed can be | |
3956 | * less that it should be by one stripe. | |
3957 | * This only happens when reshape hits sync_max and pauses. | |
3958 | * So allow wait_backup to either extent sync_max further | |
3959 | * than strictly necessary, or return before the | |
3960 | * sync has got quite as far as we would really like. | |
3961 | * This is what 'blocks2' is for. | |
3962 | * The various caller give appropriate values so that | |
3963 | * every works. | |
3964 | */ | |
3965 | /* FIXME return value is often ignored */ | |
3966 | static int forget_backup(int dests, int *destfd, | |
3967 | unsigned long long *destoffsets, | |
3968 | int part) | |
3969 | { | |
3970 | /* | |
3971 | * Erase backup 'part' (which is 0 or 1) | |
3972 | */ | |
3973 | int i; | |
3974 | int rv; | |
3975 | ||
3976 | if (part) { | |
3977 | bsb.arraystart2 = __cpu_to_le64(0); | |
3978 | bsb.length2 = __cpu_to_le64(0); | |
3979 | } else { | |
3980 | bsb.arraystart = __cpu_to_le64(0); | |
3981 | bsb.length = __cpu_to_le64(0); | |
3982 | } | |
3983 | bsb.mtime = __cpu_to_le64(time(0)); | |
3984 | rv = 0; | |
3985 | for (i = 0; i < dests; i++) { | |
3986 | bsb.devstart = __cpu_to_le64(destoffsets[i]/512); | |
3987 | bsb.sb_csum = bsb_csum((char*)&bsb, ((char*)&bsb.sb_csum)-((char*)&bsb)); | |
3988 | if (memcmp(bsb.magic, "md_backup_data-2", 16) == 0) | |
3989 | bsb.sb_csum2 = bsb_csum((char*)&bsb, | |
3990 | ((char*)&bsb.sb_csum2)-((char*)&bsb)); | |
3991 | if ((unsigned long long)lseek64(destfd[i], destoffsets[i]-4096, 0) != | |
3992 | destoffsets[i]-4096) | |
3993 | rv = -1; | |
3994 | if (rv == 0 && | |
3995 | write(destfd[i], &bsb, 512) != 512) | |
3996 | rv = -1; | |
3997 | fsync(destfd[i]); | |
3998 | } | |
3999 | return rv; | |
4000 | } | |
4001 | ||
4002 | static void fail(char *msg) | |
4003 | { | |
4004 | int rv; | |
4005 | rv = (write(2, msg, strlen(msg)) != (int)strlen(msg)); | |
4006 | rv |= (write(2, "\n", 1) != 1); | |
4007 | exit(rv ? 1 : 2); | |
4008 | } | |
4009 | ||
4010 | static char *abuf, *bbuf; | |
4011 | static unsigned long long abuflen; | |
4012 | static void validate(int afd, int bfd, unsigned long long offset) | |
4013 | { | |
4014 | /* check that the data in the backup against the array. | |
4015 | * This is only used for regression testing and should not | |
4016 | * be used while the array is active | |
4017 | */ | |
4018 | if (afd < 0) | |
4019 | return; | |
4020 | lseek64(bfd, offset - 4096, 0); | |
4021 | if (read(bfd, &bsb2, 512) != 512) | |
4022 | fail("cannot read bsb"); | |
4023 | if (bsb2.sb_csum != bsb_csum((char*)&bsb2, | |
4024 | ((char*)&bsb2.sb_csum)-((char*)&bsb2))) | |
4025 | fail("first csum bad"); | |
4026 | if (memcmp(bsb2.magic, "md_backup_data", 14) != 0) | |
4027 | fail("magic is bad"); | |
4028 | if (memcmp(bsb2.magic, "md_backup_data-2", 16) == 0 && | |
4029 | bsb2.sb_csum2 != bsb_csum((char*)&bsb2, | |
4030 | ((char*)&bsb2.sb_csum2)-((char*)&bsb2))) | |
4031 | fail("second csum bad"); | |
4032 | ||
4033 | if (__le64_to_cpu(bsb2.devstart)*512 != offset) | |
4034 | fail("devstart is wrong"); | |
4035 | ||
4036 | if (bsb2.length) { | |
4037 | unsigned long long len = __le64_to_cpu(bsb2.length)*512; | |
4038 | ||
4039 | if (abuflen < len) { | |
4040 | free(abuf); | |
4041 | free(bbuf); | |
4042 | abuflen = len; | |
4043 | if (posix_memalign((void**)&abuf, 4096, abuflen) || | |
4044 | posix_memalign((void**)&bbuf, 4096, abuflen)) { | |
4045 | abuflen = 0; | |
4046 | /* just stop validating on mem-alloc failure */ | |
4047 | return; | |
4048 | } | |
4049 | } | |
4050 | ||
4051 | lseek64(bfd, offset, 0); | |
4052 | if ((unsigned long long)read(bfd, bbuf, len) != len) { | |
4053 | //printf("len %llu\n", len); | |
4054 | fail("read first backup failed"); | |
4055 | } | |
4056 | lseek64(afd, __le64_to_cpu(bsb2.arraystart)*512, 0); | |
4057 | if ((unsigned long long)read(afd, abuf, len) != len) | |
4058 | fail("read first from array failed"); | |
4059 | if (memcmp(bbuf, abuf, len) != 0) { | |
4060 | #if 0 | |
4061 | int i; | |
4062 | printf("offset=%llu len=%llu\n", | |
4063 | (unsigned long long)__le64_to_cpu(bsb2.arraystart)*512, len); | |
4064 | for (i=0; i<len; i++) | |
4065 | if (bbuf[i] != abuf[i]) { | |
4066 | printf("first diff byte %d\n", i); | |
4067 | break; | |
4068 | } | |
4069 | #endif | |
4070 | fail("data1 compare failed"); | |
4071 | } | |
4072 | } | |
4073 | if (bsb2.length2) { | |
4074 | unsigned long long len = __le64_to_cpu(bsb2.length2)*512; | |
4075 | ||
4076 | if (abuflen < len) { | |
4077 | free(abuf); | |
4078 | free(bbuf); | |
4079 | abuflen = len; | |
4080 | abuf = xmalloc(abuflen); | |
4081 | bbuf = xmalloc(abuflen); | |
4082 | } | |
4083 | ||
4084 | lseek64(bfd, offset+__le64_to_cpu(bsb2.devstart2)*512, 0); | |
4085 | if ((unsigned long long)read(bfd, bbuf, len) != len) | |
4086 | fail("read second backup failed"); | |
4087 | lseek64(afd, __le64_to_cpu(bsb2.arraystart2)*512, 0); | |
4088 | if ((unsigned long long)read(afd, abuf, len) != len) | |
4089 | fail("read second from array failed"); | |
4090 | if (memcmp(bbuf, abuf, len) != 0) | |
4091 | fail("data2 compare failed"); | |
4092 | } | |
4093 | } | |
4094 | ||
4095 | int child_monitor(int afd, struct mdinfo *sra, struct reshape *reshape, | |
4096 | struct supertype *st, unsigned long blocks, | |
4097 | int *fds, unsigned long long *offsets, | |
4098 | int dests, int *destfd, unsigned long long *destoffsets) | |
4099 | { | |
4100 | /* Monitor a reshape where backup is being performed using | |
4101 | * 'native' mechanism - either to a backup file, or | |
4102 | * to some space in a spare. | |
4103 | */ | |
4104 | char *buf; | |
4105 | int degraded = -1; | |
4106 | unsigned long long speed; | |
4107 | unsigned long long suspend_point, array_size; | |
4108 | unsigned long long backup_point, wait_point; | |
4109 | unsigned long long reshape_completed; | |
4110 | int done = 0; | |
4111 | int increasing = reshape->after.data_disks >= reshape->before.data_disks; | |
4112 | int part = 0; /* The next part of the backup area to fill. It may already | |
4113 | * be full, so we need to check */ | |
4114 | int level = reshape->level; | |
4115 | int layout = reshape->before.layout; | |
4116 | int data = reshape->before.data_disks; | |
4117 | int disks = reshape->before.data_disks + reshape->parity; | |
4118 | int chunk = sra->array.chunk_size; | |
4119 | struct mdinfo *sd; | |
4120 | unsigned long stripes; | |
4121 | int uuid[4]; | |
4122 | ||
4123 | /* set up the backup-super-block. This requires the | |
4124 | * uuid from the array. | |
4125 | */ | |
4126 | /* Find a superblock */ | |
4127 | for (sd = sra->devs; sd; sd = sd->next) { | |
4128 | char *dn; | |
4129 | int devfd; | |
4130 | int ok; | |
4131 | if (sd->disk.state & (1<<MD_DISK_FAULTY)) | |
4132 | continue; | |
4133 | dn = map_dev(sd->disk.major, sd->disk.minor, 1); | |
4134 | devfd = dev_open(dn, O_RDONLY); | |
4135 | if (devfd < 0) | |
4136 | continue; | |
4137 | ok = st->ss->load_super(st, devfd, NULL); | |
4138 | close(devfd); | |
4139 | if (ok == 0) | |
4140 | break; | |
4141 | } | |
4142 | if (!sd) { | |
4143 | pr_err("Cannot find a superblock\n"); | |
4144 | return 0; | |
4145 | } | |
4146 | ||
4147 | memset(&bsb, 0, 512); | |
4148 | memcpy(bsb.magic, "md_backup_data-1", 16); | |
4149 | st->ss->uuid_from_super(st, uuid); | |
4150 | memcpy(bsb.set_uuid, uuid, 16); | |
4151 | bsb.mtime = __cpu_to_le64(time(0)); | |
4152 | bsb.devstart2 = blocks; | |
4153 | ||
4154 | stripes = blocks / (sra->array.chunk_size/512) / | |
4155 | reshape->before.data_disks; | |
4156 | ||
4157 | if (posix_memalign((void**)&buf, 4096, disks * chunk)) | |
4158 | /* Don't start the 'reshape' */ | |
4159 | return 0; | |
4160 | if (reshape->before.data_disks == reshape->after.data_disks) { | |
4161 | sysfs_get_ll(sra, NULL, "sync_speed_min", &speed); | |
4162 | sysfs_set_num(sra, NULL, "sync_speed_min", 200000); | |
4163 | } | |
4164 | ||
4165 | if (increasing) { | |
4166 | array_size = sra->component_size * reshape->after.data_disks; | |
4167 | backup_point = sra->reshape_progress; | |
4168 | suspend_point = 0; | |
4169 | } else { | |
4170 | array_size = sra->component_size * reshape->before.data_disks; | |
4171 | backup_point = reshape->backup_blocks; | |
4172 | suspend_point = array_size; | |
4173 | } | |
4174 | ||
4175 | while (!done) { | |
4176 | int rv; | |
4177 | ||
4178 | /* Want to return as soon the oldest backup slot can | |
4179 | * be released as that allows us to start backing up | |
4180 | * some more, providing suspend_point has been | |
4181 | * advanced, which it should have. | |
4182 | */ | |
4183 | if (increasing) { | |
4184 | wait_point = array_size; | |
4185 | if (part == 0 && __le64_to_cpu(bsb.length) > 0) | |
4186 | wait_point = (__le64_to_cpu(bsb.arraystart) + | |
4187 | __le64_to_cpu(bsb.length)); | |
4188 | if (part == 1 && __le64_to_cpu(bsb.length2) > 0) | |
4189 | wait_point = (__le64_to_cpu(bsb.arraystart2) + | |
4190 | __le64_to_cpu(bsb.length2)); | |
4191 | } else { | |
4192 | wait_point = 0; | |
4193 | if (part == 0 && __le64_to_cpu(bsb.length) > 0) | |
4194 | wait_point = __le64_to_cpu(bsb.arraystart); | |
4195 | if (part == 1 && __le64_to_cpu(bsb.length2) > 0) | |
4196 | wait_point = __le64_to_cpu(bsb.arraystart2); | |
4197 | } | |
4198 | ||
4199 | rv = progress_reshape(sra, reshape, | |
4200 | backup_point, wait_point, | |
4201 | &suspend_point, &reshape_completed); | |
4202 | /* external metadata would need to ping_monitor here */ | |
4203 | sra->reshape_progress = reshape_completed; | |
4204 | ||
4205 | /* Clear any backup region that is before 'here' */ | |
4206 | if (increasing) { | |
4207 | if (__le64_to_cpu(bsb.length) > 0 && | |
4208 | reshape_completed >= (__le64_to_cpu(bsb.arraystart) + | |
4209 | __le64_to_cpu(bsb.length))) | |
4210 | forget_backup(dests, destfd, | |
4211 | destoffsets, 0); | |
4212 | if (__le64_to_cpu(bsb.length2) > 0 && | |
4213 | reshape_completed >= (__le64_to_cpu(bsb.arraystart2) + | |
4214 | __le64_to_cpu(bsb.length2))) | |
4215 | forget_backup(dests, destfd, | |
4216 | destoffsets, 1); | |
4217 | } else { | |
4218 | if (__le64_to_cpu(bsb.length) > 0 && | |
4219 | reshape_completed <= (__le64_to_cpu(bsb.arraystart))) | |
4220 | forget_backup(dests, destfd, | |
4221 | destoffsets, 0); | |
4222 | if (__le64_to_cpu(bsb.length2) > 0 && | |
4223 | reshape_completed <= (__le64_to_cpu(bsb.arraystart2))) | |
4224 | forget_backup(dests, destfd, | |
4225 | destoffsets, 1); | |
4226 | } | |
4227 | ||
4228 | if (rv < 0) { | |
4229 | if (rv == -1) | |
4230 | done = 1; | |
4231 | break; | |
4232 | } | |
4233 | if (rv == 0 && increasing && !st->ss->external) { | |
4234 | /* No longer need to monitor this reshape */ | |
4235 | done = 1; | |
4236 | break; | |
4237 | } | |
4238 | ||
4239 | while (rv) { | |
4240 | unsigned long long offset; | |
4241 | unsigned long actual_stripes; | |
4242 | /* Need to backup some data. | |
4243 | * If 'part' is not used and the desired | |
4244 | * backup size is suspended, do a backup, | |
4245 | * then consider the next part. | |
4246 | */ | |
4247 | /* Check that 'part' is unused */ | |
4248 | if (part == 0 && __le64_to_cpu(bsb.length) != 0) | |
4249 | break; | |
4250 | if (part == 1 && __le64_to_cpu(bsb.length2) != 0) | |
4251 | break; | |
4252 | ||
4253 | offset = backup_point / data; | |
4254 | actual_stripes = stripes; | |
4255 | if (increasing) { | |
4256 | if (offset + actual_stripes * (chunk/512) > | |
4257 | sra->component_size) | |
4258 | actual_stripes = ((sra->component_size - offset) | |
4259 | / (chunk/512)); | |
4260 | if (offset + actual_stripes * (chunk/512) > | |
4261 | suspend_point/data) | |
4262 | break; | |
4263 | } else { | |
4264 | if (offset < actual_stripes * (chunk/512)) | |
4265 | actual_stripes = offset / (chunk/512); | |
4266 | offset -= actual_stripes * (chunk/512); | |
4267 | if (offset < suspend_point/data) | |
4268 | break; | |
4269 | } | |
4270 | if (actual_stripes == 0) | |
4271 | break; | |
4272 | grow_backup(sra, offset, actual_stripes, | |
4273 | fds, offsets, | |
4274 | disks, chunk, level, layout, | |
4275 | dests, destfd, destoffsets, | |
4276 | part, °raded, buf); | |
4277 | validate(afd, destfd[0], destoffsets[0]); | |
4278 | /* record where 'part' is up to */ | |
4279 | part = !part; | |
4280 | if (increasing) | |
4281 | backup_point += actual_stripes * (chunk/512) * data; | |
4282 | else | |
4283 | backup_point -= actual_stripes * (chunk/512) * data; | |
4284 | } | |
4285 | } | |
4286 | ||
4287 | /* FIXME maybe call progress_reshape one more time instead */ | |
4288 | abort_reshape(sra); /* remove any remaining suspension */ | |
4289 | if (reshape->before.data_disks == reshape->after.data_disks) | |
4290 | sysfs_set_num(sra, NULL, "sync_speed_min", speed); | |
4291 | free(buf); | |
4292 | return done; | |
4293 | } | |
4294 | ||
4295 | /* | |
4296 | * If any spare contains md_back_data-1 which is recent wrt mtime, | |
4297 | * write that data into the array and update the super blocks with | |
4298 | * the new reshape_progress | |
4299 | */ | |
4300 | int Grow_restart(struct supertype *st, struct mdinfo *info, int *fdlist, int cnt, | |
4301 | char *backup_file, int verbose) | |
4302 | { | |
4303 | int i, j; | |
4304 | int old_disks; | |
4305 | unsigned long long *offsets; | |
4306 | unsigned long long nstripe, ostripe; | |
4307 | int ndata, odata; | |
4308 | ||
4309 | odata = info->array.raid_disks - info->delta_disks - 1; | |
4310 | if (info->array.level == 6) odata--; /* number of data disks */ | |
4311 | ndata = info->array.raid_disks - 1; | |
4312 | if (info->new_level == 6) ndata--; | |
4313 | ||
4314 | old_disks = info->array.raid_disks - info->delta_disks; | |
4315 | ||
4316 | if (info->delta_disks <= 0) | |
4317 | /* Didn't grow, so the backup file must have | |
4318 | * been used | |
4319 | */ | |
4320 | old_disks = cnt; | |
4321 | for (i=old_disks-(backup_file?1:0); i<cnt; i++) { | |
4322 | struct mdinfo dinfo; | |
4323 | int fd; | |
4324 | int bsbsize; | |
4325 | char *devname, namebuf[20]; | |
4326 | unsigned long long lo, hi; | |
4327 | ||
4328 | /* This was a spare and may have some saved data on it. | |
4329 | * Load the superblock, find and load the | |
4330 | * backup_super_block. | |
4331 | * If either fail, go on to next device. | |
4332 | * If the backup contains no new info, just return | |
4333 | * else restore data and update all superblocks | |
4334 | */ | |
4335 | if (i == old_disks-1) { | |
4336 | fd = open(backup_file, O_RDONLY); | |
4337 | if (fd<0) { | |
4338 | pr_err("backup file %s inaccessible: %s\n", | |
4339 | backup_file, strerror(errno)); | |
4340 | continue; | |
4341 | } | |
4342 | devname = backup_file; | |
4343 | } else { | |
4344 | fd = fdlist[i]; | |
4345 | if (fd < 0) | |
4346 | continue; | |
4347 | if (st->ss->load_super(st, fd, NULL)) | |
4348 | continue; | |
4349 | ||
4350 | st->ss->getinfo_super(st, &dinfo, NULL); | |
4351 | st->ss->free_super(st); | |
4352 | ||
4353 | if (lseek64(fd, | |
4354 | (dinfo.data_offset + dinfo.component_size - 8) <<9, | |
4355 | 0) < 0) { | |
4356 | pr_err("Cannot seek on device %d\n", i); | |
4357 | continue; /* Cannot seek */ | |
4358 | } | |
4359 | sprintf(namebuf, "device-%d", i); | |
4360 | devname = namebuf; | |
4361 | } | |
4362 | if (read(fd, &bsb, sizeof(bsb)) != sizeof(bsb)) { | |
4363 | if (verbose) | |
4364 | pr_err("Cannot read from %s\n", devname); | |
4365 | continue; /* Cannot read */ | |
4366 | } | |
4367 | if (memcmp(bsb.magic, "md_backup_data-1", 16) != 0 && | |
4368 | memcmp(bsb.magic, "md_backup_data-2", 16) != 0) { | |
4369 | if (verbose) | |
4370 | pr_err("No backup metadata on %s\n", devname); | |
4371 | continue; | |
4372 | } | |
4373 | if (bsb.sb_csum != bsb_csum((char*)&bsb, ((char*)&bsb.sb_csum)-((char*)&bsb))) { | |
4374 | if (verbose) | |
4375 | pr_err("Bad backup-metadata checksum on %s\n", devname); | |
4376 | continue; /* bad checksum */ | |
4377 | } | |
4378 | if (memcmp(bsb.magic, "md_backup_data-2", 16) == 0 && | |
4379 | bsb.sb_csum2 != bsb_csum((char*)&bsb, ((char*)&bsb.sb_csum2)-((char*)&bsb))) { | |
4380 | if (verbose) | |
4381 | pr_err("Bad backup-metadata checksum2 on %s\n", devname); | |
4382 | continue; /* Bad second checksum */ | |
4383 | } | |
4384 | if (memcmp(bsb.set_uuid,info->uuid, 16) != 0) { | |
4385 | if (verbose) | |
4386 | pr_err("Wrong uuid on backup-metadata on %s\n", devname); | |
4387 | continue; /* Wrong uuid */ | |
4388 | } | |
4389 | ||
4390 | /* array utime and backup-mtime should be updated at much the same time, but it seems that | |
4391 | * sometimes they aren't... So allow considerable flexability in matching, and allow | |
4392 | * this test to be overridden by an environment variable. | |
4393 | */ | |
4394 | if (info->array.utime > (int)__le64_to_cpu(bsb.mtime) + 2*60*60 || | |
4395 | info->array.utime < (int)__le64_to_cpu(bsb.mtime) - 10*60) { | |
4396 | if (check_env("MDADM_GROW_ALLOW_OLD")) { | |
4397 | pr_err("accepting backup with timestamp %lu " | |
4398 | "for array with timestamp %lu\n", | |
4399 | (unsigned long)__le64_to_cpu(bsb.mtime), | |
4400 | (unsigned long)info->array.utime); | |
4401 | } else { | |
4402 | pr_err("too-old timestamp on backup-metadata on %s\n", devname); | |
4403 | pr_err("If you think it is should be safe, try 'export MDADM_GROW_ALLOW_OLD=1'\n"); | |
4404 | continue; /* time stamp is too bad */ | |
4405 | } | |
4406 | } | |
4407 | ||
4408 | if (bsb.magic[15] == '1') { | |
4409 | if (bsb.length == 0) | |
4410 | continue; | |
4411 | if (info->delta_disks >= 0) { | |
4412 | /* reshape_progress is increasing */ | |
4413 | if (__le64_to_cpu(bsb.arraystart) | |
4414 | + __le64_to_cpu(bsb.length) | |
4415 | < info->reshape_progress) { | |
4416 | nonew: | |
4417 | if (verbose) | |
4418 | pr_err("backup-metadata found on %s but is not needed\n", devname); | |
4419 | continue; /* No new data here */ | |
4420 | } | |
4421 | } else { | |
4422 | /* reshape_progress is decreasing */ | |
4423 | if (__le64_to_cpu(bsb.arraystart) >= | |
4424 | info->reshape_progress) | |
4425 | goto nonew; /* No new data here */ | |
4426 | } | |
4427 | } else { | |
4428 | if (bsb.length == 0 && bsb.length2 == 0) | |
4429 | continue; | |
4430 | if (info->delta_disks >= 0) { | |
4431 | /* reshape_progress is increasing */ | |
4432 | if ((__le64_to_cpu(bsb.arraystart) | |
4433 | + __le64_to_cpu(bsb.length) | |
4434 | < info->reshape_progress) | |
4435 | && | |
4436 | (__le64_to_cpu(bsb.arraystart2) | |
4437 | + __le64_to_cpu(bsb.length2) | |
4438 | < info->reshape_progress)) | |
4439 | goto nonew; /* No new data here */ | |
4440 | } else { | |
4441 | /* reshape_progress is decreasing */ | |
4442 | if (__le64_to_cpu(bsb.arraystart) >= | |
4443 | info->reshape_progress && | |
4444 | __le64_to_cpu(bsb.arraystart2) >= | |
4445 | info->reshape_progress) | |
4446 | goto nonew; /* No new data here */ | |
4447 | } | |
4448 | } | |
4449 | if (lseek64(fd, __le64_to_cpu(bsb.devstart)*512, 0)< 0) { | |
4450 | second_fail: | |
4451 | if (verbose) | |
4452 | pr_err("Failed to verify secondary backup-metadata block on %s\n", | |
4453 | devname); | |
4454 | continue; /* Cannot seek */ | |
4455 | } | |
4456 | /* There should be a duplicate backup superblock 4k before here */ | |
4457 | if (lseek64(fd, -4096, 1) < 0 || | |
4458 | read(fd, &bsb2, sizeof(bsb2)) != sizeof(bsb2)) | |
4459 | goto second_fail; /* Cannot find leading superblock */ | |
4460 | if (bsb.magic[15] == '1') | |
4461 | bsbsize = offsetof(struct mdp_backup_super, pad1); | |
4462 | else | |
4463 | bsbsize = offsetof(struct mdp_backup_super, pad); | |
4464 | if (memcmp(&bsb2, &bsb, bsbsize) != 0) | |
4465 | goto second_fail; /* Cannot find leading superblock */ | |
4466 | ||
4467 | /* Now need the data offsets for all devices. */ | |
4468 | offsets = xmalloc(sizeof(*offsets)*info->array.raid_disks); | |
4469 | for(j=0; j<info->array.raid_disks; j++) { | |
4470 | if (fdlist[j] < 0) | |
4471 | continue; | |
4472 | if (st->ss->load_super(st, fdlist[j], NULL)) | |
4473 | /* FIXME should be this be an error */ | |
4474 | continue; | |
4475 | st->ss->getinfo_super(st, &dinfo, NULL); | |
4476 | st->ss->free_super(st); | |
4477 | offsets[j] = dinfo.data_offset * 512; | |
4478 | } | |
4479 | printf(Name ": restoring critical section\n"); | |
4480 | ||
4481 | if (restore_stripes(fdlist, offsets, | |
4482 | info->array.raid_disks, | |
4483 | info->new_chunk, | |
4484 | info->new_level, | |
4485 | info->new_layout, | |
4486 | fd, __le64_to_cpu(bsb.devstart)*512, | |
4487 | __le64_to_cpu(bsb.arraystart)*512, | |
4488 | __le64_to_cpu(bsb.length)*512, NULL)) { | |
4489 | /* didn't succeed, so giveup */ | |
4490 | if (verbose) | |
4491 | pr_err("Error restoring backup from %s\n", | |
4492 | devname); | |
4493 | free(offsets); | |
4494 | return 1; | |
4495 | } | |
4496 | ||
4497 | if (bsb.magic[15] == '2' && | |
4498 | restore_stripes(fdlist, offsets, | |
4499 | info->array.raid_disks, | |
4500 | info->new_chunk, | |
4501 | info->new_level, | |
4502 | info->new_layout, | |
4503 | fd, __le64_to_cpu(bsb.devstart)*512 + | |
4504 | __le64_to_cpu(bsb.devstart2)*512, | |
4505 | __le64_to_cpu(bsb.arraystart2)*512, | |
4506 | __le64_to_cpu(bsb.length2)*512, NULL)) { | |
4507 | /* didn't succeed, so giveup */ | |
4508 | if (verbose) | |
4509 | pr_err("Error restoring second backup from %s\n", | |
4510 | devname); | |
4511 | free(offsets); | |
4512 | return 1; | |
4513 | } | |
4514 | ||
4515 | free(offsets); | |
4516 | ||
4517 | /* Ok, so the data is restored. Let's update those superblocks. */ | |
4518 | ||
4519 | lo = hi = 0; | |
4520 | if (bsb.length) { | |
4521 | lo = __le64_to_cpu(bsb.arraystart); | |
4522 | hi = lo + __le64_to_cpu(bsb.length); | |
4523 | } | |
4524 | if (bsb.magic[15] == '2' && bsb.length2) { | |
4525 | unsigned long long lo1, hi1; | |
4526 | lo1 = __le64_to_cpu(bsb.arraystart2); | |
4527 | hi1 = lo1 + __le64_to_cpu(bsb.length2); | |
4528 | if (lo == hi) { | |
4529 | lo = lo1; | |
4530 | hi = hi1; | |
4531 | } else if (lo < lo1) | |
4532 | hi = hi1; | |
4533 | else | |
4534 | lo = lo1; | |
4535 | } | |
4536 | if (lo < hi && | |
4537 | (info->reshape_progress < lo || | |
4538 | info->reshape_progress > hi)) | |
4539 | /* backup does not affect reshape_progress*/ ; | |
4540 | else if (info->delta_disks >= 0) { | |
4541 | info->reshape_progress = __le64_to_cpu(bsb.arraystart) + | |
4542 | __le64_to_cpu(bsb.length); | |
4543 | if (bsb.magic[15] == '2') { | |
4544 | unsigned long long p2 = __le64_to_cpu(bsb.arraystart2) + | |
4545 | __le64_to_cpu(bsb.length2); | |
4546 | if (p2 > info->reshape_progress) | |
4547 | info->reshape_progress = p2; | |
4548 | } | |
4549 | } else { | |
4550 | info->reshape_progress = __le64_to_cpu(bsb.arraystart); | |
4551 | if (bsb.magic[15] == '2') { | |
4552 | unsigned long long p2 = __le64_to_cpu(bsb.arraystart2); | |
4553 | if (p2 < info->reshape_progress) | |
4554 | info->reshape_progress = p2; | |
4555 | } | |
4556 | } | |
4557 | for (j=0; j<info->array.raid_disks; j++) { | |
4558 | if (fdlist[j] < 0) | |
4559 | continue; | |
4560 | if (st->ss->load_super(st, fdlist[j], NULL)) | |
4561 | continue; | |
4562 | st->ss->getinfo_super(st, &dinfo, NULL); | |
4563 | dinfo.reshape_progress = info->reshape_progress; | |
4564 | st->ss->update_super(st, &dinfo, | |
4565 | "_reshape_progress", | |
4566 | NULL,0, 0, NULL); | |
4567 | st->ss->store_super(st, fdlist[j]); | |
4568 | st->ss->free_super(st); | |
4569 | } | |
4570 | return 0; | |
4571 | } | |
4572 | /* Didn't find any backup data, try to see if any | |
4573 | * was needed. | |
4574 | */ | |
4575 | if (info->delta_disks < 0) { | |
4576 | /* When shrinking, the critical section is at the end. | |
4577 | * So see if we are before the critical section. | |
4578 | */ | |
4579 | unsigned long long first_block; | |
4580 | nstripe = ostripe = 0; | |
4581 | first_block = 0; | |
4582 | while (ostripe >= nstripe) { | |
4583 | ostripe += info->array.chunk_size / 512; | |
4584 | first_block = ostripe * odata; | |
4585 | nstripe = first_block / ndata / (info->new_chunk/512) * | |
4586 | (info->new_chunk/512); | |
4587 | } | |
4588 | ||
4589 | if (info->reshape_progress >= first_block) | |
4590 | return 0; | |
4591 | } | |
4592 | if (info->delta_disks > 0) { | |
4593 | /* See if we are beyond the critical section. */ | |
4594 | unsigned long long last_block; | |
4595 | nstripe = ostripe = 0; | |
4596 | last_block = 0; | |
4597 | while (nstripe >= ostripe) { | |
4598 | nstripe += info->new_chunk / 512; | |
4599 | last_block = nstripe * ndata; | |
4600 | ostripe = last_block / odata / (info->array.chunk_size/512) * | |
4601 | (info->array.chunk_size/512); | |
4602 | } | |
4603 | ||
4604 | if (info->reshape_progress >= last_block) | |
4605 | return 0; | |
4606 | } | |
4607 | /* needed to recover critical section! */ | |
4608 | if (verbose) | |
4609 | pr_err("Failed to find backup of critical section\n"); | |
4610 | return 1; | |
4611 | } | |
4612 | ||
4613 | int Grow_continue_command(char *devname, int fd, | |
4614 | char *backup_file, int verbose) | |
4615 | { | |
4616 | int ret_val = 0; | |
4617 | struct supertype *st = NULL; | |
4618 | struct mdinfo *content = NULL; | |
4619 | struct mdinfo array; | |
4620 | char *subarray = NULL; | |
4621 | struct mdinfo *cc = NULL; | |
4622 | struct mdstat_ent *mdstat = NULL; | |
4623 | int cfd = -1; | |
4624 | int fd2 = -1; | |
4625 | ||
4626 | dprintf("Grow continue from command line called for %s\n", | |
4627 | devname); | |
4628 | ||
4629 | st = super_by_fd(fd, &subarray); | |
4630 | if (!st || !st->ss) { | |
4631 | pr_err("Unable to determine metadata format for %s\n", | |
4632 | devname); | |
4633 | return 1; | |
4634 | } | |
4635 | dprintf("Grow continue is run for "); | |
4636 | if (st->ss->external == 0) { | |
4637 | int d; | |
4638 | dprintf("native array (%s)\n", devname); | |
4639 | if (ioctl(fd, GET_ARRAY_INFO, &array.array) < 0) { | |
4640 | pr_err("%s is not an active md array -" | |
4641 | " aborting\n", devname); | |
4642 | ret_val = 1; | |
4643 | goto Grow_continue_command_exit; | |
4644 | } | |
4645 | content = &array; | |
4646 | /* Need to load a superblock. | |
4647 | * FIXME we should really get what we need from | |
4648 | * sysfs | |
4649 | */ | |
4650 | for (d = 0; d < MAX_DISKS; d++) { | |
4651 | mdu_disk_info_t disk; | |
4652 | char *dv; | |
4653 | int err; | |
4654 | disk.number = d; | |
4655 | if (ioctl(fd, GET_DISK_INFO, &disk) < 0) | |
4656 | continue; | |
4657 | if (disk.major == 0 && disk.minor == 0) | |
4658 | continue; | |
4659 | if ((disk.state & (1 << MD_DISK_ACTIVE)) == 0) | |
4660 | continue; | |
4661 | dv = map_dev(disk.major, disk.minor, 1); | |
4662 | if (!dv) | |
4663 | continue; | |
4664 | fd2 = dev_open(dv, O_RDONLY); | |
4665 | if (fd2 < 0) | |
4666 | continue; | |
4667 | err = st->ss->load_super(st, fd2, NULL); | |
4668 | close(fd2); | |
4669 | if (err) | |
4670 | continue; | |
4671 | break; | |
4672 | } | |
4673 | if (d == MAX_DISKS) { | |
4674 | pr_err("Unable to load metadata for %s\n", | |
4675 | devname); | |
4676 | ret_val = 1; | |
4677 | goto Grow_continue_command_exit; | |
4678 | } | |
4679 | st->ss->getinfo_super(st, content, NULL); | |
4680 | } else { | |
4681 | char *container; | |
4682 | ||
4683 | if (subarray) { | |
4684 | dprintf("subarray (%s)\n", subarray); | |
4685 | container = st->container_devnm; | |
4686 | cfd = open_dev_excl(st->container_devnm); | |
4687 | } else { | |
4688 | container = st->devnm; | |
4689 | close(fd); | |
4690 | cfd = open_dev_excl(st->devnm); | |
4691 | dprintf("container (%s)\n", container); | |
4692 | fd = cfd; | |
4693 | } | |
4694 | if (cfd < 0) { | |
4695 | pr_err("Unable to open container " | |
4696 | "for %s\n", devname); | |
4697 | ret_val = 1; | |
4698 | goto Grow_continue_command_exit; | |
4699 | } | |
4700 | ||
4701 | /* find in container array under reshape | |
4702 | */ | |
4703 | ret_val = st->ss->load_container(st, cfd, NULL); | |
4704 | if (ret_val) { | |
4705 | pr_err("Cannot read superblock for %s\n", | |
4706 | devname); | |
4707 | ret_val = 1; | |
4708 | goto Grow_continue_command_exit; | |
4709 | } | |
4710 | ||
4711 | cc = st->ss->container_content(st, subarray); | |
4712 | for (content = cc; content ; content = content->next) { | |
4713 | char *array; | |
4714 | int allow_reshape = 1; | |
4715 | ||
4716 | if (content->reshape_active == 0) | |
4717 | continue; | |
4718 | /* The decision about array or container wide | |
4719 | * reshape is taken in Grow_continue based | |
4720 | * content->reshape_active state, therefore we | |
4721 | * need to check_reshape based on | |
4722 | * reshape_active and subarray name | |
4723 | */ | |
4724 | if (content->array.state & (1<<MD_SB_BLOCK_VOLUME)) | |
4725 | allow_reshape = 0; | |
4726 | if (content->reshape_active == CONTAINER_RESHAPE && | |
4727 | (content->array.state | |
4728 | & (1<<MD_SB_BLOCK_CONTAINER_RESHAPE))) | |
4729 | allow_reshape = 0; | |
4730 | ||
4731 | if (!allow_reshape) { | |
4732 | pr_err("cannot continue reshape of an array" | |
4733 | " in container with unsupported" | |
4734 | " metadata: %s(%s)\n", | |
4735 | devname, container); | |
4736 | ret_val = 1; | |
4737 | goto Grow_continue_command_exit; | |
4738 | } | |
4739 | ||
4740 | array = strchr(content->text_version+1, '/')+1; | |
4741 | mdstat = mdstat_by_subdev(array, container); | |
4742 | if (!mdstat) | |
4743 | continue; | |
4744 | if (mdstat->active == 0) { | |
4745 | pr_err("Skipping inactive array %s.\n", | |
4746 | mdstat->devnm); | |
4747 | free_mdstat(mdstat); | |
4748 | mdstat = NULL; | |
4749 | continue; | |
4750 | } | |
4751 | break; | |
4752 | } | |
4753 | if (!content) { | |
4754 | pr_err("Unable to determine reshaped " | |
4755 | "array for %s\n", devname); | |
4756 | ret_val = 1; | |
4757 | goto Grow_continue_command_exit; | |
4758 | } | |
4759 | fd2 = open_dev(mdstat->devnm); | |
4760 | if (fd2 < 0) { | |
4761 | pr_err("cannot open (%s)\n", mdstat->devnm); | |
4762 | ret_val = 1; | |
4763 | goto Grow_continue_command_exit; | |
4764 | } | |
4765 | ||
4766 | sysfs_init(content, fd2, mdstat->devnm); | |
4767 | ||
4768 | /* start mdmon in case it is not running | |
4769 | */ | |
4770 | if (!mdmon_running(container)) | |
4771 | start_mdmon(container); | |
4772 | ping_monitor(container); | |
4773 | ||
4774 | if (mdmon_running(container)) | |
4775 | st->update_tail = &st->updates; | |
4776 | else { | |
4777 | pr_err("No mdmon found. " | |
4778 | "Grow cannot continue.\n"); | |
4779 | ret_val = 1; | |
4780 | goto Grow_continue_command_exit; | |
4781 | } | |
4782 | } | |
4783 | ||
4784 | /* verify that array under reshape is started from | |
4785 | * correct position | |
4786 | */ | |
4787 | if (verify_reshape_position(content, content->array.level) < 0) { | |
4788 | ret_val = 1; | |
4789 | goto Grow_continue_command_exit; | |
4790 | } | |
4791 | ||
4792 | /* continue reshape | |
4793 | */ | |
4794 | ret_val = Grow_continue(fd, st, content, backup_file, 0); | |
4795 | ||
4796 | Grow_continue_command_exit: | |
4797 | if (fd2 > -1) | |
4798 | close(fd2); | |
4799 | if (cfd > -1) | |
4800 | close(cfd); | |
4801 | st->ss->free_super(st); | |
4802 | free_mdstat(mdstat); | |
4803 | sysfs_free(cc); | |
4804 | free(subarray); | |
4805 | ||
4806 | return ret_val; | |
4807 | } | |
4808 | ||
4809 | int Grow_continue(int mdfd, struct supertype *st, struct mdinfo *info, | |
4810 | char *backup_file, int freeze_reshape) | |
4811 | { | |
4812 | int ret_val = 2; | |
4813 | ||
4814 | if (!info->reshape_active) | |
4815 | return ret_val; | |
4816 | ||
4817 | if (st->ss->external) { | |
4818 | int cfd = open_dev(st->container_devnm); | |
4819 | ||
4820 | if (cfd < 0) | |
4821 | return 1; | |
4822 | ||
4823 | st->ss->load_container(st, cfd, st->container_devnm); | |
4824 | close(cfd); | |
4825 | ret_val = reshape_container(st->container_devnm, NULL, mdfd, | |
4826 | st, info, 0, backup_file, | |
4827 | 0, | |
4828 | 1 | info->reshape_active, | |
4829 | freeze_reshape); | |
4830 | } else | |
4831 | ret_val = reshape_array(NULL, mdfd, "array", st, info, 1, | |
4832 | NULL, 0ULL, backup_file, 0, 0, | |
4833 | 1 | info->reshape_active, | |
4834 | freeze_reshape); | |
4835 | ||
4836 | return ret_val; | |
4837 | } |