]> git.ipfire.org Git - thirdparty/xfsprogs-dev.git/blob - libxfs/xfs_bmap_btree.c
xfs: rename flist/free_list to dfops
[thirdparty/xfsprogs-dev.git] / libxfs / xfs_bmap_btree.c
1 /*
2 * Copyright (c) 2000-2003,2005 Silicon Graphics, Inc.
3 * All Rights Reserved.
4 *
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of the GNU General Public License as
7 * published by the Free Software Foundation.
8 *
9 * This program is distributed in the hope that it would be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
13 *
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write the Free Software Foundation,
16 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
17 */
18 #include "libxfs_priv.h"
19 #include "xfs_fs.h"
20 #include "xfs_shared.h"
21 #include "xfs_format.h"
22 #include "xfs_log_format.h"
23 #include "xfs_trans_resv.h"
24 #include "xfs_bit.h"
25 #include "xfs_mount.h"
26 #include "xfs_defer.h"
27 #include "xfs_inode.h"
28 #include "xfs_trans.h"
29 #include "xfs_alloc.h"
30 #include "xfs_btree.h"
31 #include "xfs_bmap_btree.h"
32 #include "xfs_bmap.h"
33 #include "xfs_trace.h"
34 #include "xfs_cksum.h"
35
36 /*
37 * Determine the extent state.
38 */
39 /* ARGSUSED */
40 STATIC xfs_exntst_t
41 xfs_extent_state(
42 xfs_filblks_t blks,
43 int extent_flag)
44 {
45 if (extent_flag) {
46 ASSERT(blks != 0); /* saved for DMIG */
47 return XFS_EXT_UNWRITTEN;
48 }
49 return XFS_EXT_NORM;
50 }
51
52 /*
53 * Convert on-disk form of btree root to in-memory form.
54 */
55 void
56 xfs_bmdr_to_bmbt(
57 struct xfs_inode *ip,
58 xfs_bmdr_block_t *dblock,
59 int dblocklen,
60 struct xfs_btree_block *rblock,
61 int rblocklen)
62 {
63 struct xfs_mount *mp = ip->i_mount;
64 int dmxr;
65 xfs_bmbt_key_t *fkp;
66 __be64 *fpp;
67 xfs_bmbt_key_t *tkp;
68 __be64 *tpp;
69
70 if (xfs_sb_version_hascrc(&mp->m_sb))
71 xfs_btree_init_block_int(mp, rblock, XFS_BUF_DADDR_NULL,
72 XFS_BMAP_CRC_MAGIC, 0, 0, ip->i_ino,
73 XFS_BTREE_LONG_PTRS | XFS_BTREE_CRC_BLOCKS);
74 else
75 xfs_btree_init_block_int(mp, rblock, XFS_BUF_DADDR_NULL,
76 XFS_BMAP_MAGIC, 0, 0, ip->i_ino,
77 XFS_BTREE_LONG_PTRS);
78
79 rblock->bb_level = dblock->bb_level;
80 ASSERT(be16_to_cpu(rblock->bb_level) > 0);
81 rblock->bb_numrecs = dblock->bb_numrecs;
82 dmxr = xfs_bmdr_maxrecs(dblocklen, 0);
83 fkp = XFS_BMDR_KEY_ADDR(dblock, 1);
84 tkp = XFS_BMBT_KEY_ADDR(mp, rblock, 1);
85 fpp = XFS_BMDR_PTR_ADDR(dblock, 1, dmxr);
86 tpp = XFS_BMAP_BROOT_PTR_ADDR(mp, rblock, 1, rblocklen);
87 dmxr = be16_to_cpu(dblock->bb_numrecs);
88 memcpy(tkp, fkp, sizeof(*fkp) * dmxr);
89 memcpy(tpp, fpp, sizeof(*fpp) * dmxr);
90 }
91
92 /*
93 * Convert a compressed bmap extent record to an uncompressed form.
94 * This code must be in sync with the routines xfs_bmbt_get_startoff,
95 * xfs_bmbt_get_startblock, xfs_bmbt_get_blockcount and xfs_bmbt_get_state.
96 */
97 STATIC void
98 __xfs_bmbt_get_all(
99 __uint64_t l0,
100 __uint64_t l1,
101 xfs_bmbt_irec_t *s)
102 {
103 int ext_flag;
104 xfs_exntst_t st;
105
106 ext_flag = (int)(l0 >> (64 - BMBT_EXNTFLAG_BITLEN));
107 s->br_startoff = ((xfs_fileoff_t)l0 &
108 xfs_mask64lo(64 - BMBT_EXNTFLAG_BITLEN)) >> 9;
109 s->br_startblock = (((xfs_fsblock_t)l0 & xfs_mask64lo(9)) << 43) |
110 (((xfs_fsblock_t)l1) >> 21);
111 s->br_blockcount = (xfs_filblks_t)(l1 & xfs_mask64lo(21));
112 /* This is xfs_extent_state() in-line */
113 if (ext_flag) {
114 ASSERT(s->br_blockcount != 0); /* saved for DMIG */
115 st = XFS_EXT_UNWRITTEN;
116 } else
117 st = XFS_EXT_NORM;
118 s->br_state = st;
119 }
120
121 void
122 xfs_bmbt_get_all(
123 xfs_bmbt_rec_host_t *r,
124 xfs_bmbt_irec_t *s)
125 {
126 __xfs_bmbt_get_all(r->l0, r->l1, s);
127 }
128
129 /*
130 * Extract the blockcount field from an in memory bmap extent record.
131 */
132 xfs_filblks_t
133 xfs_bmbt_get_blockcount(
134 xfs_bmbt_rec_host_t *r)
135 {
136 return (xfs_filblks_t)(r->l1 & xfs_mask64lo(21));
137 }
138
139 /*
140 * Extract the startblock field from an in memory bmap extent record.
141 */
142 xfs_fsblock_t
143 xfs_bmbt_get_startblock(
144 xfs_bmbt_rec_host_t *r)
145 {
146 return (((xfs_fsblock_t)r->l0 & xfs_mask64lo(9)) << 43) |
147 (((xfs_fsblock_t)r->l1) >> 21);
148 }
149
150 /*
151 * Extract the startoff field from an in memory bmap extent record.
152 */
153 xfs_fileoff_t
154 xfs_bmbt_get_startoff(
155 xfs_bmbt_rec_host_t *r)
156 {
157 return ((xfs_fileoff_t)r->l0 &
158 xfs_mask64lo(64 - BMBT_EXNTFLAG_BITLEN)) >> 9;
159 }
160
161 xfs_exntst_t
162 xfs_bmbt_get_state(
163 xfs_bmbt_rec_host_t *r)
164 {
165 int ext_flag;
166
167 ext_flag = (int)((r->l0) >> (64 - BMBT_EXNTFLAG_BITLEN));
168 return xfs_extent_state(xfs_bmbt_get_blockcount(r),
169 ext_flag);
170 }
171
172 /*
173 * Extract the blockcount field from an on disk bmap extent record.
174 */
175 xfs_filblks_t
176 xfs_bmbt_disk_get_blockcount(
177 xfs_bmbt_rec_t *r)
178 {
179 return (xfs_filblks_t)(be64_to_cpu(r->l1) & xfs_mask64lo(21));
180 }
181
182 /*
183 * Extract the startoff field from a disk format bmap extent record.
184 */
185 xfs_fileoff_t
186 xfs_bmbt_disk_get_startoff(
187 xfs_bmbt_rec_t *r)
188 {
189 return ((xfs_fileoff_t)be64_to_cpu(r->l0) &
190 xfs_mask64lo(64 - BMBT_EXNTFLAG_BITLEN)) >> 9;
191 }
192
193
194 /*
195 * Set all the fields in a bmap extent record from the arguments.
196 */
197 void
198 xfs_bmbt_set_allf(
199 xfs_bmbt_rec_host_t *r,
200 xfs_fileoff_t startoff,
201 xfs_fsblock_t startblock,
202 xfs_filblks_t blockcount,
203 xfs_exntst_t state)
204 {
205 int extent_flag = (state == XFS_EXT_NORM) ? 0 : 1;
206
207 ASSERT(state == XFS_EXT_NORM || state == XFS_EXT_UNWRITTEN);
208 ASSERT((startoff & xfs_mask64hi(64-BMBT_STARTOFF_BITLEN)) == 0);
209 ASSERT((blockcount & xfs_mask64hi(64-BMBT_BLOCKCOUNT_BITLEN)) == 0);
210
211 ASSERT((startblock & xfs_mask64hi(64-BMBT_STARTBLOCK_BITLEN)) == 0);
212
213 r->l0 = ((xfs_bmbt_rec_base_t)extent_flag << 63) |
214 ((xfs_bmbt_rec_base_t)startoff << 9) |
215 ((xfs_bmbt_rec_base_t)startblock >> 43);
216 r->l1 = ((xfs_bmbt_rec_base_t)startblock << 21) |
217 ((xfs_bmbt_rec_base_t)blockcount &
218 (xfs_bmbt_rec_base_t)xfs_mask64lo(21));
219 }
220
221 /*
222 * Set all the fields in a bmap extent record from the uncompressed form.
223 */
224 void
225 xfs_bmbt_set_all(
226 xfs_bmbt_rec_host_t *r,
227 xfs_bmbt_irec_t *s)
228 {
229 xfs_bmbt_set_allf(r, s->br_startoff, s->br_startblock,
230 s->br_blockcount, s->br_state);
231 }
232
233
234 /*
235 * Set all the fields in a disk format bmap extent record from the arguments.
236 */
237 void
238 xfs_bmbt_disk_set_allf(
239 xfs_bmbt_rec_t *r,
240 xfs_fileoff_t startoff,
241 xfs_fsblock_t startblock,
242 xfs_filblks_t blockcount,
243 xfs_exntst_t state)
244 {
245 int extent_flag = (state == XFS_EXT_NORM) ? 0 : 1;
246
247 ASSERT(state == XFS_EXT_NORM || state == XFS_EXT_UNWRITTEN);
248 ASSERT((startoff & xfs_mask64hi(64-BMBT_STARTOFF_BITLEN)) == 0);
249 ASSERT((blockcount & xfs_mask64hi(64-BMBT_BLOCKCOUNT_BITLEN)) == 0);
250 ASSERT((startblock & xfs_mask64hi(64-BMBT_STARTBLOCK_BITLEN)) == 0);
251
252 r->l0 = cpu_to_be64(
253 ((xfs_bmbt_rec_base_t)extent_flag << 63) |
254 ((xfs_bmbt_rec_base_t)startoff << 9) |
255 ((xfs_bmbt_rec_base_t)startblock >> 43));
256 r->l1 = cpu_to_be64(
257 ((xfs_bmbt_rec_base_t)startblock << 21) |
258 ((xfs_bmbt_rec_base_t)blockcount &
259 (xfs_bmbt_rec_base_t)xfs_mask64lo(21)));
260 }
261
262 /*
263 * Set all the fields in a bmap extent record from the uncompressed form.
264 */
265 STATIC void
266 xfs_bmbt_disk_set_all(
267 xfs_bmbt_rec_t *r,
268 xfs_bmbt_irec_t *s)
269 {
270 xfs_bmbt_disk_set_allf(r, s->br_startoff, s->br_startblock,
271 s->br_blockcount, s->br_state);
272 }
273
274 /*
275 * Set the blockcount field in a bmap extent record.
276 */
277 void
278 xfs_bmbt_set_blockcount(
279 xfs_bmbt_rec_host_t *r,
280 xfs_filblks_t v)
281 {
282 ASSERT((v & xfs_mask64hi(43)) == 0);
283 r->l1 = (r->l1 & (xfs_bmbt_rec_base_t)xfs_mask64hi(43)) |
284 (xfs_bmbt_rec_base_t)(v & xfs_mask64lo(21));
285 }
286
287 /*
288 * Set the startblock field in a bmap extent record.
289 */
290 void
291 xfs_bmbt_set_startblock(
292 xfs_bmbt_rec_host_t *r,
293 xfs_fsblock_t v)
294 {
295 ASSERT((v & xfs_mask64hi(12)) == 0);
296 r->l0 = (r->l0 & (xfs_bmbt_rec_base_t)xfs_mask64hi(55)) |
297 (xfs_bmbt_rec_base_t)(v >> 43);
298 r->l1 = (r->l1 & (xfs_bmbt_rec_base_t)xfs_mask64lo(21)) |
299 (xfs_bmbt_rec_base_t)(v << 21);
300 }
301
302 /*
303 * Set the startoff field in a bmap extent record.
304 */
305 void
306 xfs_bmbt_set_startoff(
307 xfs_bmbt_rec_host_t *r,
308 xfs_fileoff_t v)
309 {
310 ASSERT((v & xfs_mask64hi(9)) == 0);
311 r->l0 = (r->l0 & (xfs_bmbt_rec_base_t) xfs_mask64hi(1)) |
312 ((xfs_bmbt_rec_base_t)v << 9) |
313 (r->l0 & (xfs_bmbt_rec_base_t)xfs_mask64lo(9));
314 }
315
316 /*
317 * Set the extent state field in a bmap extent record.
318 */
319 void
320 xfs_bmbt_set_state(
321 xfs_bmbt_rec_host_t *r,
322 xfs_exntst_t v)
323 {
324 ASSERT(v == XFS_EXT_NORM || v == XFS_EXT_UNWRITTEN);
325 if (v == XFS_EXT_NORM)
326 r->l0 &= xfs_mask64lo(64 - BMBT_EXNTFLAG_BITLEN);
327 else
328 r->l0 |= xfs_mask64hi(BMBT_EXNTFLAG_BITLEN);
329 }
330
331 /*
332 * Convert in-memory form of btree root to on-disk form.
333 */
334 void
335 xfs_bmbt_to_bmdr(
336 struct xfs_mount *mp,
337 struct xfs_btree_block *rblock,
338 int rblocklen,
339 xfs_bmdr_block_t *dblock,
340 int dblocklen)
341 {
342 int dmxr;
343 xfs_bmbt_key_t *fkp;
344 __be64 *fpp;
345 xfs_bmbt_key_t *tkp;
346 __be64 *tpp;
347
348 if (xfs_sb_version_hascrc(&mp->m_sb)) {
349 ASSERT(rblock->bb_magic == cpu_to_be32(XFS_BMAP_CRC_MAGIC));
350 ASSERT(uuid_equal(&rblock->bb_u.l.bb_uuid,
351 &mp->m_sb.sb_meta_uuid));
352 ASSERT(rblock->bb_u.l.bb_blkno ==
353 cpu_to_be64(XFS_BUF_DADDR_NULL));
354 } else
355 ASSERT(rblock->bb_magic == cpu_to_be32(XFS_BMAP_MAGIC));
356 ASSERT(rblock->bb_u.l.bb_leftsib == cpu_to_be64(NULLFSBLOCK));
357 ASSERT(rblock->bb_u.l.bb_rightsib == cpu_to_be64(NULLFSBLOCK));
358 ASSERT(rblock->bb_level != 0);
359 dblock->bb_level = rblock->bb_level;
360 dblock->bb_numrecs = rblock->bb_numrecs;
361 dmxr = xfs_bmdr_maxrecs(dblocklen, 0);
362 fkp = XFS_BMBT_KEY_ADDR(mp, rblock, 1);
363 tkp = XFS_BMDR_KEY_ADDR(dblock, 1);
364 fpp = XFS_BMAP_BROOT_PTR_ADDR(mp, rblock, 1, rblocklen);
365 tpp = XFS_BMDR_PTR_ADDR(dblock, 1, dmxr);
366 dmxr = be16_to_cpu(dblock->bb_numrecs);
367 memcpy(tkp, fkp, sizeof(*fkp) * dmxr);
368 memcpy(tpp, fpp, sizeof(*fpp) * dmxr);
369 }
370
371 /*
372 * Check extent records, which have just been read, for
373 * any bit in the extent flag field. ASSERT on debug
374 * kernels, as this condition should not occur.
375 * Return an error condition (1) if any flags found,
376 * otherwise return 0.
377 */
378
379 int
380 xfs_check_nostate_extents(
381 xfs_ifork_t *ifp,
382 xfs_extnum_t idx,
383 xfs_extnum_t num)
384 {
385 for (; num > 0; num--, idx++) {
386 xfs_bmbt_rec_host_t *ep = xfs_iext_get_ext(ifp, idx);
387 if ((ep->l0 >>
388 (64 - BMBT_EXNTFLAG_BITLEN)) != 0) {
389 ASSERT(0);
390 return 1;
391 }
392 }
393 return 0;
394 }
395
396
397 STATIC struct xfs_btree_cur *
398 xfs_bmbt_dup_cursor(
399 struct xfs_btree_cur *cur)
400 {
401 struct xfs_btree_cur *new;
402
403 new = xfs_bmbt_init_cursor(cur->bc_mp, cur->bc_tp,
404 cur->bc_private.b.ip, cur->bc_private.b.whichfork);
405
406 /*
407 * Copy the firstblock, dfops, and flags values,
408 * since init cursor doesn't get them.
409 */
410 new->bc_private.b.firstblock = cur->bc_private.b.firstblock;
411 new->bc_private.b.dfops = cur->bc_private.b.dfops;
412 new->bc_private.b.flags = cur->bc_private.b.flags;
413
414 return new;
415 }
416
417 STATIC void
418 xfs_bmbt_update_cursor(
419 struct xfs_btree_cur *src,
420 struct xfs_btree_cur *dst)
421 {
422 ASSERT((dst->bc_private.b.firstblock != NULLFSBLOCK) ||
423 (dst->bc_private.b.ip->i_d.di_flags & XFS_DIFLAG_REALTIME));
424 ASSERT(dst->bc_private.b.dfops == src->bc_private.b.dfops);
425
426 dst->bc_private.b.allocated += src->bc_private.b.allocated;
427 dst->bc_private.b.firstblock = src->bc_private.b.firstblock;
428
429 src->bc_private.b.allocated = 0;
430 }
431
432 STATIC int
433 xfs_bmbt_alloc_block(
434 struct xfs_btree_cur *cur,
435 union xfs_btree_ptr *start,
436 union xfs_btree_ptr *new,
437 int *stat)
438 {
439 xfs_alloc_arg_t args; /* block allocation args */
440 int error; /* error return value */
441
442 memset(&args, 0, sizeof(args));
443 args.tp = cur->bc_tp;
444 args.mp = cur->bc_mp;
445 args.fsbno = cur->bc_private.b.firstblock;
446 args.firstblock = args.fsbno;
447
448 if (args.fsbno == NULLFSBLOCK) {
449 args.fsbno = be64_to_cpu(start->l);
450 args.type = XFS_ALLOCTYPE_START_BNO;
451 /*
452 * Make sure there is sufficient room left in the AG to
453 * complete a full tree split for an extent insert. If
454 * we are converting the middle part of an extent then
455 * we may need space for two tree splits.
456 *
457 * We are relying on the caller to make the correct block
458 * reservation for this operation to succeed. If the
459 * reservation amount is insufficient then we may fail a
460 * block allocation here and corrupt the filesystem.
461 */
462 args.minleft = args.tp->t_blk_res;
463 } else if (cur->bc_private.b.dfops->dop_low) {
464 args.type = XFS_ALLOCTYPE_START_BNO;
465 } else {
466 args.type = XFS_ALLOCTYPE_NEAR_BNO;
467 }
468
469 args.minlen = args.maxlen = args.prod = 1;
470 args.wasdel = cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL;
471 if (!args.wasdel && args.tp->t_blk_res == 0) {
472 error = -ENOSPC;
473 goto error0;
474 }
475 error = xfs_alloc_vextent(&args);
476 if (error)
477 goto error0;
478
479 if (args.fsbno == NULLFSBLOCK && args.minleft) {
480 /*
481 * Could not find an AG with enough free space to satisfy
482 * a full btree split. Try again without minleft and if
483 * successful activate the lowspace algorithm.
484 */
485 args.fsbno = 0;
486 args.type = XFS_ALLOCTYPE_FIRST_AG;
487 args.minleft = 0;
488 error = xfs_alloc_vextent(&args);
489 if (error)
490 goto error0;
491 cur->bc_private.b.dfops->dop_low = true;
492 }
493 if (args.fsbno == NULLFSBLOCK) {
494 XFS_BTREE_TRACE_CURSOR(cur, XBT_EXIT);
495 *stat = 0;
496 return 0;
497 }
498 ASSERT(args.len == 1);
499 cur->bc_private.b.firstblock = args.fsbno;
500 cur->bc_private.b.allocated++;
501 cur->bc_private.b.ip->i_d.di_nblocks++;
502 xfs_trans_log_inode(args.tp, cur->bc_private.b.ip, XFS_ILOG_CORE);
503 xfs_trans_mod_dquot_byino(args.tp, cur->bc_private.b.ip,
504 XFS_TRANS_DQ_BCOUNT, 1L);
505
506 new->l = cpu_to_be64(args.fsbno);
507
508 XFS_BTREE_TRACE_CURSOR(cur, XBT_EXIT);
509 *stat = 1;
510 return 0;
511
512 error0:
513 XFS_BTREE_TRACE_CURSOR(cur, XBT_ERROR);
514 return error;
515 }
516
517 STATIC int
518 xfs_bmbt_free_block(
519 struct xfs_btree_cur *cur,
520 struct xfs_buf *bp)
521 {
522 struct xfs_mount *mp = cur->bc_mp;
523 struct xfs_inode *ip = cur->bc_private.b.ip;
524 struct xfs_trans *tp = cur->bc_tp;
525 xfs_fsblock_t fsbno = XFS_DADDR_TO_FSB(mp, XFS_BUF_ADDR(bp));
526
527 xfs_bmap_add_free(mp, cur->bc_private.b.dfops, fsbno, 1);
528 ip->i_d.di_nblocks--;
529
530 xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
531 xfs_trans_mod_dquot_byino(tp, ip, XFS_TRANS_DQ_BCOUNT, -1L);
532 return 0;
533 }
534
535 STATIC int
536 xfs_bmbt_get_minrecs(
537 struct xfs_btree_cur *cur,
538 int level)
539 {
540 if (level == cur->bc_nlevels - 1) {
541 struct xfs_ifork *ifp;
542
543 ifp = XFS_IFORK_PTR(cur->bc_private.b.ip,
544 cur->bc_private.b.whichfork);
545
546 return xfs_bmbt_maxrecs(cur->bc_mp,
547 ifp->if_broot_bytes, level == 0) / 2;
548 }
549
550 return cur->bc_mp->m_bmap_dmnr[level != 0];
551 }
552
553 int
554 xfs_bmbt_get_maxrecs(
555 struct xfs_btree_cur *cur,
556 int level)
557 {
558 if (level == cur->bc_nlevels - 1) {
559 struct xfs_ifork *ifp;
560
561 ifp = XFS_IFORK_PTR(cur->bc_private.b.ip,
562 cur->bc_private.b.whichfork);
563
564 return xfs_bmbt_maxrecs(cur->bc_mp,
565 ifp->if_broot_bytes, level == 0);
566 }
567
568 return cur->bc_mp->m_bmap_dmxr[level != 0];
569
570 }
571
572 /*
573 * Get the maximum records we could store in the on-disk format.
574 *
575 * For non-root nodes this is equivalent to xfs_bmbt_get_maxrecs, but
576 * for the root node this checks the available space in the dinode fork
577 * so that we can resize the in-memory buffer to match it. After a
578 * resize to the maximum size this function returns the same value
579 * as xfs_bmbt_get_maxrecs for the root node, too.
580 */
581 STATIC int
582 xfs_bmbt_get_dmaxrecs(
583 struct xfs_btree_cur *cur,
584 int level)
585 {
586 if (level != cur->bc_nlevels - 1)
587 return cur->bc_mp->m_bmap_dmxr[level != 0];
588 return xfs_bmdr_maxrecs(cur->bc_private.b.forksize, level == 0);
589 }
590
591 STATIC void
592 xfs_bmbt_init_key_from_rec(
593 union xfs_btree_key *key,
594 union xfs_btree_rec *rec)
595 {
596 key->bmbt.br_startoff =
597 cpu_to_be64(xfs_bmbt_disk_get_startoff(&rec->bmbt));
598 }
599
600 STATIC void
601 xfs_bmbt_init_rec_from_cur(
602 struct xfs_btree_cur *cur,
603 union xfs_btree_rec *rec)
604 {
605 xfs_bmbt_disk_set_all(&rec->bmbt, &cur->bc_rec.b);
606 }
607
608 STATIC void
609 xfs_bmbt_init_ptr_from_cur(
610 struct xfs_btree_cur *cur,
611 union xfs_btree_ptr *ptr)
612 {
613 ptr->l = 0;
614 }
615
616 STATIC __int64_t
617 xfs_bmbt_key_diff(
618 struct xfs_btree_cur *cur,
619 union xfs_btree_key *key)
620 {
621 return (__int64_t)be64_to_cpu(key->bmbt.br_startoff) -
622 cur->bc_rec.b.br_startoff;
623 }
624
625 static bool
626 xfs_bmbt_verify(
627 struct xfs_buf *bp)
628 {
629 struct xfs_mount *mp = bp->b_target->bt_mount;
630 struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
631 unsigned int level;
632
633 switch (block->bb_magic) {
634 case cpu_to_be32(XFS_BMAP_CRC_MAGIC):
635 if (!xfs_sb_version_hascrc(&mp->m_sb))
636 return false;
637 if (!uuid_equal(&block->bb_u.l.bb_uuid, &mp->m_sb.sb_meta_uuid))
638 return false;
639 if (be64_to_cpu(block->bb_u.l.bb_blkno) != bp->b_bn)
640 return false;
641 /*
642 * XXX: need a better way of verifying the owner here. Right now
643 * just make sure there has been one set.
644 */
645 if (be64_to_cpu(block->bb_u.l.bb_owner) == 0)
646 return false;
647 /* fall through */
648 case cpu_to_be32(XFS_BMAP_MAGIC):
649 break;
650 default:
651 return false;
652 }
653
654 /*
655 * numrecs and level verification.
656 *
657 * We don't know what fork we belong to, so just verify that the level
658 * is less than the maximum of the two. Later checks will be more
659 * precise.
660 */
661 level = be16_to_cpu(block->bb_level);
662 if (level > max(mp->m_bm_maxlevels[0], mp->m_bm_maxlevels[1]))
663 return false;
664 if (be16_to_cpu(block->bb_numrecs) > mp->m_bmap_dmxr[level != 0])
665 return false;
666
667 /* sibling pointer verification */
668 if (!block->bb_u.l.bb_leftsib ||
669 (block->bb_u.l.bb_leftsib != cpu_to_be64(NULLFSBLOCK) &&
670 !XFS_FSB_SANITY_CHECK(mp, be64_to_cpu(block->bb_u.l.bb_leftsib))))
671 return false;
672 if (!block->bb_u.l.bb_rightsib ||
673 (block->bb_u.l.bb_rightsib != cpu_to_be64(NULLFSBLOCK) &&
674 !XFS_FSB_SANITY_CHECK(mp, be64_to_cpu(block->bb_u.l.bb_rightsib))))
675 return false;
676
677 return true;
678 }
679
680 static void
681 xfs_bmbt_read_verify(
682 struct xfs_buf *bp)
683 {
684 if (!xfs_btree_lblock_verify_crc(bp))
685 xfs_buf_ioerror(bp, -EFSBADCRC);
686 else if (!xfs_bmbt_verify(bp))
687 xfs_buf_ioerror(bp, -EFSCORRUPTED);
688
689 if (bp->b_error) {
690 trace_xfs_btree_corrupt(bp, _RET_IP_);
691 xfs_verifier_error(bp);
692 }
693 }
694
695 static void
696 xfs_bmbt_write_verify(
697 struct xfs_buf *bp)
698 {
699 if (!xfs_bmbt_verify(bp)) {
700 trace_xfs_btree_corrupt(bp, _RET_IP_);
701 xfs_buf_ioerror(bp, -EFSCORRUPTED);
702 xfs_verifier_error(bp);
703 return;
704 }
705 xfs_btree_lblock_calc_crc(bp);
706 }
707
708 const struct xfs_buf_ops xfs_bmbt_buf_ops = {
709 .name = "xfs_bmbt",
710 .verify_read = xfs_bmbt_read_verify,
711 .verify_write = xfs_bmbt_write_verify,
712 };
713
714
715 #if defined(DEBUG) || defined(XFS_WARN)
716 STATIC int
717 xfs_bmbt_keys_inorder(
718 struct xfs_btree_cur *cur,
719 union xfs_btree_key *k1,
720 union xfs_btree_key *k2)
721 {
722 return be64_to_cpu(k1->bmbt.br_startoff) <
723 be64_to_cpu(k2->bmbt.br_startoff);
724 }
725
726 STATIC int
727 xfs_bmbt_recs_inorder(
728 struct xfs_btree_cur *cur,
729 union xfs_btree_rec *r1,
730 union xfs_btree_rec *r2)
731 {
732 return xfs_bmbt_disk_get_startoff(&r1->bmbt) +
733 xfs_bmbt_disk_get_blockcount(&r1->bmbt) <=
734 xfs_bmbt_disk_get_startoff(&r2->bmbt);
735 }
736 #endif /* DEBUG */
737
738 static const struct xfs_btree_ops xfs_bmbt_ops = {
739 .rec_len = sizeof(xfs_bmbt_rec_t),
740 .key_len = sizeof(xfs_bmbt_key_t),
741
742 .dup_cursor = xfs_bmbt_dup_cursor,
743 .update_cursor = xfs_bmbt_update_cursor,
744 .alloc_block = xfs_bmbt_alloc_block,
745 .free_block = xfs_bmbt_free_block,
746 .get_maxrecs = xfs_bmbt_get_maxrecs,
747 .get_minrecs = xfs_bmbt_get_minrecs,
748 .get_dmaxrecs = xfs_bmbt_get_dmaxrecs,
749 .init_key_from_rec = xfs_bmbt_init_key_from_rec,
750 .init_rec_from_cur = xfs_bmbt_init_rec_from_cur,
751 .init_ptr_from_cur = xfs_bmbt_init_ptr_from_cur,
752 .key_diff = xfs_bmbt_key_diff,
753 .buf_ops = &xfs_bmbt_buf_ops,
754 #if defined(DEBUG) || defined(XFS_WARN)
755 .keys_inorder = xfs_bmbt_keys_inorder,
756 .recs_inorder = xfs_bmbt_recs_inorder,
757 #endif
758
759 .get_leaf_keys = xfs_btree_get_leaf_keys,
760 .get_node_keys = xfs_btree_get_node_keys,
761 .update_keys = xfs_btree_update_keys,
762 };
763
764 /*
765 * Allocate a new bmap btree cursor.
766 */
767 struct xfs_btree_cur * /* new bmap btree cursor */
768 xfs_bmbt_init_cursor(
769 struct xfs_mount *mp, /* file system mount point */
770 struct xfs_trans *tp, /* transaction pointer */
771 struct xfs_inode *ip, /* inode owning the btree */
772 int whichfork) /* data or attr fork */
773 {
774 struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, whichfork);
775 struct xfs_btree_cur *cur;
776
777 cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_SLEEP);
778
779 cur->bc_tp = tp;
780 cur->bc_mp = mp;
781 cur->bc_nlevels = be16_to_cpu(ifp->if_broot->bb_level) + 1;
782 cur->bc_btnum = XFS_BTNUM_BMAP;
783 cur->bc_blocklog = mp->m_sb.sb_blocklog;
784
785 cur->bc_ops = &xfs_bmbt_ops;
786 cur->bc_flags = XFS_BTREE_LONG_PTRS | XFS_BTREE_ROOT_IN_INODE;
787 if (xfs_sb_version_hascrc(&mp->m_sb))
788 cur->bc_flags |= XFS_BTREE_CRC_BLOCKS;
789
790 cur->bc_private.b.forksize = XFS_IFORK_SIZE(ip, whichfork);
791 cur->bc_private.b.ip = ip;
792 cur->bc_private.b.firstblock = NULLFSBLOCK;
793 cur->bc_private.b.dfops = NULL;
794 cur->bc_private.b.allocated = 0;
795 cur->bc_private.b.flags = 0;
796 cur->bc_private.b.whichfork = whichfork;
797
798 return cur;
799 }
800
801 /*
802 * Calculate number of records in a bmap btree block.
803 */
804 int
805 xfs_bmbt_maxrecs(
806 struct xfs_mount *mp,
807 int blocklen,
808 int leaf)
809 {
810 blocklen -= XFS_BMBT_BLOCK_LEN(mp);
811
812 if (leaf)
813 return blocklen / sizeof(xfs_bmbt_rec_t);
814 return blocklen / (sizeof(xfs_bmbt_key_t) + sizeof(xfs_bmbt_ptr_t));
815 }
816
817 /*
818 * Calculate number of records in a bmap btree inode root.
819 */
820 int
821 xfs_bmdr_maxrecs(
822 int blocklen,
823 int leaf)
824 {
825 blocklen -= sizeof(xfs_bmdr_block_t);
826
827 if (leaf)
828 return blocklen / sizeof(xfs_bmdr_rec_t);
829 return blocklen / (sizeof(xfs_bmdr_key_t) + sizeof(xfs_bmdr_ptr_t));
830 }
831
832 /*
833 * Change the owner of a btree format fork fo the inode passed in. Change it to
834 * the owner of that is passed in so that we can change owners before or after
835 * we switch forks between inodes. The operation that the caller is doing will
836 * determine whether is needs to change owner before or after the switch.
837 *
838 * For demand paged transactional modification, the fork switch should be done
839 * after reading in all the blocks, modifying them and pinning them in the
840 * transaction. For modification when the buffers are already pinned in memory,
841 * the fork switch can be done before changing the owner as we won't need to
842 * validate the owner until the btree buffers are unpinned and writes can occur
843 * again.
844 *
845 * For recovery based ownership change, there is no transactional context and
846 * so a buffer list must be supplied so that we can record the buffers that we
847 * modified for the caller to issue IO on.
848 */
849 int
850 xfs_bmbt_change_owner(
851 struct xfs_trans *tp,
852 struct xfs_inode *ip,
853 int whichfork,
854 xfs_ino_t new_owner,
855 struct list_head *buffer_list)
856 {
857 struct xfs_btree_cur *cur;
858 int error;
859
860 ASSERT(tp || buffer_list);
861 ASSERT(!(tp && buffer_list));
862 if (whichfork == XFS_DATA_FORK)
863 ASSERT(ip->i_d.di_format == XFS_DINODE_FMT_BTREE);
864 else
865 ASSERT(ip->i_d.di_aformat == XFS_DINODE_FMT_BTREE);
866
867 cur = xfs_bmbt_init_cursor(ip->i_mount, tp, ip, whichfork);
868 if (!cur)
869 return -ENOMEM;
870
871 error = xfs_btree_change_owner(cur, new_owner, buffer_list);
872 xfs_btree_del_cursor(cur, error ? XFS_BTREE_ERROR : XFS_BTREE_NOERROR);
873 return error;
874 }