1 /* 2 * Copyright (c) 2008 The DragonFly Project. All rights reserved. 3 * 4 * This code is derived from software contributed to The DragonFly Project 5 * by Matthew Dillon <dillon@backplane.com> 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 11 * 1. Redistributions of source code must retain the above copyright 12 * notice, this list of conditions and the following disclaimer. 13 * 2. Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in 15 * the documentation and/or other materials provided with the 16 * distribution. 17 * 3. Neither the name of The DragonFly Project nor the names of its 18 * contributors may be used to endorse or promote products derived 19 * from this software without specific, prior written permission. 20 * 21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 22 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 23 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS 24 * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE 25 * COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, 26 * INCIDENTAL, SPECIAL, EXEMPLARY OR CONSEQUENTIAL DAMAGES (INCLUDING, 27 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; 28 * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED 29 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, 30 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT 31 * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 32 * SUCH DAMAGE. 33 */ 34 35 #include <libutil.h> 36 37 #include "hammer.h" 38 39 #define FLAG_TOOFARLEFT 0x0001 40 #define FLAG_TOOFARRIGHT 0x0002 41 #define FLAG_BADTYPE 0x0004 42 #define FLAG_BADCHILDPARENT 0x0008 43 #define FLAG_BADMIRRORTID 0x0010 44 45 struct { 46 struct hammer_base_elm base; 47 int limit; /* # of fields to test */ 48 int filter; /* filter type (default -1) */ 49 int obfuscate; /* obfuscate direntry name */ 50 int indent; /* use depth indentation */ 51 zone_stat_t stats; 52 } opt; 53 54 static __inline void print_btree(hammer_off_t node_offset); 55 static __inline void print_subtree(hammer_btree_elm_t elm); 56 static void print_btree_node(hammer_off_t node_offset, hammer_tid_t mirror_tid, 57 hammer_btree_elm_t lbe); 58 static int test_node_count(hammer_node_ondisk_t node, char *badmp); 59 static void print_btree_elm(hammer_node_ondisk_t node, hammer_off_t node_offset, 60 hammer_btree_elm_t elm, hammer_btree_elm_t lbe, const char *ext); 61 static int get_elm_flags(hammer_node_ondisk_t node, hammer_off_t node_offset, 62 hammer_btree_elm_t elm, hammer_btree_elm_t lbe); 63 static int test_lr(hammer_btree_elm_t elm, hammer_btree_elm_t lbe); 64 static int test_rbn_lr(hammer_btree_elm_t elm, hammer_btree_elm_t lbe); 65 static void print_bigblock_fill(hammer_off_t offset); 66 static const char *check_data_crc(hammer_btree_elm_t elm, const char **whichp); 67 static hammer_crc_t get_inode_crc(hammer_btree_leaf_elm_t leaf, 68 const char **whichp); 69 static hammer_crc_t get_buf_crc(hammer_btree_leaf_elm_t leaf, 70 const char **whichp); 71 static void print_record(hammer_btree_elm_t elm); 72 static int init_btree_search(const char *arg); 73 static int test_btree_search(hammer_btree_elm_t elm); 74 static int test_btree_match(hammer_btree_elm_t elm); 75 static int test_btree_out_of_range(hammer_btree_elm_t elm); 76 static void hexdump_record(const void *ptr, int length, const char *hdr); 77 78 static int num_bad_node = 0; 79 static int num_bad_elm = 0; 80 static int num_bad_rec = 0; 81 static int depth; 82 83 #define _X "\t" 84 static const char* _indents[] = { 85 "", 86 _X, 87 _X _X, 88 _X _X _X, 89 _X _X _X _X, 90 _X _X _X _X _X, 91 _X _X _X _X _X _X, 92 _X _X _X _X _X _X _X, 93 _X _X _X _X _X _X _X _X, 94 _X _X _X _X _X _X _X _X _X, 95 _X _X _X _X _X _X _X _X _X _X, 96 _X _X _X _X _X _X _X _X _X _X _X, 97 _X _X _X _X _X _X _X _X _X _X _X _X, 98 _X _X _X _X _X _X _X _X _X _X _X _X _X, 99 _X _X _X _X _X _X _X _X _X _X _X _X _X _X, 100 _X _X _X _X _X _X _X _X _X _X _X _X _X _X _X, 101 _X _X _X _X _X _X _X _X _X _X _X _X _X _X _X _X, 102 /* deep enough */ 103 }; 104 #define INDENT _indents[opt.indent ? depth : 0] 105 106 void 107 hammer_cmd_show(const char *arg, int filter, int obfuscate, int indent) 108 { 109 volume_info_t volume; 110 hammer_volume_ondisk_t ondisk; 111 zone_stat_t stats = NULL; 112 113 volume = get_root_volume(); 114 ondisk = volume->ondisk; 115 116 print_blockmap(volume); 117 118 if (VerboseOpt) 119 stats = hammer_init_zone_stat_bits(); 120 121 bzero(&opt, sizeof(opt)); 122 opt.filter = filter; 123 opt.obfuscate = obfuscate; 124 opt.indent = indent; 125 opt.stats = stats; 126 127 if (init_btree_search(arg) > 0) { 128 printf("arg=\"%s\"", arg); 129 if (opt.limit > 0) 130 printf(" lo=%08x", opt.base.localization); 131 if (opt.limit > 1) 132 printf(" objid=%016jx", (uintmax_t)opt.base.obj_id); 133 if (opt.limit > 2) 134 printf(" rt=%02x", opt.base.rec_type); 135 if (opt.limit > 3) 136 printf(" key=%016jx", (uintmax_t)opt.base.key); 137 if (opt.limit > 4) 138 printf(" tid=%016jx", (uintmax_t)opt.base.create_tid); 139 printf("\n"); 140 } 141 print_btree(ondisk->vol0_btree_root); 142 143 if (stats) { 144 hammer_print_zone_stat(stats); 145 hammer_cleanup_zone_stat(stats); 146 } 147 148 if (num_bad_node || VerboseOpt) 149 printf("%d bad nodes\n", num_bad_node); 150 if (num_bad_elm || VerboseOpt) 151 printf("%d bad elms\n", num_bad_elm); 152 if (num_bad_rec || VerboseOpt) 153 printf("%d bad records\n", num_bad_rec); 154 } 155 156 static __inline 157 void 158 print_btree(hammer_off_t node_offset) 159 { 160 depth = -1; 161 print_btree_node(node_offset, HAMMER_MAX_TID, NULL); 162 assert(depth == -1); 163 } 164 165 static __inline 166 void 167 print_subtree(hammer_btree_elm_t elm) 168 { 169 hammer_btree_internal_elm_t i = &elm->internal; 170 print_btree_node(i->subtree_offset, i->mirror_tid, elm); 171 } 172 173 static 174 void 175 print_btree_node(hammer_off_t node_offset, hammer_tid_t mirror_tid, 176 hammer_btree_elm_t lbe) 177 { 178 buffer_info_t buffer = NULL; 179 hammer_node_ondisk_t node; 180 hammer_btree_elm_t elm; 181 int i; 182 char badc = ' '; /* good */ 183 char badm = ' '; /* good */ 184 const char *ext; 185 186 depth++; 187 node = get_buffer_data(node_offset, &buffer, 0); 188 189 if (node == NULL) { 190 badc = 'B'; 191 badm = 'I'; 192 } else { 193 if (!hammer_crc_test_btree(HammerVersion, node)) 194 badc = 'B'; 195 if (node->mirror_tid > mirror_tid) { 196 badc = 'B'; 197 badm = 'M'; 198 } 199 if (test_node_count(node, &badm) == -1) { 200 badc = 'B'; 201 assert(badm != ' '); 202 } 203 } 204 205 if (badm != ' ' || badc != ' ') /* not good */ 206 ++num_bad_node; 207 208 printf("%s%c%c NODE %016jx ", 209 INDENT, badc, badm, (uintmax_t)node_offset); 210 printf("cnt=%02d p=%016jx type=%c depth=%d mirror=%016jx", 211 node->count, 212 (uintmax_t)node->parent, 213 (node->type ? node->type : '?'), 214 depth, 215 (uintmax_t)node->mirror_tid); 216 printf(" fill="); 217 print_bigblock_fill(node_offset); 218 printf(" {\n"); 219 220 if (opt.stats) 221 hammer_add_zone_stat(opt.stats, node_offset, sizeof(*node)); 222 223 for (i = 0; i < node->count; ++i) { 224 elm = &node->elms[i]; 225 ext = NULL; 226 if (opt.limit) { 227 switch (node->type) { 228 case HAMMER_BTREE_TYPE_INTERNAL: 229 if (!test_btree_out_of_range(elm)) 230 ext = "*"; 231 break; 232 case HAMMER_BTREE_TYPE_LEAF: 233 if (test_btree_match(elm)) 234 ext = "*"; 235 break; 236 } 237 } 238 print_btree_elm(node, node_offset, elm, lbe, ext); 239 } 240 if (node->type == HAMMER_BTREE_TYPE_INTERNAL) { 241 assert(i == node->count); /* boundary */ 242 elm = &node->elms[i]; 243 print_btree_elm(node, node_offset, elm, lbe, NULL); 244 } 245 printf("%s }\n", INDENT); 246 247 if (node->type == HAMMER_BTREE_TYPE_INTERNAL) { 248 for (i = 0; i < node->count; ++i) { 249 elm = &node->elms[i]; 250 if (opt.limit && opt.filter) { 251 if (test_btree_out_of_range(elm)) 252 continue; 253 } 254 if (elm->internal.subtree_offset) { 255 print_subtree(elm); 256 /* 257 * Cause show to do normal iteration after 258 * seeking to the lo:objid:rt:key:tid 259 * by default 260 */ 261 if (opt.limit && opt.filter == -1) /* default */ 262 opt.filter = 0; 263 } 264 } 265 } 266 rel_buffer(buffer); 267 depth--; 268 } 269 270 static 271 int 272 test_node_count(hammer_node_ondisk_t node, char *badmp) 273 { 274 hammer_node_ondisk_t parent_node; 275 buffer_info_t buffer = NULL; 276 int maxcount; 277 278 maxcount = hammer_node_max_elements(node->type); 279 280 if (maxcount == -1) { 281 *badmp = 'U'; 282 return(-1); 283 } else if (node->count > maxcount) { 284 *badmp = 'C'; 285 return(-1); 286 } else if (node->count == 0) { 287 parent_node = get_buffer_data(node->parent, &buffer, 0); 288 if (parent_node->count != 1) { 289 *badmp = 'C'; 290 rel_buffer(buffer); 291 return(-1); 292 } 293 rel_buffer(buffer); 294 } 295 296 return(0); 297 } 298 299 static __inline 300 int 301 is_root_btree_beg(uint8_t type, int i, hammer_btree_elm_t elm) 302 { 303 /* 304 * elm->base.btype depends on what the original node had 305 * so it could be anything but HAMMER_BTREE_TYPE_NONE. 306 */ 307 return (type == HAMMER_BTREE_TYPE_INTERNAL && 308 i == 0 && 309 elm->base.localization == HAMMER_MIN_ONDISK_LOCALIZATION && 310 elm->base.obj_id == (int64_t)HAMMER_MIN_OBJID && 311 elm->base.key == (int64_t)HAMMER_MIN_KEY && 312 elm->base.create_tid == 1 && 313 elm->base.delete_tid == 1 && 314 elm->base.rec_type == HAMMER_MIN_RECTYPE && 315 elm->base.obj_type == 0 && 316 elm->base.btype != HAMMER_BTREE_TYPE_NONE); 317 } 318 319 static __inline 320 int 321 is_root_btree_end(uint8_t type, int i, hammer_btree_elm_t elm) 322 { 323 return (type == HAMMER_BTREE_TYPE_INTERNAL && 324 i != 0 && 325 elm->base.localization == HAMMER_MAX_ONDISK_LOCALIZATION && 326 elm->base.obj_id == HAMMER_MAX_OBJID && 327 elm->base.key == HAMMER_MAX_KEY && 328 elm->base.create_tid == HAMMER_MAX_TID && 329 elm->base.delete_tid == 0 && 330 elm->base.rec_type == HAMMER_MAX_RECTYPE && 331 elm->base.obj_type == 0 && 332 elm->base.btype == HAMMER_BTREE_TYPE_NONE); 333 } 334 335 static 336 void 337 print_btree_elm(hammer_node_ondisk_t node, hammer_off_t node_offset, 338 hammer_btree_elm_t elm, hammer_btree_elm_t lbe, const char *ext) 339 { 340 char flagstr[8] = { 0, '-', '-', '-', '-', '-', '-', 0 }; 341 char deleted; 342 char rootelm; 343 const char *label; 344 const char *p; 345 const char *which; 346 int flags; 347 int i = ((char*)elm - (char*)node) / (int)sizeof(*elm) - 1; 348 349 flags = get_elm_flags(node, node_offset, elm, lbe); 350 flagstr[0] = flags ? 'B' : 'G'; 351 if (flags & FLAG_TOOFARLEFT) 352 flagstr[2] = 'L'; 353 if (flags & FLAG_TOOFARRIGHT) 354 flagstr[3] = 'R'; 355 if (flags & FLAG_BADTYPE) 356 flagstr[4] = 'T'; 357 if (flags & FLAG_BADCHILDPARENT) 358 flagstr[5] = 'C'; 359 if (flags & FLAG_BADMIRRORTID) 360 flagstr[6] = 'M'; 361 if (flagstr[0] == 'B') 362 ++num_bad_elm; 363 364 /* 365 * Check if elm is derived from root split 366 */ 367 if (is_root_btree_beg(node->type, i, elm)) 368 rootelm = '>'; 369 else if (is_root_btree_end(node->type, i, elm)) 370 rootelm = '<'; 371 else 372 rootelm = ' '; 373 374 if (elm->base.delete_tid) 375 deleted = 'd'; 376 else 377 deleted = ' '; 378 379 if (node->type == HAMMER_BTREE_TYPE_INTERNAL && node->count == i) 380 label = "RBN"; 381 else 382 label = "ELM"; 383 384 printf("%s%s %s %2d %c ", 385 INDENT, flagstr, label, i, hammer_elm_btype(elm)); 386 printf("lo=%08x objid=%016jx rt=%02x key=%016jx tid=%016jx\n", 387 elm->base.localization, 388 (uintmax_t)elm->base.obj_id, 389 elm->base.rec_type, 390 (uintmax_t)elm->base.key, 391 (uintmax_t)elm->base.create_tid); 392 printf("%s %c del=%016jx ot=%02x", 393 INDENT, 394 (rootelm == ' ' ? deleted : rootelm), 395 (uintmax_t)elm->base.delete_tid, 396 elm->base.obj_type); 397 398 switch(node->type) { 399 case HAMMER_BTREE_TYPE_INTERNAL: 400 printf(" suboff=%016jx mirror=%016jx", 401 (uintmax_t)elm->internal.subtree_offset, 402 (uintmax_t)elm->internal.mirror_tid); 403 if (ext) 404 printf(" %s", ext); 405 break; 406 case HAMMER_BTREE_TYPE_LEAF: 407 switch(elm->base.btype) { 408 case HAMMER_BTREE_TYPE_RECORD: 409 printf(" dataoff=%016jx/%d", 410 (uintmax_t)elm->leaf.data_offset, 411 elm->leaf.data_len); 412 p = check_data_crc(elm, &which); 413 printf(" %scrc=%08x", which, elm->leaf.data_crc); 414 if (p) { 415 printf(" error=%s", p); 416 ++num_bad_rec; 417 } 418 printf(" fill="); 419 print_bigblock_fill(elm->leaf.data_offset); 420 if (QuietOpt < 2) 421 print_record(elm); 422 if (opt.stats) { 423 hammer_add_zone_stat(opt.stats, 424 elm->leaf.data_offset, 425 elm->leaf.data_len); 426 } 427 break; 428 default: 429 printf(" badtype=%d", elm->base.btype); 430 break; 431 } 432 if (ext) 433 printf(" %s", ext); 434 break; 435 } 436 printf("\n"); 437 } 438 439 static 440 int 441 get_elm_flags(hammer_node_ondisk_t node, hammer_off_t node_offset, 442 hammer_btree_elm_t elm, hammer_btree_elm_t lbe) 443 { 444 hammer_off_t child_offset; 445 int flags = 0; 446 int i = ((char*)elm - (char*)node) / (int)sizeof(*elm) - 1; 447 448 switch(node->type) { 449 case HAMMER_BTREE_TYPE_INTERNAL: 450 child_offset = elm->internal.subtree_offset; 451 if (elm->internal.mirror_tid > node->mirror_tid) 452 flags |= FLAG_BADMIRRORTID; 453 454 if (i == node->count) { 455 if (child_offset != 0) 456 flags |= FLAG_BADCHILDPARENT; 457 switch(elm->base.btype) { 458 case HAMMER_BTREE_TYPE_NONE: 459 flags |= test_rbn_lr(elm, lbe); 460 break; 461 default: 462 flags |= FLAG_BADTYPE; 463 break; 464 } 465 } else { 466 if (child_offset == 0) { 467 flags |= FLAG_BADCHILDPARENT; 468 } else { 469 buffer_info_t buffer = NULL; 470 hammer_node_ondisk_t subnode; 471 subnode = get_buffer_data(child_offset, &buffer, 0); 472 if (subnode == NULL) 473 flags |= FLAG_BADCHILDPARENT; 474 else if (subnode->parent != node_offset) 475 flags |= FLAG_BADCHILDPARENT; 476 rel_buffer(buffer); 477 } 478 switch(elm->base.btype) { 479 case HAMMER_BTREE_TYPE_INTERNAL: 480 case HAMMER_BTREE_TYPE_LEAF: 481 flags |= test_lr(elm, lbe); 482 break; 483 default: 484 flags |= FLAG_BADTYPE; 485 break; 486 } 487 } 488 break; 489 case HAMMER_BTREE_TYPE_LEAF: 490 if (elm->leaf.data_offset == 0) 491 flags |= FLAG_BADCHILDPARENT; 492 if (elm->leaf.data_len == 0) 493 flags |= FLAG_BADCHILDPARENT; 494 495 if (node->mirror_tid == 0 && 496 !(node->parent == 0 && node->count == 2)) { 497 flags |= FLAG_BADMIRRORTID; 498 } 499 if (elm->base.create_tid && node->mirror_tid && 500 elm->base.create_tid > node->mirror_tid) { 501 flags |= FLAG_BADMIRRORTID; 502 } 503 if (elm->base.delete_tid && node->mirror_tid && 504 elm->base.delete_tid > node->mirror_tid) { 505 flags |= FLAG_BADMIRRORTID; 506 } 507 switch(elm->base.btype) { 508 case HAMMER_BTREE_TYPE_RECORD: 509 flags |= test_lr(elm, lbe); 510 break; 511 default: 512 flags |= FLAG_BADTYPE; 513 break; 514 } 515 break; 516 default: 517 flags |= FLAG_BADTYPE; 518 break; 519 } 520 return(flags); 521 } 522 523 /* 524 * Taken from /usr/src/sys/vfs/hammer/hammer_btree.c. 525 */ 526 static 527 int 528 hammer_btree_cmp(hammer_base_elm_t key1, hammer_base_elm_t key2) 529 { 530 if (key1->localization < key2->localization) 531 return(-5); 532 if (key1->localization > key2->localization) 533 return(5); 534 535 if (key1->obj_id < key2->obj_id) 536 return(-4); 537 if (key1->obj_id > key2->obj_id) 538 return(4); 539 540 if (key1->rec_type < key2->rec_type) 541 return(-3); 542 if (key1->rec_type > key2->rec_type) 543 return(3); 544 545 if (key1->key < key2->key) 546 return(-2); 547 if (key1->key > key2->key) 548 return(2); 549 550 if (key1->create_tid == 0) { 551 if (key2->create_tid == 0) 552 return(0); 553 return(1); 554 } 555 if (key2->create_tid == 0) 556 return(-1); 557 if (key1->create_tid < key2->create_tid) 558 return(-1); 559 if (key1->create_tid > key2->create_tid) 560 return(1); 561 return(0); 562 } 563 564 static 565 int 566 test_lr(hammer_btree_elm_t elm, hammer_btree_elm_t lbe) 567 { 568 if (lbe) { 569 hammer_btree_elm_t rbe = lbe + 1; 570 if (hammer_btree_cmp(&elm->base, &lbe->base) < 0) 571 return(FLAG_TOOFARLEFT); 572 if (hammer_btree_cmp(&elm->base, &rbe->base) >= 0) 573 return(FLAG_TOOFARRIGHT); 574 } 575 return(0); 576 } 577 578 static 579 int 580 test_rbn_lr(hammer_btree_elm_t rbn, hammer_btree_elm_t lbe) 581 { 582 if (lbe) { 583 hammer_btree_elm_t rbe = lbe + 1; 584 if (hammer_btree_cmp(&rbn->base, &lbe->base) < 0) 585 return(FLAG_TOOFARLEFT); 586 if (hammer_btree_cmp(&rbn->base, &rbe->base) > 0) 587 return(FLAG_TOOFARRIGHT); 588 } 589 return(0); 590 } 591 592 static 593 void 594 print_bigblock_fill(hammer_off_t offset) 595 { 596 struct hammer_blockmap_layer1 layer1; 597 struct hammer_blockmap_layer2 layer2; 598 int fill; 599 int error; 600 601 blockmap_lookup_save(offset, &layer1, &layer2, &error); 602 printf("z%d:v%d:%d:%d:%lu=", 603 HAMMER_ZONE_DECODE(offset), 604 HAMMER_VOL_DECODE(offset), 605 HAMMER_BLOCKMAP_LAYER1_INDEX(offset), 606 HAMMER_BLOCKMAP_LAYER2_INDEX(offset), 607 offset & HAMMER_BIGBLOCK_MASK64); 608 609 if (error) { 610 printf("B%d", error); 611 } else { 612 fill = layer2.bytes_free * 100 / HAMMER_BIGBLOCK_SIZE; 613 printf("%d%%", 100 - fill); 614 } 615 } 616 617 /* 618 * Check the generic crc on a data element. Inodes record types are 619 * special in that some of their fields are not CRCed. 620 * 621 * Also check that the zone is valid. 622 */ 623 static 624 const char * 625 check_data_crc(hammer_btree_elm_t elm, const char **whichp) 626 { 627 hammer_crc_t crc; 628 629 *whichp = ""; 630 if (elm->leaf.data_offset == 0 || elm->leaf.data_len == 0) 631 return("ZO"); /* zero offset or length */ 632 633 switch (elm->leaf.base.rec_type) { 634 case HAMMER_RECTYPE_INODE: 635 if (elm->leaf.data_len != sizeof(struct hammer_inode_data)) 636 return("BI"); /* bad inode size */ 637 crc = get_inode_crc(&elm->leaf, whichp); 638 break; 639 default: 640 crc = get_buf_crc(&elm->leaf, whichp); 641 break; 642 } 643 644 if (crc == 0) 645 return("Bx"); /* bad crc */ 646 if (crc != elm->leaf.data_crc) 647 return("BX"); /* bad crc */ 648 return(NULL); /* success */ 649 } 650 651 static 652 hammer_crc_t 653 get_inode_crc(hammer_btree_leaf_elm_t leaf, const char **whichp) 654 { 655 buffer_info_t data_buffer = NULL; 656 hammer_crc_t crc; 657 char *ptr; 658 659 ptr = get_buffer_data(leaf->data_offset, &data_buffer, 0); 660 661 if (HammerVersion >= HAMMER_VOL_VERSION_SEVEN) { 662 crc = iscsi_crc32(ptr, HAMMER_INODE_CRCSIZE); 663 if (crc == leaf->data_crc) { 664 *whichp = "i"; 665 goto end; 666 } 667 } 668 669 crc = crc32(ptr, HAMMER_INODE_CRCSIZE); 670 if (crc == leaf->data_crc) { 671 *whichp = "o"; 672 goto end; 673 } 674 675 *whichp = ""; 676 end: 677 rel_buffer(data_buffer); 678 return(crc); 679 } 680 681 typedef uint32_t (*crc32_ext_fn)(const void *, size_t, uint32_t); 682 683 static 684 hammer_crc_t 685 __get_buf_crc(hammer_btree_leaf_elm_t leaf, crc32_ext_fn f) 686 { 687 buffer_info_t data_buffer = NULL; 688 hammer_off_t buf_offset; 689 hammer_crc_t crc = 0; 690 int32_t buf_len, len; 691 char *ptr; 692 693 buf_offset = leaf->data_offset; 694 buf_len = leaf->data_len; 695 696 while (buf_len) { 697 ptr = get_buffer_data(buf_offset, &data_buffer, 0); 698 len = HAMMER_BUFSIZE - ((int)buf_offset & HAMMER_BUFMASK); 699 if (len > buf_len) 700 len = (int)buf_len; 701 assert(len <= HAMMER_BUFSIZE); 702 crc = f(ptr, len, crc); 703 buf_len -= len; 704 buf_offset += len; 705 } 706 rel_buffer(data_buffer); 707 708 return(crc); 709 } 710 711 static 712 hammer_crc_t 713 get_buf_crc(hammer_btree_leaf_elm_t leaf, const char **whichp) 714 { 715 hammer_crc_t crc; 716 717 if (HammerVersion >= HAMMER_VOL_VERSION_SEVEN) { 718 crc = __get_buf_crc(leaf, iscsi_crc32_ext); 719 if (crc == leaf->data_crc) { 720 *whichp = "i"; 721 goto end; 722 } 723 } 724 725 crc = __get_buf_crc(leaf, crc32_ext); 726 if (crc == leaf->data_crc) { 727 *whichp = "o"; 728 goto end; 729 } 730 731 *whichp = ""; 732 end: 733 return(crc); 734 } 735 736 static 737 void 738 print_config(char *cfgtxt) 739 { 740 char *token; 741 742 printf("\n%s%17s", INDENT, ""); 743 printf("config text=\"\n"); 744 if (cfgtxt != NULL) { 745 while((token = strsep(&cfgtxt, "\r\n")) != NULL) { 746 if (strlen(token)) { 747 printf("%s%17s %s\n", 748 INDENT, "", token); 749 } 750 } 751 } 752 printf("%s%17s \"", INDENT, ""); 753 } 754 755 static 756 void 757 print_record(hammer_btree_elm_t elm) 758 { 759 buffer_info_t data_buffer; 760 hammer_off_t data_offset; 761 int32_t data_len; 762 hammer_data_ondisk_t data; 763 uint32_t status; 764 char *str1 = NULL; 765 char *str2 = NULL; 766 767 data_offset = elm->leaf.data_offset; 768 data_len = elm->leaf.data_len; 769 assert(data_offset != 0); 770 assert(data_len != 0); 771 772 data_buffer = NULL; 773 data = get_buffer_data(data_offset, &data_buffer, 0); 774 assert(data != NULL); 775 776 switch(elm->leaf.base.rec_type) { 777 case HAMMER_RECTYPE_UNKNOWN: 778 printf("\n%s%17s", INDENT, ""); 779 printf("unknown"); 780 break; 781 case HAMMER_RECTYPE_INODE: 782 printf("\n%s%17s", INDENT, ""); 783 printf("inode size=%jd nlinks=%jd", 784 (intmax_t)data->inode.size, 785 (intmax_t)data->inode.nlinks); 786 printf(" mode=%05o uflags=%08x caps=%02x", 787 data->inode.mode, 788 data->inode.uflags, 789 data->inode.cap_flags); 790 printf(" pobjid=%016jx ot=%02x\n", 791 (uintmax_t)data->inode.parent_obj_id, 792 data->inode.obj_type); 793 printf("%s%17s", INDENT, ""); 794 printf(" ctime=%016jx mtime=%016jx atime=%016jx", 795 (uintmax_t)data->inode.ctime, 796 (uintmax_t)data->inode.mtime, 797 (uintmax_t)data->inode.atime); 798 if (data->inode.ext.symlink[0]) { 799 printf(" symlink=\"%s\"", 800 data->inode.ext.symlink); 801 } 802 break; 803 case HAMMER_RECTYPE_DIRENTRY: 804 data_len -= HAMMER_ENTRY_NAME_OFF; 805 printf("\n%s%17s", INDENT, ""); 806 printf("dir-entry objid=%016jx lo=%08x", 807 (uintmax_t)data->entry.obj_id, 808 data->entry.localization); 809 if (!opt.obfuscate) { 810 printf(" name=\"%*.*s\"", 811 data_len, data_len, data->entry.name); 812 } 813 break; 814 case HAMMER_RECTYPE_FIX: 815 switch(elm->leaf.base.key) { 816 case HAMMER_FIXKEY_SYMLINK: 817 data_len -= HAMMER_SYMLINK_NAME_OFF; 818 printf("\n%s%17s", INDENT, ""); 819 printf("fix-symlink name=\"%*.*s\"", 820 data_len, data_len, data->symlink.name); 821 break; 822 } 823 break; 824 case HAMMER_RECTYPE_PFS: 825 printf("\n%s%17s", INDENT, ""); 826 printf("pfs sync_beg_tid=%016jx sync_end_tid=%016jx\n", 827 (intmax_t)data->pfsd.sync_beg_tid, 828 (intmax_t)data->pfsd.sync_end_tid); 829 uuid_to_string(&data->pfsd.shared_uuid, &str1, &status); 830 uuid_to_string(&data->pfsd.unique_uuid, &str2, &status); 831 printf("%17s", ""); 832 printf(" shared_uuid=%s\n", str1); 833 printf("%17s", ""); 834 printf(" unique_uuid=%s\n", str2); 835 printf("%17s", ""); 836 printf(" mirror_flags=%08x label=\"%s\"", 837 data->pfsd.mirror_flags, data->pfsd.label); 838 if (data->pfsd.snapshots[0]) 839 printf(" snapshots=\"%s\"", data->pfsd.snapshots); 840 free(str1); 841 free(str2); 842 break; 843 case HAMMER_RECTYPE_SNAPSHOT: 844 printf("\n%s%17s", INDENT, ""); 845 printf("snapshot tid=%016jx label=\"%s\"", 846 (intmax_t)data->snap.tid, data->snap.label); 847 break; 848 case HAMMER_RECTYPE_CONFIG: 849 if (VerboseOpt > 2) { 850 char *p = strdup(data->config.text); 851 print_config(p); 852 free(p); 853 } 854 break; 855 case HAMMER_RECTYPE_DATA: 856 if (VerboseOpt > 3) { 857 printf("\n"); 858 hexdump_record(data, data_len, "\t\t "); 859 } 860 break; 861 case HAMMER_RECTYPE_EXT: 862 case HAMMER_RECTYPE_DB: 863 if (VerboseOpt > 2) { 864 printf("\n"); 865 hexdump_record(data, data_len, "\t\t "); 866 } 867 break; 868 default: 869 assert(0); 870 break; 871 } 872 rel_buffer(data_buffer); 873 } 874 875 /* 876 * HAMMER userspace only supports buffer size upto HAMMER_BUFSIZE 877 * which is 16KB. Passing record data length larger than 16KB to 878 * hexdump(3) is invalid even if the leaf node elm says >16KB data. 879 */ 880 static 881 void 882 hexdump_record(const void *ptr, int length, const char *hdr) 883 { 884 int data_len = length; 885 886 if (data_len > HAMMER_BUFSIZE) /* XXX */ 887 data_len = HAMMER_BUFSIZE; 888 hexdump(ptr, data_len, hdr, 0); 889 890 if (length > data_len) 891 printf("%s....\n", hdr); 892 } 893 894 static __inline __always_inline 895 unsigned long 896 _strtoul(const char *p, int base) 897 { 898 unsigned long retval; 899 900 errno = 0; /* clear */ 901 retval = strtoul(p, NULL, base); 902 if (errno == ERANGE && retval == ULONG_MAX) { 903 err(1, "strtoul"); 904 /* not reached */ 905 } 906 return retval; 907 } 908 909 static __inline __always_inline 910 unsigned long long 911 _strtoull(const char *p, int base) 912 { 913 unsigned long long retval; 914 915 errno = 0; /* clear */ 916 retval = strtoull(p, NULL, base); 917 if (errno == ERANGE && retval == ULLONG_MAX) { 918 err(1, "strtoull"); 919 /* not reached */ 920 } 921 return retval; 922 } 923 924 static 925 int 926 init_btree_search(const char *arg) 927 { 928 char *s, *p; 929 int i = 0; 930 931 bzero(&opt.base, sizeof(opt.base)); 932 opt.limit = 0; 933 934 if (arg == NULL) 935 return(-1); 936 if (strcmp(arg, "none") == 0) 937 return(-1); 938 939 s = strdup(arg); 940 if (s == NULL) 941 return(-1); 942 943 while ((p = s) != NULL) { 944 if ((s = strchr(s, ':')) != NULL) 945 *s++ = 0; 946 if (++i == 1) { 947 opt.base.localization = _strtoul(p, 16); 948 } else if (i == 2) { 949 opt.base.obj_id = _strtoull(p, 16); 950 } else if (i == 3) { 951 opt.base.rec_type = _strtoul(p, 16); 952 } else if (i == 4) { 953 opt.base.key = _strtoull(p, 16); 954 } else if (i == 5) { 955 opt.base.create_tid = _strtoull(p, 16); 956 break; 957 } 958 } 959 opt.limit = i; 960 free(s); 961 962 return(i); 963 } 964 965 static 966 int 967 test_btree_search(hammer_btree_elm_t elm) 968 { 969 hammer_base_elm_t base1 = &elm->base; 970 hammer_base_elm_t base2 = &opt.base; 971 int limit = opt.limit; 972 973 if (base1->localization < base2->localization) 974 return(-5); 975 if (base1->localization > base2->localization) 976 return(5); 977 if (limit == 1) 978 return(0); /* ignore below */ 979 980 if (base1->obj_id < base2->obj_id) 981 return(-4); 982 if (base1->obj_id > base2->obj_id) 983 return(4); 984 if (limit == 2) 985 return(0); /* ignore below */ 986 987 if (base1->rec_type < base2->rec_type) 988 return(-3); 989 if (base1->rec_type > base2->rec_type) 990 return(3); 991 if (limit == 3) 992 return(0); /* ignore below */ 993 994 if (base1->key < base2->key) 995 return(-2); 996 if (base1->key > base2->key) 997 return(2); 998 if (limit == 4) 999 return(0); /* ignore below */ 1000 1001 if (base1->create_tid == 0) { 1002 if (base2->create_tid == 0) 1003 return(0); 1004 return(1); 1005 } 1006 if (base2->create_tid == 0) 1007 return(-1); 1008 if (base1->create_tid < base2->create_tid) 1009 return(-1); 1010 if (base1->create_tid > base2->create_tid) 1011 return(1); 1012 return(0); 1013 } 1014 1015 static __inline 1016 int 1017 test_btree_match(hammer_btree_elm_t elm) 1018 { 1019 if (test_btree_search(elm) == 0) 1020 return(1); 1021 return(0); 1022 } 1023 1024 static 1025 int 1026 test_btree_out_of_range(hammer_btree_elm_t elm) 1027 { 1028 if (test_btree_search(elm) > 0) 1029 return(1); /* conditions < this elm */ 1030 1031 if (opt.limit >= 5) { 1032 if (test_btree_search(elm + 1) <= 0) 1033 return(1); /* next elm <= conditions */ 1034 } else { 1035 if (test_btree_search(elm + 1) < 0) 1036 return(1); /* next elm < conditions */ 1037 } 1038 return(0); 1039 } 1040 1041 /* 1042 * Dump the UNDO FIFO 1043 */ 1044 void 1045 hammer_cmd_show_undo(void) 1046 { 1047 volume_info_t volume; 1048 hammer_blockmap_t rootmap; 1049 hammer_off_t scan_offset; 1050 hammer_fifo_any_t head; 1051 hammer_fifo_head_t hdr; 1052 buffer_info_t data_buffer = NULL; 1053 zone_stat_t stats = NULL; 1054 1055 volume = get_root_volume(); 1056 rootmap = &volume->ondisk->vol0_blockmap[HAMMER_ZONE_UNDO_INDEX]; 1057 1058 print_blockmap(volume); 1059 1060 if (VerboseOpt) 1061 stats = hammer_init_zone_stat_bits(); 1062 1063 scan_offset = HAMMER_ENCODE_UNDO(0); 1064 while (scan_offset < rootmap->alloc_offset) { 1065 head = get_buffer_data(scan_offset, &data_buffer, 0); 1066 hdr = &head->head; 1067 printf("%016jx ", scan_offset); 1068 1069 switch(hdr->hdr_type) { 1070 case HAMMER_HEAD_TYPE_PAD: 1071 printf("PAD(%d)", hdr->hdr_size); 1072 break; 1073 case HAMMER_HEAD_TYPE_DUMMY: 1074 printf("DUMMY(%d)\tseq=%08x", 1075 hdr->hdr_size, hdr->hdr_seq); 1076 break; 1077 case HAMMER_HEAD_TYPE_UNDO: 1078 printf("UNDO(%d)\tseq=%08x offset=%016jx bytes=%d", 1079 hdr->hdr_size, hdr->hdr_seq, 1080 (intmax_t)head->undo.undo_offset, 1081 head->undo.undo_data_bytes); 1082 break; 1083 case HAMMER_HEAD_TYPE_REDO: 1084 printf("REDO(%d)\tseq=%08x offset=%016jx bytes=%d " 1085 "objid=%016jx flags=%08x lo=%08x", 1086 hdr->hdr_size, hdr->hdr_seq, 1087 (intmax_t)head->redo.redo_offset, 1088 head->redo.redo_data_bytes, 1089 (intmax_t)head->redo.redo_objid, 1090 head->redo.redo_flags, 1091 head->redo.redo_localization); 1092 break; 1093 default: 1094 printf("%04x(%d)\tseq=%08x", 1095 hdr->hdr_type, hdr->hdr_size, hdr->hdr_seq); 1096 break; 1097 } 1098 1099 if (scan_offset == rootmap->first_offset) 1100 printf(" >"); 1101 if (scan_offset == rootmap->next_offset) 1102 printf(" <"); 1103 printf("\n"); 1104 1105 if (stats) 1106 hammer_add_zone_stat(stats, scan_offset, hdr->hdr_size); 1107 1108 if ((hdr->hdr_size & HAMMER_HEAD_ALIGN_MASK) || 1109 hdr->hdr_size == 0 || 1110 hdr->hdr_size > HAMMER_UNDO_ALIGN - 1111 ((u_int)scan_offset & HAMMER_UNDO_MASK)) { 1112 printf("Illegal size field, skipping to " 1113 "next boundary\n"); 1114 scan_offset = HAMMER_UNDO_DOALIGN(scan_offset); 1115 } else { 1116 scan_offset += hdr->hdr_size; 1117 } 1118 } 1119 rel_buffer(data_buffer); 1120 1121 if (stats) { 1122 hammer_print_zone_stat(stats); 1123 hammer_cleanup_zone_stat(stats); 1124 } 1125 } 1126