]> git.saurik.com Git - apple/xnu.git/blob - bsd/hfs/hfs_btreeio.c
xnu-1228.7.58.tar.gz
[apple/xnu.git] / bsd / hfs / hfs_btreeio.c
1 /*
2 * Copyright (c) 2000-2008 Apple Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28
29 #include <sys/param.h>
30 #include <sys/systm.h>
31 #include <sys/buf.h>
32 #include <sys/buf_internal.h>
33 #include <sys/kernel.h>
34 #include <sys/malloc.h>
35 #include <sys/mount.h>
36 #include <sys/vnode.h>
37
38
39 #include "hfs.h"
40 #include "hfs_cnode.h"
41 #include "hfs_dbg.h"
42 #include "hfs_endian.h"
43 #include "hfs_btreeio.h"
44
45 #include "hfscommon/headers/FileMgrInternal.h"
46 #include "hfscommon/headers/BTreesPrivate.h"
47
48 #define FORCESYNCBTREEWRITES 0
49
50 /* From bsd/vfs/vfs_bio.c */
51 extern int bdwrite_internal(struct buf *, int);
52
53 static int ClearBTNodes(struct vnode *vp, long blksize, off_t offset, off_t amount);
54 static int btree_journal_modify_block_end(struct hfsmount *hfsmp, struct buf *bp);
55
56
57 __private_extern__
58 OSStatus SetBTreeBlockSize(FileReference vp, ByteCount blockSize, __unused ItemCount minBlockCount)
59 {
60 BTreeControlBlockPtr bTreePtr;
61
62 DBG_ASSERT(vp != NULL);
63 DBG_ASSERT(blockSize >= kMinNodeSize);
64 if (blockSize > MAXBSIZE )
65 return (fsBTBadNodeSize);
66
67 bTreePtr = (BTreeControlBlockPtr)VTOF(vp)->fcbBTCBPtr;
68 bTreePtr->nodeSize = blockSize;
69
70 return (E_NONE);
71 }
72
73
74 __private_extern__
75 OSStatus GetBTreeBlock(FileReference vp, u_int32_t blockNum, GetBlockOptions options, BlockDescriptor *block)
76 {
77 OSStatus retval = E_NONE;
78 struct buf *bp = NULL;
79 u_int8_t allow_empty_node;
80
81 /* If the btree block is being read using hint, it is
82 * fine for the swap code to find zeroed out nodes.
83 */
84 if (options & kGetBlockHint) {
85 allow_empty_node = true;
86 } else {
87 allow_empty_node = false;
88 }
89
90 if (options & kGetEmptyBlock) {
91 daddr64_t blkno;
92 off_t offset;
93
94 offset = (daddr64_t)blockNum * (daddr64_t)block->blockSize;
95 bp = buf_getblk(vp, (daddr64_t)blockNum, block->blockSize, 0, 0, BLK_META);
96 if (bp &&
97 VNOP_BLOCKMAP(vp, offset, block->blockSize, &blkno, NULL, NULL, 0, NULL) == 0) {
98 buf_setblkno(bp, blkno);
99 }
100 } else {
101 retval = buf_meta_bread(vp, (daddr64_t)blockNum, block->blockSize, NOCRED, &bp);
102 }
103 if (bp == NULL)
104 retval = -1; //XXX need better error
105
106 if (retval == E_NONE) {
107 block->blockHeader = bp;
108 block->buffer = (char *)buf_dataptr(bp);
109 block->blockNum = buf_lblkno(bp);
110 block->blockReadFromDisk = (buf_fromcache(bp) == 0); /* not found in cache ==> came from disk */
111
112 // XXXdbg
113 block->isModified = 0;
114
115 /* Check and endian swap B-Tree node (only if it's a valid block) */
116 if (!(options & kGetEmptyBlock)) {
117 /* This happens when we first open the b-tree, we might not have all the node data on hand */
118 if ((((BTNodeDescriptor *)block->buffer)->kind == kBTHeaderNode) &&
119 (((BTHeaderRec *)((char *)block->buffer + 14))->nodeSize != buf_count(bp)) &&
120 (SWAP_BE16 (((BTHeaderRec *)((char *)block->buffer + 14))->nodeSize) != buf_count(bp))) {
121
122 /*
123 * Don't swap the node descriptor, record offsets, or other records.
124 * This record will be invalidated and re-read with the correct node
125 * size once the B-tree control block is set up with the node size
126 * from the header record.
127 */
128 retval = hfs_swap_BTNode (block, vp, kSwapBTNodeHeaderRecordOnly, allow_empty_node);
129
130 } else if (block->blockReadFromDisk) {
131 /*
132 * The node was just read from disk, so always swap/check it.
133 * This is necessary on big endian since the test below won't trigger.
134 */
135 retval = hfs_swap_BTNode (block, vp, kSwapBTNodeBigToHost, allow_empty_node);
136 } else if (*((u_int16_t *)((char *)block->buffer + (block->blockSize - sizeof (u_int16_t)))) == 0x0e00) {
137 /*
138 * The node was left in the cache in non-native order, so swap it.
139 * This only happens on little endian, after the node is written
140 * back to disk.
141 */
142 retval = hfs_swap_BTNode (block, vp, kSwapBTNodeBigToHost, allow_empty_node);
143 }
144
145 /*
146 * If we got an error, then the node is only partially swapped.
147 * We mark the buffer invalid so that the next attempt to get the
148 * node will read it and attempt to swap again, and will notice
149 * the error again. If we didn't do this, the next attempt to get
150 * the node might use the partially swapped node as-is.
151 */
152 if (retval)
153 buf_markinvalid(bp);
154 }
155 }
156
157 if (retval) {
158 if (bp)
159 buf_brelse(bp);
160 block->blockHeader = NULL;
161 block->buffer = NULL;
162 }
163
164 return (retval);
165 }
166
167
168 __private_extern__
169 void ModifyBlockStart(FileReference vp, BlockDescPtr blockPtr)
170 {
171 struct hfsmount *hfsmp = VTOHFS(vp);
172 struct buf *bp = NULL;
173
174 if (hfsmp->jnl == NULL) {
175 return;
176 }
177
178 bp = (struct buf *) blockPtr->blockHeader;
179 if (bp == NULL) {
180 panic("ModifyBlockStart: null bp for blockdescptr %p?!?\n", blockPtr);
181 return;
182 }
183
184 journal_modify_block_start(hfsmp->jnl, bp);
185 blockPtr->isModified = 1;
186 }
187
188 static void
189 btree_swap_node(struct buf *bp, __unused void *arg)
190 {
191 // struct hfsmount *hfsmp = (struct hfsmount *)arg;
192 int retval;
193 struct vnode *vp = buf_vnode(bp);
194 BlockDescriptor block;
195
196 /* Prepare the block pointer */
197 block.blockHeader = bp;
198 block.buffer = (char *)buf_dataptr(bp);
199 block.blockNum = buf_lblkno(bp);
200 /* not found in cache ==> came from disk */
201 block.blockReadFromDisk = (buf_fromcache(bp) == 0);
202 block.blockSize = buf_count(bp);
203
204 /* Swap the data now that this node is ready to go to disk.
205 * We allow swapping of zeroed out nodes here because we might
206 * be writing node whose last record just got deleted.
207 */
208 retval = hfs_swap_BTNode (&block, vp, kSwapBTNodeHostToBig, true);
209 if (retval)
210 panic("btree_swap_node: about to write corrupt node!\n");
211 }
212
213
214 static int
215 btree_journal_modify_block_end(struct hfsmount *hfsmp, struct buf *bp)
216 {
217 return journal_modify_block_end(hfsmp->jnl, bp, btree_swap_node, hfsmp);
218 }
219
220
221 __private_extern__
222 OSStatus ReleaseBTreeBlock(FileReference vp, BlockDescPtr blockPtr, ReleaseBlockOptions options)
223 {
224 struct hfsmount *hfsmp = VTOHFS(vp);
225 OSStatus retval = E_NONE;
226 struct buf *bp = NULL;
227
228 bp = (struct buf *) blockPtr->blockHeader;
229
230 if (bp == NULL) {
231 retval = -1;
232 goto exit;
233 }
234
235 if (options & kTrashBlock) {
236 buf_markinvalid(bp);
237
238 if (hfsmp->jnl && (buf_flags(bp) & B_LOCKED)) {
239 journal_kill_block(hfsmp->jnl, bp);
240 } else {
241 buf_brelse(bp); /* note: B-tree code will clear blockPtr->blockHeader and blockPtr->buffer */
242 }
243 } else {
244 if (options & kForceWriteBlock) {
245 if (hfsmp->jnl) {
246 if (blockPtr->isModified == 0) {
247 panic("hfs: releaseblock: modified is 0 but forcewrite set! bp %p\n", bp);
248 }
249
250 retval = btree_journal_modify_block_end(hfsmp, bp);
251 blockPtr->isModified = 0;
252 } else {
253 retval = VNOP_BWRITE(bp);
254 }
255 } else if (options & kMarkBlockDirty) {
256 struct timeval tv;
257 microuptime(&tv);
258 if ((options & kLockTransaction) && hfsmp->jnl == NULL) {
259 /*
260 *
261 * Set the B_LOCKED flag and unlock the buffer, causing buf_brelse to move
262 * the buffer onto the LOCKED free list. This is necessary, otherwise
263 * getnewbuf() would try to reclaim the buffers using buf_bawrite, which
264 * isn't going to work.
265 *
266 */
267 /* Don't hog all the buffers... */
268 if (count_lock_queue() > kMaxLockedMetaBuffers) {
269 hfs_btsync(vp, HFS_SYNCTRANS);
270 /* Rollback sync time to cause a sync on lock release... */
271 (void) BTSetLastSync(VTOF(vp), tv.tv_sec - (kMaxSecsForFsync + 1));
272 }
273 buf_setflags(bp, B_LOCKED);
274 }
275
276 /*
277 * Delay-write this block.
278 * If the maximum delayed buffers has been exceeded then
279 * free up some buffers and fall back to an asynchronous write.
280 */
281 if (hfsmp->jnl) {
282 if (blockPtr->isModified == 0) {
283 panic("hfs: releaseblock: modified is 0 but markdirty set! bp %p\n", bp);
284 }
285 retval = btree_journal_modify_block_end(hfsmp, bp);
286 blockPtr->isModified = 0;
287 } else if (bdwrite_internal(bp, 1) != 0) {
288 hfs_btsync(vp, 0);
289 /* Rollback sync time to cause a sync on lock release... */
290 (void) BTSetLastSync(VTOF(vp), tv.tv_sec - (kMaxSecsForFsync + 1));
291
292 buf_clearflags(bp, B_LOCKED);
293 buf_bawrite(bp);
294 }
295 } else {
296 // check if we had previously called journal_modify_block_start()
297 // on this block and if so, abort it (which will call buf_brelse()).
298 if (hfsmp->jnl && blockPtr->isModified) {
299 // XXXdbg - I don't want to call modify_block_abort()
300 // because I think it may be screwing up the
301 // journal and blowing away a block that has
302 // valid data in it.
303 //
304 // journal_modify_block_abort(hfsmp->jnl, bp);
305 //panic("hfs: releaseblock called for 0x%x but mod_block_start previously called.\n", bp);
306 btree_journal_modify_block_end(hfsmp, bp);
307 blockPtr->isModified = 0;
308 } else {
309 buf_brelse(bp); /* note: B-tree code will clear blockPtr->blockHeader and blockPtr->buffer */
310 }
311 };
312 };
313
314 exit:
315 return (retval);
316 }
317
318
319 __private_extern__
320 OSStatus ExtendBTreeFile(FileReference vp, FSSize minEOF, FSSize maxEOF)
321 {
322 #pragma unused (maxEOF)
323
324 OSStatus retval = 0, ret = 0;
325 int64_t actualBytesAdded, origSize;
326 u_int64_t bytesToAdd;
327 u_int32_t startAllocation;
328 u_int32_t fileblocks;
329 BTreeInfoRec btInfo;
330 ExtendedVCB *vcb;
331 FCB *filePtr;
332 struct proc *p = NULL;
333 int64_t trim = 0;
334 int lockflags = 0;
335
336 filePtr = GetFileControlBlock(vp);
337
338 if ( (off_t)minEOF > filePtr->fcbEOF )
339 {
340 bytesToAdd = minEOF - filePtr->fcbEOF;
341
342 if (bytesToAdd < filePtr->ff_clumpsize)
343 bytesToAdd = filePtr->ff_clumpsize; //XXX why not always be a mutiple of clump size?
344 }
345 else
346 {
347 return -1;
348 }
349
350 vcb = VTOVCB(vp);
351
352 /*
353 * The Extents B-tree can't have overflow extents. ExtendFileC will
354 * return an error if an attempt is made to extend the Extents B-tree
355 * when the resident extents are exhausted.
356 */
357
358 /* Protect allocation bitmap and extents overflow file. */
359 lockflags = SFL_BITMAP;
360 if (VTOC(vp)->c_fileid != kHFSExtentsFileID)
361 lockflags |= SFL_EXTENTS;
362 lockflags = hfs_systemfile_lock(vcb, lockflags, HFS_EXCLUSIVE_LOCK);
363
364 (void) BTGetInformation(filePtr, 0, &btInfo);
365
366 #if 0 // XXXdbg
367 /*
368 * The b-tree code expects nodes to be contiguous. So when
369 * the allocation block size is less than the b-tree node
370 * size, we need to force disk allocations to be contiguous.
371 */
372 if (vcb->blockSize >= btInfo.nodeSize) {
373 extendFlags = 0;
374 } else {
375 /* Ensure that all b-tree nodes are contiguous on disk */
376 extendFlags = kEFContigMask;
377 }
378 #endif
379
380 origSize = filePtr->fcbEOF;
381 fileblocks = filePtr->ff_blocks;
382 startAllocation = vcb->nextAllocation;
383
384 // loop trying to get a contiguous chunk that's an integer multiple
385 // of the btree node size. if we can't get a contiguous chunk that
386 // is at least the node size then we break out of the loop and let
387 // the error propagate back up.
388 while((off_t)bytesToAdd >= btInfo.nodeSize) {
389 do {
390 retval = ExtendFileC(vcb, filePtr, bytesToAdd, 0,
391 kEFContigMask | kEFMetadataMask | kEFNoClumpMask,
392 (int64_t *)&actualBytesAdded);
393 if (retval == dskFulErr && actualBytesAdded == 0) {
394 bytesToAdd >>= 1;
395 if (bytesToAdd < btInfo.nodeSize) {
396 break;
397 } else if ((bytesToAdd % btInfo.nodeSize) != 0) {
398 // make sure it's an integer multiple of the nodeSize
399 bytesToAdd -= (bytesToAdd % btInfo.nodeSize);
400 }
401 }
402 } while (retval == dskFulErr && actualBytesAdded == 0);
403
404 if (retval == dskFulErr && actualBytesAdded == 0 && bytesToAdd <= btInfo.nodeSize) {
405 break;
406 }
407
408 filePtr->fcbEOF = (u_int64_t)filePtr->ff_blocks * (u_int64_t)vcb->blockSize;
409 bytesToAdd = minEOF - filePtr->fcbEOF;
410 }
411
412 /*
413 * If a new extent was added then move the roving allocator
414 * reference forward by the current b-tree file size so
415 * there's plenty of room to grow.
416 */
417 if ((retval == 0) &&
418 ((VCBTOHFS(vcb)->hfs_flags & HFS_METADATA_ZONE) == 0) &&
419 (vcb->nextAllocation > startAllocation) &&
420 ((vcb->nextAllocation + fileblocks) < vcb->allocLimit)) {
421 HFS_UPDATE_NEXT_ALLOCATION(vcb, vcb->nextAllocation + fileblocks);
422 }
423
424 filePtr->fcbEOF = (u_int64_t)filePtr->ff_blocks * (u_int64_t)vcb->blockSize;
425
426 // XXXdbg ExtendFileC() could have returned an error even though
427 // it grew the file to be big enough for our needs. If this is
428 // the case, we don't care about retval so we blow it away.
429 //
430 if (filePtr->fcbEOF >= (off_t)minEOF && retval != 0) {
431 retval = 0;
432 }
433
434 // XXXdbg if the file grew but isn't large enough or isn't an
435 // even multiple of the nodeSize then trim things back. if
436 // the file isn't large enough we trim back to the original
437 // size. otherwise we trim back to be an even multiple of the
438 // btree node size.
439 //
440 if ((filePtr->fcbEOF < (off_t)minEOF) || ((filePtr->fcbEOF - origSize) % btInfo.nodeSize) != 0) {
441
442 if (filePtr->fcbEOF < (off_t)minEOF) {
443 retval = dskFulErr;
444
445 if (filePtr->fcbEOF < origSize) {
446 panic("hfs: btree file eof %lld less than orig size %lld!\n",
447 filePtr->fcbEOF, origSize);
448 }
449
450 trim = filePtr->fcbEOF - origSize;
451 } else {
452 trim = ((filePtr->fcbEOF - origSize) % btInfo.nodeSize);
453 }
454
455 ret = TruncateFileC(vcb, filePtr, filePtr->fcbEOF - trim, 0);
456 filePtr->fcbEOF = (u_int64_t)filePtr->ff_blocks * (u_int64_t)vcb->blockSize;
457
458 // XXXdbg - panic if the file didn't get trimmed back properly
459 if ((filePtr->fcbEOF % btInfo.nodeSize) != 0) {
460 panic("hfs: truncate file didn't! fcbEOF %lld nsize %d fcb %p\n",
461 filePtr->fcbEOF, btInfo.nodeSize, filePtr);
462 }
463
464 if (ret) {
465 // XXXdbg - this probably doesn't need to be a panic()
466 panic("hfs: error truncating btree files (sz 0x%llx, trim %lld, ret %ld)\n",
467 filePtr->fcbEOF, trim, ret);
468 goto out;
469 }
470 }
471
472 if(VTOC(vp)->c_fileid != kHFSExtentsFileID) {
473 /*
474 * Get any extents overflow b-tree changes to disk ASAP!
475 */
476 (void) BTFlushPath(VTOF(vcb->extentsRefNum));
477 (void) hfs_fsync(vcb->extentsRefNum, MNT_WAIT, 0, p);
478 }
479 hfs_systemfile_unlock(vcb, lockflags);
480 lockflags = 0;
481
482 if ((filePtr->fcbEOF % btInfo.nodeSize) != 0) {
483 panic("hfs: extendbtree: fcb %p has eof 0x%llx not a multiple of 0x%x (trim %llx)\n",
484 filePtr, filePtr->fcbEOF, btInfo.nodeSize, trim);
485 }
486
487 /*
488 * Update the Alternate MDB or Alternate VolumeHeader
489 */
490 if ((VTOC(vp)->c_fileid == kHFSExtentsFileID) ||
491 (VTOC(vp)->c_fileid == kHFSCatalogFileID) ||
492 (VTOC(vp)->c_fileid == kHFSAttributesFileID)
493 ) {
494 VTOC(vp)->c_flag |= C_MODIFIED;
495 MarkVCBDirty( vcb );
496 ret = hfs_flushvolumeheader(VCBTOHFS(vcb), MNT_WAIT, HFS_ALTFLUSH);
497 } else {
498 VTOC(vp)->c_touch_chgtime = TRUE;
499 VTOC(vp)->c_touch_modtime = TRUE;
500 (void) hfs_update(vp, TRUE);
501 }
502
503 ret = ClearBTNodes(vp, btInfo.nodeSize, origSize, (filePtr->fcbEOF - origSize));
504 out:
505 if (retval == 0)
506 retval = ret;
507
508 if (lockflags)
509 hfs_systemfile_unlock(vcb, lockflags);
510
511 return retval;
512 }
513
514
515 /*
516 * Clear out (zero) new b-tree nodes on disk.
517 */
518 static int
519 ClearBTNodes(struct vnode *vp, long blksize, off_t offset, off_t amount)
520 {
521 struct hfsmount *hfsmp = VTOHFS(vp);
522 struct buf *bp = NULL;
523 daddr64_t blk;
524 daddr64_t blkcnt;
525
526 blk = offset / blksize;
527 blkcnt = amount / blksize;
528
529 while (blkcnt > 0) {
530 bp = buf_getblk(vp, blk, blksize, 0, 0, BLK_META);
531 if (bp == NULL)
532 continue;
533
534 // XXXdbg
535 if (hfsmp->jnl) {
536 // XXXdbg -- skipping this for now since it makes a transaction
537 // become *way* too large
538 //journal_modify_block_start(hfsmp->jnl, bp);
539 }
540 bzero((char *)buf_dataptr(bp), blksize);
541
542 buf_markaged(bp);
543
544 // XXXdbg
545 if (hfsmp->jnl) {
546 // XXXdbg -- skipping this for now since it makes a transaction
547 // become *way* too large
548 //journal_modify_block_end(hfsmp->jnl, bp);
549
550 // XXXdbg - remove this once we decide what to do with the
551 // writes to the journal
552 if ((blk % 32) == 0)
553 VNOP_BWRITE(bp);
554 else
555 buf_bawrite(bp);
556 } else {
557 /* wait/yield every 32 blocks so we don't hog all the buffers */
558 if ((blk % 32) == 0)
559 VNOP_BWRITE(bp);
560 else
561 buf_bawrite(bp);
562 }
563 --blkcnt;
564 ++blk;
565 }
566
567 return (0);
568 }
569
570
571 extern char hfs_attrname[];
572
573 /*
574 * Create an HFS+ Attribute B-tree File.
575 *
576 * No global resources should be held.
577 */
578 int
579 hfs_create_attr_btree(struct hfsmount *hfsmp, u_int32_t nodesize, u_int32_t nodecnt)
580 {
581 struct vnode* vp = NULLVP;
582 struct cat_desc cndesc;
583 struct cat_attr cnattr;
584 struct cat_fork cfork;
585 BlockDescriptor blkdesc;
586 BTNodeDescriptor *ndp;
587 BTHeaderRec *bthp;
588 BTreeControlBlockPtr btcb = NULL;
589 struct buf *bp = NULL;
590 void * buffer;
591 u_int16_t *index;
592 u_int16_t offset;
593 int intrans = 0;
594 int result;
595 again:
596 /*
597 * Serialize creation using HFS_CREATING_BTREE flag.
598 */
599 lck_mtx_lock(&hfsmp->hfs_mutex);
600 if (hfsmp->hfs_flags & HFS_CREATING_BTREE) {
601 /* Someone else beat us, wait for them to finish. */
602 (void) msleep(hfsmp->hfs_attribute_cp, &hfsmp->hfs_mutex,
603 PDROP | PINOD, "hfs_create_attr_btree", 0);
604 if (hfsmp->hfs_attribute_vp) {
605 return (0);
606 }
607 goto again;
608 }
609 hfsmp->hfs_flags |= HFS_CREATING_BTREE;
610 lck_mtx_unlock(&hfsmp->hfs_mutex);
611
612 /* Check if were out of usable disk space. */
613 if ((hfs_freeblks(hfsmp, 1) == 0)) {
614 result = ENOSPC;
615 goto exit;
616 }
617
618 /*
619 * Set up Attribute B-tree vnode
620 * (this must be done before we start a transaction
621 * or take any system file locks)
622 */
623 bzero(&cndesc, sizeof(cndesc));
624 cndesc.cd_parentcnid = kHFSRootParentID;
625 cndesc.cd_flags |= CD_ISMETA;
626 cndesc.cd_nameptr = (const u_int8_t *)hfs_attrname;
627 cndesc.cd_namelen = strlen(hfs_attrname);
628 cndesc.cd_cnid = kHFSAttributesFileID;
629
630 bzero(&cnattr, sizeof(cnattr));
631 cnattr.ca_linkcount = 1;
632 cnattr.ca_mode = S_IFREG;
633 cnattr.ca_fileid = cndesc.cd_cnid;
634
635 bzero(&cfork, sizeof(cfork));
636 cfork.cf_clump = nodesize * nodecnt;
637
638 result = hfs_getnewvnode(hfsmp, NULL, NULL, &cndesc, 0, &cnattr, &cfork, &vp);
639 if (result) {
640 goto exit;
641 }
642 /*
643 * Set up Attribute B-tree control block
644 */
645 MALLOC(btcb, BTreeControlBlock *, sizeof(BTreeControlBlock), M_TEMP, M_WAITOK);
646 bzero(btcb, sizeof(BTreeControlBlock));
647
648 btcb->nodeSize = nodesize;
649 btcb->maxKeyLength = kHFSPlusAttrKeyMaximumLength;
650 btcb->btreeType = 0xFF;
651 btcb->attributes = kBTVariableIndexKeysMask | kBTBigKeysMask;
652 btcb->version = kBTreeVersion;
653 btcb->writeCount = 1;
654 btcb->flags = 0; /* kBTHeaderDirty */
655 btcb->fileRefNum = vp;
656 btcb->getBlockProc = GetBTreeBlock;
657 btcb->releaseBlockProc = ReleaseBTreeBlock;
658 btcb->setEndOfForkProc = ExtendBTreeFile;
659 btcb->keyCompareProc = (KeyCompareProcPtr)hfs_attrkeycompare;
660 VTOF(vp)->fcbBTCBPtr = btcb;
661
662 /*
663 * Allocate some space
664 */
665 if (hfs_start_transaction(hfsmp) != 0) {
666 result = EINVAL;
667 goto exit;
668 }
669 intrans = 1;
670
671 /* Note ExtendBTreeFile will acquire the necessary system file locks. */
672 result = ExtendBTreeFile(vp, nodesize, cfork.cf_clump);
673 if (result)
674 goto exit;
675
676 btcb->totalNodes = VTOF(vp)->ff_size / nodesize;
677 btcb->freeNodes = btcb->totalNodes - 1;
678
679 /*
680 * Initialize the b-tree header on disk
681 */
682 bp = buf_getblk(vp, 0, nodesize, 0, 0, BLK_META);
683 if (bp == NULL) {
684 result = EIO;
685 goto exit;
686 }
687
688 buffer = (void *)buf_dataptr(bp);
689 blkdesc.buffer = buffer;
690 blkdesc.blockHeader = (void *)bp;
691 blkdesc.blockReadFromDisk = 0;
692 blkdesc.isModified = 0;
693
694 ModifyBlockStart(vp, &blkdesc);
695
696 if (buf_size(bp) != nodesize)
697 panic("hfs_create_attr_btree: bad buffer size (%d)\n", buf_size(bp));
698
699 bzero(buffer, nodesize);
700 index = (u_int16_t *)buffer;
701
702 /* FILL IN THE NODE DESCRIPTOR: */
703 ndp = (BTNodeDescriptor *)buffer;
704 ndp->kind = kBTHeaderNode;
705 ndp->numRecords = 3;
706 offset = sizeof(BTNodeDescriptor);
707 index[(nodesize / 2) - 1] = offset;
708
709 /* FILL IN THE HEADER RECORD: */
710 bthp = (BTHeaderRec *)((u_int8_t *)buffer + offset);
711 bthp->nodeSize = nodesize;
712 bthp->totalNodes = btcb->totalNodes;
713 bthp->freeNodes = btcb->freeNodes;
714 bthp->clumpSize = cfork.cf_clump;
715 bthp->btreeType = 0xFF;
716 bthp->attributes = kBTVariableIndexKeysMask | kBTBigKeysMask;
717 bthp->maxKeyLength = kHFSPlusAttrKeyMaximumLength;
718 bthp->keyCompareType = kHFSBinaryCompare;
719 offset += sizeof(BTHeaderRec);
720 index[(nodesize / 2) - 2] = offset;
721
722 /* FILL IN THE USER RECORD: */
723 offset += kBTreeHeaderUserBytes;
724 index[(nodesize / 2) - 3] = offset;
725
726 /* FILL IN THE MAP RECORD (only one node in use). */
727 *((u_int8_t *)buffer + offset) = 0x80;
728 offset += nodesize - sizeof(BTNodeDescriptor) - sizeof(BTHeaderRec)
729 - kBTreeHeaderUserBytes - (4 * sizeof(int16_t));
730 index[(nodesize / 2) - 4] = offset;
731
732 if (hfsmp->jnl) {
733 result = btree_journal_modify_block_end(hfsmp, bp);
734 } else {
735 result = VNOP_BWRITE(bp);
736 }
737 if (result)
738 goto exit;
739
740 /* Update vp/cp for attribute btree */
741 lck_mtx_lock(&hfsmp->hfs_mutex);
742 hfsmp->hfs_attribute_cp = VTOC(vp);
743 hfsmp->hfs_attribute_vp = vp;
744 lck_mtx_unlock(&hfsmp->hfs_mutex);
745
746 (void) hfs_flushvolumeheader(hfsmp, MNT_WAIT, HFS_ALTFLUSH);
747 exit:
748 if (vp) {
749 hfs_unlock(VTOC(vp));
750 }
751 if (result) {
752 if (btcb) {
753 FREE (btcb, M_TEMP);
754 }
755 if (vp) {
756 vnode_put(vp);
757 }
758 /* XXX need to give back blocks ? */
759 }
760 if (intrans) {
761 hfs_end_transaction(hfsmp);
762 }
763
764 /*
765 * All done, clear HFS_CREATING_BTREE, and wake up any sleepers.
766 */
767 lck_mtx_lock(&hfsmp->hfs_mutex);
768 hfsmp->hfs_flags &= ~HFS_CREATING_BTREE;
769 wakeup((caddr_t)hfsmp->hfs_attribute_cp);
770 lck_mtx_unlock(&hfsmp->hfs_mutex);
771
772 return (result);
773 }
774