]> git.saurik.com Git - apple/xnu.git/blame - bsd/hfs/hfs_btreeio.c
xnu-1504.9.37.tar.gz
[apple/xnu.git] / bsd / hfs / hfs_btreeio.c
CommitLineData
1c79356b 1/*
935ed37a 2 * Copyright (c) 2000-2008 Apple Inc. All rights reserved.
5d5c5d0d 3 *
2d21ac55 4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
1c79356b 5 *
2d21ac55
A
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
8f6c56a5 14 *
2d21ac55
A
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
8f6c56a5
A
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
2d21ac55
A
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
8f6c56a5 25 *
2d21ac55 26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
1c79356b 27 */
1c79356b
A
28
29#include <sys/param.h>
30#include <sys/systm.h>
31#include <sys/buf.h>
2d21ac55 32#include <sys/buf_internal.h>
9bccf70c 33#include <sys/kernel.h>
91447636 34#include <sys/malloc.h>
1c79356b
A
35#include <sys/mount.h>
36#include <sys/vnode.h>
37
38
39#include "hfs.h"
9bccf70c 40#include "hfs_cnode.h"
1c79356b
A
41#include "hfs_dbg.h"
42#include "hfs_endian.h"
2d21ac55 43#include "hfs_btreeio.h"
1c79356b
A
44
45#include "hfscommon/headers/FileMgrInternal.h"
46#include "hfscommon/headers/BTreesPrivate.h"
47
48#define FORCESYNCBTREEWRITES 0
49
2d21ac55
A
50/* From bsd/vfs/vfs_bio.c */
51extern int bdwrite_internal(struct buf *, int);
1c79356b
A
52
53static int ClearBTNodes(struct vnode *vp, long blksize, off_t offset, off_t amount);
2d21ac55 54static int btree_journal_modify_block_end(struct hfsmount *hfsmp, struct buf *bp);
1c79356b
A
55
56
9bccf70c 57__private_extern__
2d21ac55 58OSStatus SetBTreeBlockSize(FileReference vp, ByteCount blockSize, __unused ItemCount minBlockCount)
1c79356b
A
59{
60 BTreeControlBlockPtr bTreePtr;
61
62 DBG_ASSERT(vp != NULL);
1c79356b
A
63 DBG_ASSERT(blockSize >= kMinNodeSize);
64 if (blockSize > MAXBSIZE )
65 return (fsBTBadNodeSize);
66
9bccf70c 67 bTreePtr = (BTreeControlBlockPtr)VTOF(vp)->fcbBTCBPtr;
1c79356b
A
68 bTreePtr->nodeSize = blockSize;
69
70 return (E_NONE);
71}
72
73
9bccf70c 74__private_extern__
2d21ac55 75OSStatus GetBTreeBlock(FileReference vp, u_int32_t blockNum, GetBlockOptions options, BlockDescriptor *block)
1c79356b
A
76{
77 OSStatus retval = E_NONE;
78 struct buf *bp = NULL;
935ed37a
A
79 u_int8_t allow_empty_node;
80
81 /* If the btree block is being read using hint, it is
82 * fine for the swap code to find zeroed out nodes.
83 */
84 if (options & kGetBlockHint) {
85 allow_empty_node = true;
86 } else {
87 allow_empty_node = false;
88 }
1c79356b 89
91447636
A
90 if (options & kGetEmptyBlock) {
91 daddr64_t blkno;
92 off_t offset;
1c79356b 93
91447636
A
94 offset = (daddr64_t)blockNum * (daddr64_t)block->blockSize;
95 bp = buf_getblk(vp, (daddr64_t)blockNum, block->blockSize, 0, 0, BLK_META);
96 if (bp &&
97 VNOP_BLOCKMAP(vp, offset, block->blockSize, &blkno, NULL, NULL, 0, NULL) == 0) {
98 buf_setblkno(bp, blkno);
99 }
100 } else {
101 retval = buf_meta_bread(vp, (daddr64_t)blockNum, block->blockSize, NOCRED, &bp);
102 }
1c79356b
A
103 if (bp == NULL)
104 retval = -1; //XXX need better error
105
106 if (retval == E_NONE) {
107 block->blockHeader = bp;
91447636 108 block->buffer = (char *)buf_dataptr(bp);
3a60a9f5 109 block->blockNum = buf_lblkno(bp);
91447636 110 block->blockReadFromDisk = (buf_fromcache(bp) == 0); /* not found in cache ==> came from disk */
1c79356b 111
b4c24cb9
A
112 // XXXdbg
113 block->isModified = 0;
114
3a60a9f5 115 /* Check and endian swap B-Tree node (only if it's a valid block) */
1c79356b
A
116 if (!(options & kGetEmptyBlock)) {
117 /* This happens when we first open the b-tree, we might not have all the node data on hand */
118 if ((((BTNodeDescriptor *)block->buffer)->kind == kBTHeaderNode) &&
91447636
A
119 (((BTHeaderRec *)((char *)block->buffer + 14))->nodeSize != buf_count(bp)) &&
120 (SWAP_BE16 (((BTHeaderRec *)((char *)block->buffer + 14))->nodeSize) != buf_count(bp))) {
1c79356b 121
3a60a9f5
A
122 /*
123 * Don't swap the node descriptor, record offsets, or other records.
124 * This record will be invalidated and re-read with the correct node
125 * size once the B-tree control block is set up with the node size
126 * from the header record.
127 */
935ed37a 128 retval = hfs_swap_BTNode (block, vp, kSwapBTNodeHeaderRecordOnly, allow_empty_node);
3a60a9f5
A
129
130 } else if (block->blockReadFromDisk) {
131 /*
132 * The node was just read from disk, so always swap/check it.
133 * This is necessary on big endian since the test below won't trigger.
134 */
935ed37a 135 retval = hfs_swap_BTNode (block, vp, kSwapBTNodeBigToHost, allow_empty_node);
2d21ac55 136 } else if (*((u_int16_t *)((char *)block->buffer + (block->blockSize - sizeof (u_int16_t)))) == 0x0e00) {
3a60a9f5
A
137 /*
138 * The node was left in the cache in non-native order, so swap it.
139 * This only happens on little endian, after the node is written
140 * back to disk.
141 */
935ed37a 142 retval = hfs_swap_BTNode (block, vp, kSwapBTNodeBigToHost, allow_empty_node);
1c79356b 143 }
3a60a9f5
A
144
145 /*
146 * If we got an error, then the node is only partially swapped.
147 * We mark the buffer invalid so that the next attempt to get the
148 * node will read it and attempt to swap again, and will notice
149 * the error again. If we didn't do this, the next attempt to get
150 * the node might use the partially swapped node as-is.
151 */
152 if (retval)
153 buf_markinvalid(bp);
1c79356b 154 }
3a60a9f5
A
155 }
156
157 if (retval) {
1c79356b 158 if (bp)
3a60a9f5 159 buf_brelse(bp);
1c79356b
A
160 block->blockHeader = NULL;
161 block->buffer = NULL;
162 }
163
164 return (retval);
165}
166
167
b4c24cb9
A
168__private_extern__
169void ModifyBlockStart(FileReference vp, BlockDescPtr blockPtr)
170{
171 struct hfsmount *hfsmp = VTOHFS(vp);
172 struct buf *bp = NULL;
173
174 if (hfsmp->jnl == NULL) {
175 return;
176 }
177
178 bp = (struct buf *) blockPtr->blockHeader;
179 if (bp == NULL) {
b0d623f7 180 panic("hfs: ModifyBlockStart: null bp for blockdescptr %p?!?\n", blockPtr);
b4c24cb9
A
181 return;
182 }
183
184 journal_modify_block_start(hfsmp->jnl, bp);
185 blockPtr->isModified = 1;
186}
187
2d21ac55
A
188static void
189btree_swap_node(struct buf *bp, __unused void *arg)
55e303ae 190{
2d21ac55 191 // struct hfsmount *hfsmp = (struct hfsmount *)arg;
3a60a9f5 192 int retval;
91447636 193 struct vnode *vp = buf_vnode(bp);
55e303ae
A
194 BlockDescriptor block;
195
196 /* Prepare the block pointer */
197 block.blockHeader = bp;
91447636 198 block.buffer = (char *)buf_dataptr(bp);
3a60a9f5 199 block.blockNum = buf_lblkno(bp);
55e303ae 200 /* not found in cache ==> came from disk */
91447636
A
201 block.blockReadFromDisk = (buf_fromcache(bp) == 0);
202 block.blockSize = buf_count(bp);
55e303ae 203
935ed37a
A
204 /* Swap the data now that this node is ready to go to disk.
205 * We allow swapping of zeroed out nodes here because we might
206 * be writing node whose last record just got deleted.
207 */
208 retval = hfs_swap_BTNode (&block, vp, kSwapBTNodeHostToBig, true);
3a60a9f5 209 if (retval)
b0d623f7 210 panic("hfs: btree_swap_node: about to write corrupt node!\n");
2d21ac55
A
211}
212
55e303ae 213
2d21ac55
A
214static int
215btree_journal_modify_block_end(struct hfsmount *hfsmp, struct buf *bp)
216{
217 return journal_modify_block_end(hfsmp->jnl, bp, btree_swap_node, hfsmp);
55e303ae
A
218}
219
b4c24cb9 220
9bccf70c 221__private_extern__
1c79356b
A
222OSStatus ReleaseBTreeBlock(FileReference vp, BlockDescPtr blockPtr, ReleaseBlockOptions options)
223{
b4c24cb9 224 struct hfsmount *hfsmp = VTOHFS(vp);
1c79356b
A
225 OSStatus retval = E_NONE;
226 struct buf *bp = NULL;
227
228 bp = (struct buf *) blockPtr->blockHeader;
229
230 if (bp == NULL) {
1c79356b
A
231 retval = -1;
232 goto exit;
233 }
234
235 if (options & kTrashBlock) {
91447636
A
236 buf_markinvalid(bp);
237
238 if (hfsmp->jnl && (buf_flags(bp) & B_LOCKED)) {
b4c24cb9
A
239 journal_kill_block(hfsmp->jnl, bp);
240 } else {
91447636 241 buf_brelse(bp); /* note: B-tree code will clear blockPtr->blockHeader and blockPtr->buffer */
b4c24cb9 242 }
d1ecb069
A
243
244 /* Don't let anyone else try to use this bp, it's been consumed */
245 blockPtr->blockHeader = NULL;
246
1c79356b
A
247 } else {
248 if (options & kForceWriteBlock) {
b4c24cb9
A
249 if (hfsmp->jnl) {
250 if (blockPtr->isModified == 0) {
2d21ac55 251 panic("hfs: releaseblock: modified is 0 but forcewrite set! bp %p\n", bp);
b4c24cb9 252 }
55e303ae
A
253
254 retval = btree_journal_modify_block_end(hfsmp, bp);
b4c24cb9
A
255 blockPtr->isModified = 0;
256 } else {
91447636 257 retval = VNOP_BWRITE(bp);
b4c24cb9 258 }
d1ecb069
A
259
260 /* Don't let anyone else try to use this bp, it's been consumed */
261 blockPtr->blockHeader = NULL;
262
1c79356b 263 } else if (options & kMarkBlockDirty) {
91447636
A
264 struct timeval tv;
265 microuptime(&tv);
b4c24cb9 266 if ((options & kLockTransaction) && hfsmp->jnl == NULL) {
9bccf70c
A
267 /*
268 *
91447636 269 * Set the B_LOCKED flag and unlock the buffer, causing buf_brelse to move
9bccf70c 270 * the buffer onto the LOCKED free list. This is necessary, otherwise
91447636 271 * getnewbuf() would try to reclaim the buffers using buf_bawrite, which
9bccf70c
A
272 * isn't going to work.
273 *
274 */
9bccf70c
A
275 /* Don't hog all the buffers... */
276 if (count_lock_queue() > kMaxLockedMetaBuffers) {
277 hfs_btsync(vp, HFS_SYNCTRANS);
278 /* Rollback sync time to cause a sync on lock release... */
91447636 279 (void) BTSetLastSync(VTOF(vp), tv.tv_sec - (kMaxSecsForFsync + 1));
9bccf70c 280 }
91447636 281 buf_setflags(bp, B_LOCKED);
b4c24cb9
A
282 }
283
9bccf70c
A
284 /*
285 * Delay-write this block.
286 * If the maximum delayed buffers has been exceeded then
287 * free up some buffers and fall back to an asynchronous write.
288 */
b4c24cb9
A
289 if (hfsmp->jnl) {
290 if (blockPtr->isModified == 0) {
2d21ac55 291 panic("hfs: releaseblock: modified is 0 but markdirty set! bp %p\n", bp);
b4c24cb9 292 }
55e303ae 293 retval = btree_journal_modify_block_end(hfsmp, bp);
b4c24cb9
A
294 blockPtr->isModified = 0;
295 } else if (bdwrite_internal(bp, 1) != 0) {
9bccf70c
A
296 hfs_btsync(vp, 0);
297 /* Rollback sync time to cause a sync on lock release... */
91447636
A
298 (void) BTSetLastSync(VTOF(vp), tv.tv_sec - (kMaxSecsForFsync + 1));
299
300 buf_clearflags(bp, B_LOCKED);
301 buf_bawrite(bp);
9bccf70c 302 }
d1ecb069
A
303
304 /* Don't let anyone else try to use this bp, it's been consumed */
305 blockPtr->blockHeader = NULL;
306
1c79356b 307 } else {
b4c24cb9 308 // check if we had previously called journal_modify_block_start()
91447636 309 // on this block and if so, abort it (which will call buf_brelse()).
b4c24cb9
A
310 if (hfsmp->jnl && blockPtr->isModified) {
311 // XXXdbg - I don't want to call modify_block_abort()
312 // because I think it may be screwing up the
313 // journal and blowing away a block that has
314 // valid data in it.
315 //
316 // journal_modify_block_abort(hfsmp->jnl, bp);
317 //panic("hfs: releaseblock called for 0x%x but mod_block_start previously called.\n", bp);
55e303ae 318 btree_journal_modify_block_end(hfsmp, bp);
b4c24cb9
A
319 blockPtr->isModified = 0;
320 } else {
91447636 321 buf_brelse(bp); /* note: B-tree code will clear blockPtr->blockHeader and blockPtr->buffer */
b4c24cb9 322 }
d1ecb069
A
323
324 /* Don't let anyone else try to use this bp, it's been consumed */
325 blockPtr->blockHeader = NULL;
326 }
327 }
1c79356b
A
328
329exit:
330 return (retval);
331}
332
333
9bccf70c 334__private_extern__
1c79356b
A
335OSStatus ExtendBTreeFile(FileReference vp, FSSize minEOF, FSSize maxEOF)
336{
337#pragma unused (maxEOF)
338
91447636 339 OSStatus retval = 0, ret = 0;
6601e61a
A
340 int64_t actualBytesAdded, origSize;
341 u_int64_t bytesToAdd;
9bccf70c
A
342 u_int32_t startAllocation;
343 u_int32_t fileblocks;
6601e61a 344 BTreeInfoRec btInfo;
1c79356b 345 ExtendedVCB *vcb;
6601e61a
A
346 FCB *filePtr;
347 struct proc *p = NULL;
348 int64_t trim = 0;
349 int lockflags = 0;
1c79356b
A
350
351 filePtr = GetFileControlBlock(vp);
352
6601e61a 353 if ( (off_t)minEOF > filePtr->fcbEOF )
1c79356b
A
354 {
355 bytesToAdd = minEOF - filePtr->fcbEOF;
356
9bccf70c
A
357 if (bytesToAdd < filePtr->ff_clumpsize)
358 bytesToAdd = filePtr->ff_clumpsize; //XXX why not always be a mutiple of clump size?
1c79356b
A
359 }
360 else
361 {
1c79356b
A
362 return -1;
363 }
364
9bccf70c 365 vcb = VTOVCB(vp);
6601e61a 366
1c79356b
A
367 /*
368 * The Extents B-tree can't have overflow extents. ExtendFileC will
369 * return an error if an attempt is made to extend the Extents B-tree
370 * when the resident extents are exhausted.
371 */
1c79356b 372
91447636
A
373 /* Protect allocation bitmap and extents overflow file. */
374 lockflags = SFL_BITMAP;
375 if (VTOC(vp)->c_fileid != kHFSExtentsFileID)
376 lockflags |= SFL_EXTENTS;
377 lockflags = hfs_systemfile_lock(vcb, lockflags, HFS_EXCLUSIVE_LOCK);
378
379 (void) BTGetInformation(filePtr, 0, &btInfo);
1c79356b 380
b4c24cb9 381#if 0 // XXXdbg
1c79356b 382 /*
9bccf70c 383 * The b-tree code expects nodes to be contiguous. So when
1c79356b 384 * the allocation block size is less than the b-tree node
9bccf70c
A
385 * size, we need to force disk allocations to be contiguous.
386 */
1c79356b
A
387 if (vcb->blockSize >= btInfo.nodeSize) {
388 extendFlags = 0;
389 } else {
390 /* Ensure that all b-tree nodes are contiguous on disk */
b4c24cb9 391 extendFlags = kEFContigMask;
1c79356b 392 }
b4c24cb9 393#endif
1c79356b 394
b4c24cb9 395 origSize = filePtr->fcbEOF;
9bccf70c
A
396 fileblocks = filePtr->ff_blocks;
397 startAllocation = vcb->nextAllocation;
1c79356b 398
b4c24cb9
A
399 // loop trying to get a contiguous chunk that's an integer multiple
400 // of the btree node size. if we can't get a contiguous chunk that
401 // is at least the node size then we break out of the loop and let
402 // the error propagate back up.
6601e61a
A
403 while((off_t)bytesToAdd >= btInfo.nodeSize) {
404 do {
55e303ae 405 retval = ExtendFileC(vcb, filePtr, bytesToAdd, 0,
6601e61a
A
406 kEFContigMask | kEFMetadataMask | kEFNoClumpMask,
407 (int64_t *)&actualBytesAdded);
b4c24cb9 408 if (retval == dskFulErr && actualBytesAdded == 0) {
6601e61a
A
409 bytesToAdd >>= 1;
410 if (bytesToAdd < btInfo.nodeSize) {
411 break;
412 } else if ((bytesToAdd % btInfo.nodeSize) != 0) {
413 // make sure it's an integer multiple of the nodeSize
414 bytesToAdd -= (bytesToAdd % btInfo.nodeSize);
415 }
b4c24cb9 416 }
6601e61a
A
417 } while (retval == dskFulErr && actualBytesAdded == 0);
418
419 if (retval == dskFulErr && actualBytesAdded == 0 && bytesToAdd <= btInfo.nodeSize) {
420 break;
421 }
422
423 filePtr->fcbEOF = (u_int64_t)filePtr->ff_blocks * (u_int64_t)vcb->blockSize;
424 bytesToAdd = minEOF - filePtr->fcbEOF;
425 }
91447636 426
9bccf70c
A
427 /*
428 * If a new extent was added then move the roving allocator
429 * reference forward by the current b-tree file size so
430 * there's plenty of room to grow.
431 */
432 if ((retval == 0) &&
55e303ae 433 ((VCBTOHFS(vcb)->hfs_flags & HFS_METADATA_ZONE) == 0) &&
9bccf70c 434 (vcb->nextAllocation > startAllocation) &&
2d21ac55
A
435 ((vcb->nextAllocation + fileblocks) < vcb->allocLimit)) {
436 HFS_UPDATE_NEXT_ALLOCATION(vcb, vcb->nextAllocation + fileblocks);
9bccf70c
A
437 }
438
b4c24cb9
A
439 filePtr->fcbEOF = (u_int64_t)filePtr->ff_blocks * (u_int64_t)vcb->blockSize;
440
441 // XXXdbg ExtendFileC() could have returned an error even though
442 // it grew the file to be big enough for our needs. If this is
443 // the case, we don't care about retval so we blow it away.
444 //
6601e61a 445 if (filePtr->fcbEOF >= (off_t)minEOF && retval != 0) {
b4c24cb9
A
446 retval = 0;
447 }
448
449 // XXXdbg if the file grew but isn't large enough or isn't an
450 // even multiple of the nodeSize then trim things back. if
451 // the file isn't large enough we trim back to the original
452 // size. otherwise we trim back to be an even multiple of the
453 // btree node size.
454 //
6601e61a 455 if ((filePtr->fcbEOF < (off_t)minEOF) || ((filePtr->fcbEOF - origSize) % btInfo.nodeSize) != 0) {
b4c24cb9 456
6601e61a 457 if (filePtr->fcbEOF < (off_t)minEOF) {
b4c24cb9
A
458 retval = dskFulErr;
459
460 if (filePtr->fcbEOF < origSize) {
461 panic("hfs: btree file eof %lld less than orig size %lld!\n",
462 filePtr->fcbEOF, origSize);
463 }
464
465 trim = filePtr->fcbEOF - origSize;
b4c24cb9 466 } else {
6601e61a 467 trim = ((filePtr->fcbEOF - origSize) % btInfo.nodeSize);
b4c24cb9
A
468 }
469
470 ret = TruncateFileC(vcb, filePtr, filePtr->fcbEOF - trim, 0);
471 filePtr->fcbEOF = (u_int64_t)filePtr->ff_blocks * (u_int64_t)vcb->blockSize;
472
473 // XXXdbg - panic if the file didn't get trimmed back properly
474 if ((filePtr->fcbEOF % btInfo.nodeSize) != 0) {
6601e61a 475 panic("hfs: truncate file didn't! fcbEOF %lld nsize %d fcb %p\n",
b4c24cb9
A
476 filePtr->fcbEOF, btInfo.nodeSize, filePtr);
477 }
478
479 if (ret) {
480 // XXXdbg - this probably doesn't need to be a panic()
6601e61a 481 panic("hfs: error truncating btree files (sz 0x%llx, trim %lld, ret %ld)\n",
b0d623f7 482 filePtr->fcbEOF, trim, (long)ret);
91447636 483 goto out;
b4c24cb9 484 }
b4c24cb9
A
485 }
486
9bccf70c
A
487 if(VTOC(vp)->c_fileid != kHFSExtentsFileID) {
488 /*
489 * Get any extents overflow b-tree changes to disk ASAP!
490 */
b4c24cb9 491 (void) BTFlushPath(VTOF(vcb->extentsRefNum));
91447636 492 (void) hfs_fsync(vcb->extentsRefNum, MNT_WAIT, 0, p);
9bccf70c 493 }
91447636
A
494 hfs_systemfile_unlock(vcb, lockflags);
495 lockflags = 0;
1c79356b 496
b4c24cb9 497 if ((filePtr->fcbEOF % btInfo.nodeSize) != 0) {
6601e61a 498 panic("hfs: extendbtree: fcb %p has eof 0x%llx not a multiple of 0x%x (trim %llx)\n",
b4c24cb9
A
499 filePtr, filePtr->fcbEOF, btInfo.nodeSize, trim);
500 }
501
1c79356b
A
502 /*
503 * Update the Alternate MDB or Alternate VolumeHeader
504 */
9bccf70c
A
505 if ((VTOC(vp)->c_fileid == kHFSExtentsFileID) ||
506 (VTOC(vp)->c_fileid == kHFSCatalogFileID) ||
507 (VTOC(vp)->c_fileid == kHFSAttributesFileID)
1c79356b 508 ) {
91447636 509 VTOC(vp)->c_flag |= C_MODIFIED;
1c79356b 510 MarkVCBDirty( vcb );
b4c24cb9 511 ret = hfs_flushvolumeheader(VCBTOHFS(vcb), MNT_WAIT, HFS_ALTFLUSH);
55e303ae 512 } else {
91447636
A
513 VTOC(vp)->c_touch_chgtime = TRUE;
514 VTOC(vp)->c_touch_modtime = TRUE;
515 (void) hfs_update(vp, TRUE);
1c79356b 516 }
b4c24cb9 517
6601e61a 518 ret = ClearBTNodes(vp, btInfo.nodeSize, origSize, (filePtr->fcbEOF - origSize));
91447636
A
519out:
520 if (retval == 0)
521 retval = ret;
522
523 if (lockflags)
524 hfs_systemfile_unlock(vcb, lockflags);
1c79356b
A
525
526 return retval;
527}
528
529
1c79356b
A
530/*
531 * Clear out (zero) new b-tree nodes on disk.
532 */
533static int
534ClearBTNodes(struct vnode *vp, long blksize, off_t offset, off_t amount)
535{
b4c24cb9 536 struct hfsmount *hfsmp = VTOHFS(vp);
1c79356b 537 struct buf *bp = NULL;
91447636
A
538 daddr64_t blk;
539 daddr64_t blkcnt;
1c79356b
A
540
541 blk = offset / blksize;
542 blkcnt = amount / blksize;
543
544 while (blkcnt > 0) {
91447636 545 bp = buf_getblk(vp, blk, blksize, 0, 0, BLK_META);
1c79356b
A
546 if (bp == NULL)
547 continue;
b4c24cb9
A
548
549 // XXXdbg
550 if (hfsmp->jnl) {
551 // XXXdbg -- skipping this for now since it makes a transaction
552 // become *way* too large
553 //journal_modify_block_start(hfsmp->jnl, bp);
554 }
91447636 555 bzero((char *)buf_dataptr(bp), blksize);
b4c24cb9 556
91447636 557 buf_markaged(bp);
1c79356b 558
b4c24cb9
A
559 // XXXdbg
560 if (hfsmp->jnl) {
561 // XXXdbg -- skipping this for now since it makes a transaction
562 // become *way* too large
563 //journal_modify_block_end(hfsmp->jnl, bp);
564
565 // XXXdbg - remove this once we decide what to do with the
566 // writes to the journal
567 if ((blk % 32) == 0)
91447636 568 VNOP_BWRITE(bp);
b4c24cb9 569 else
91447636 570 buf_bawrite(bp);
b4c24cb9
A
571 } else {
572 /* wait/yield every 32 blocks so we don't hog all the buffers */
573 if ((blk % 32) == 0)
91447636 574 VNOP_BWRITE(bp);
b4c24cb9 575 else
91447636 576 buf_bawrite(bp);
b4c24cb9 577 }
1c79356b
A
578 --blkcnt;
579 ++blk;
580 }
581
582 return (0);
583}
91447636
A
584
585
586extern char hfs_attrname[];
587
91447636
A
588/*
589 * Create an HFS+ Attribute B-tree File.
590 *
2d21ac55 591 * No global resources should be held.
91447636
A
592 */
593int
2d21ac55 594hfs_create_attr_btree(struct hfsmount *hfsmp, u_int32_t nodesize, u_int32_t nodecnt)
91447636 595{
2d21ac55 596 struct vnode* vp = NULLVP;
91447636
A
597 struct cat_desc cndesc;
598 struct cat_attr cnattr;
599 struct cat_fork cfork;
600 BlockDescriptor blkdesc;
601 BTNodeDescriptor *ndp;
602 BTHeaderRec *bthp;
603 BTreeControlBlockPtr btcb = NULL;
604 struct buf *bp = NULL;
605 void * buffer;
b0d623f7 606 u_int8_t *bitmap;
91447636 607 u_int16_t *index;
b0d623f7
A
608 u_int32_t node_num, num_map_nodes;
609 u_int32_t bytes_per_map_record;
610 u_int32_t temp;
91447636 611 u_int16_t offset;
2d21ac55 612 int intrans = 0;
91447636 613 int result;
2d21ac55
A
614again:
615 /*
616 * Serialize creation using HFS_CREATING_BTREE flag.
617 */
618 lck_mtx_lock(&hfsmp->hfs_mutex);
619 if (hfsmp->hfs_flags & HFS_CREATING_BTREE) {
620 /* Someone else beat us, wait for them to finish. */
621 (void) msleep(hfsmp->hfs_attribute_cp, &hfsmp->hfs_mutex,
622 PDROP | PINOD, "hfs_create_attr_btree", 0);
623 if (hfsmp->hfs_attribute_vp) {
624 return (0);
625 }
626 goto again;
627 }
628 hfsmp->hfs_flags |= HFS_CREATING_BTREE;
629 lck_mtx_unlock(&hfsmp->hfs_mutex);
91447636 630
2d21ac55
A
631 /* Check if were out of usable disk space. */
632 if ((hfs_freeblks(hfsmp, 1) == 0)) {
633 result = ENOSPC;
634 goto exit;
635 }
91447636
A
636
637 /*
638 * Set up Attribute B-tree vnode
2d21ac55
A
639 * (this must be done before we start a transaction
640 * or take any system file locks)
91447636
A
641 */
642 bzero(&cndesc, sizeof(cndesc));
643 cndesc.cd_parentcnid = kHFSRootParentID;
644 cndesc.cd_flags |= CD_ISMETA;
2d21ac55 645 cndesc.cd_nameptr = (const u_int8_t *)hfs_attrname;
91447636
A
646 cndesc.cd_namelen = strlen(hfs_attrname);
647 cndesc.cd_cnid = kHFSAttributesFileID;
648
649 bzero(&cnattr, sizeof(cnattr));
2d21ac55 650 cnattr.ca_linkcount = 1;
91447636
A
651 cnattr.ca_mode = S_IFREG;
652 cnattr.ca_fileid = cndesc.cd_cnid;
653
654 bzero(&cfork, sizeof(cfork));
655 cfork.cf_clump = nodesize * nodecnt;
656
657 result = hfs_getnewvnode(hfsmp, NULL, NULL, &cndesc, 0, &cnattr, &cfork, &vp);
2d21ac55
A
658 if (result) {
659 goto exit;
660 }
91447636
A
661 /*
662 * Set up Attribute B-tree control block
663 */
664 MALLOC(btcb, BTreeControlBlock *, sizeof(BTreeControlBlock), M_TEMP, M_WAITOK);
665 bzero(btcb, sizeof(BTreeControlBlock));
666
667 btcb->nodeSize = nodesize;
668 btcb->maxKeyLength = kHFSPlusAttrKeyMaximumLength;
669 btcb->btreeType = 0xFF;
670 btcb->attributes = kBTVariableIndexKeysMask | kBTBigKeysMask;
671 btcb->version = kBTreeVersion;
672 btcb->writeCount = 1;
673 btcb->flags = 0; /* kBTHeaderDirty */
674 btcb->fileRefNum = vp;
675 btcb->getBlockProc = GetBTreeBlock;
676 btcb->releaseBlockProc = ReleaseBTreeBlock;
677 btcb->setEndOfForkProc = ExtendBTreeFile;
678 btcb->keyCompareProc = (KeyCompareProcPtr)hfs_attrkeycompare;
679 VTOF(vp)->fcbBTCBPtr = btcb;
680
681 /*
682 * Allocate some space
683 */
2d21ac55
A
684 if (hfs_start_transaction(hfsmp) != 0) {
685 result = EINVAL;
686 goto exit;
687 }
688 intrans = 1;
689
690 /* Note ExtendBTreeFile will acquire the necessary system file locks. */
91447636
A
691 result = ExtendBTreeFile(vp, nodesize, cfork.cf_clump);
692 if (result)
693 goto exit;
694
695 btcb->totalNodes = VTOF(vp)->ff_size / nodesize;
91447636 696
b0d623f7
A
697 /*
698 * Figure out how many map nodes we'll need.
699 *
700 * bytes_per_map_record = the number of bytes in the map record of a
701 * map node. Since that is the only record in the node, it is the size
702 * of the node minus the node descriptor at the start, and two record
703 * offsets at the end of the node. The "- 2" is to round the size down
704 * to a multiple of 4 bytes (since sizeof(BTNodeDescriptor) is not a
705 * multiple of 4).
706 *
707 * The value "temp" here is the number of *bits* in the map record of
708 * the header node.
709 */
710 bytes_per_map_record = nodesize - sizeof(BTNodeDescriptor) - 2*sizeof(u_int16_t) - 2;
711 temp = 8 * (nodesize - sizeof(BTNodeDescriptor)
712 - sizeof(BTHeaderRec)
713 - kBTreeHeaderUserBytes
714 - 4 * sizeof(u_int16_t));
715 if (btcb->totalNodes > temp) {
716 num_map_nodes = howmany(btcb->totalNodes - temp, bytes_per_map_record * 8);
717 }
718 else {
719 num_map_nodes = 0;
720 }
721
722 btcb->freeNodes = btcb->totalNodes - 1 - num_map_nodes;
723
91447636
A
724 /*
725 * Initialize the b-tree header on disk
726 */
727 bp = buf_getblk(vp, 0, nodesize, 0, 0, BLK_META);
728 if (bp == NULL) {
729 result = EIO;
730 goto exit;
731 }
732
733 buffer = (void *)buf_dataptr(bp);
734 blkdesc.buffer = buffer;
735 blkdesc.blockHeader = (void *)bp;
736 blkdesc.blockReadFromDisk = 0;
737 blkdesc.isModified = 0;
738
739 ModifyBlockStart(vp, &blkdesc);
740
741 if (buf_size(bp) != nodesize)
742 panic("hfs_create_attr_btree: bad buffer size (%d)\n", buf_size(bp));
743
744 bzero(buffer, nodesize);
2d21ac55 745 index = (u_int16_t *)buffer;
91447636
A
746
747 /* FILL IN THE NODE DESCRIPTOR: */
748 ndp = (BTNodeDescriptor *)buffer;
b0d623f7
A
749 if (num_map_nodes != 0)
750 ndp->fLink = 1;
91447636
A
751 ndp->kind = kBTHeaderNode;
752 ndp->numRecords = 3;
753 offset = sizeof(BTNodeDescriptor);
754 index[(nodesize / 2) - 1] = offset;
755
756 /* FILL IN THE HEADER RECORD: */
2d21ac55 757 bthp = (BTHeaderRec *)((u_int8_t *)buffer + offset);
91447636
A
758 bthp->nodeSize = nodesize;
759 bthp->totalNodes = btcb->totalNodes;
760 bthp->freeNodes = btcb->freeNodes;
761 bthp->clumpSize = cfork.cf_clump;
762 bthp->btreeType = 0xFF;
763 bthp->attributes = kBTVariableIndexKeysMask | kBTBigKeysMask;
764 bthp->maxKeyLength = kHFSPlusAttrKeyMaximumLength;
765 bthp->keyCompareType = kHFSBinaryCompare;
766 offset += sizeof(BTHeaderRec);
767 index[(nodesize / 2) - 2] = offset;
768
769 /* FILL IN THE USER RECORD: */
770 offset += kBTreeHeaderUserBytes;
771 index[(nodesize / 2) - 3] = offset;
772
b0d623f7
A
773 /* Mark the header node and map nodes in use in the map record.
774 *
775 * NOTE: Assumes that the header node's map record has at least
776 * (num_map_nodes + 1) bits.
777 */
778 bitmap = (u_int8_t *) buffer + offset;
779 temp = num_map_nodes + 1; /* +1 for the header node */
780 while (temp >= 8) {
781 *(bitmap++) = 0xFF;
782 temp -= 8;
783 }
784 *bitmap = ~(0xFF >> temp);
785
91447636
A
786 offset += nodesize - sizeof(BTNodeDescriptor) - sizeof(BTHeaderRec)
787 - kBTreeHeaderUserBytes - (4 * sizeof(int16_t));
788 index[(nodesize / 2) - 4] = offset;
789
790 if (hfsmp->jnl) {
791 result = btree_journal_modify_block_end(hfsmp, bp);
792 } else {
793 result = VNOP_BWRITE(bp);
794 }
795 if (result)
796 goto exit;
797
b0d623f7
A
798 /* Create the map nodes: node numbers 1 .. num_map_nodes */
799 for (node_num=1; node_num <= num_map_nodes; ++node_num) {
800 bp = buf_getblk(vp, node_num, nodesize, 0, 0, BLK_META);
801 if (bp == NULL) {
802 result = EIO;
803 goto exit;
804 }
805 buffer = (void *)buf_dataptr(bp);
806 blkdesc.buffer = buffer;
807 blkdesc.blockHeader = (void *)bp;
808 blkdesc.blockReadFromDisk = 0;
809 blkdesc.isModified = 0;
810
811 ModifyBlockStart(vp, &blkdesc);
812
813 bzero(buffer, nodesize);
814 index = (u_int16_t *)buffer;
815
816 /* Fill in the node descriptor */
817 ndp = (BTNodeDescriptor *)buffer;
818 if (node_num != num_map_nodes)
819 ndp->fLink = node_num + 1;
820 ndp->kind = kBTMapNode;
821 ndp->numRecords = 1;
822 offset = sizeof(BTNodeDescriptor);
823 index[(nodesize / 2) - 1] = offset;
824
825
826 /* Fill in the map record's offset */
827 /* Note: We assume that the map record is all zeroes */
828 offset = sizeof(BTNodeDescriptor) + bytes_per_map_record;
829 index[(nodesize / 2) - 2] = offset;
830
831 if (hfsmp->jnl) {
832 result = btree_journal_modify_block_end(hfsmp, bp);
833 } else {
834 result = VNOP_BWRITE(bp);
835 }
836 if (result)
837 goto exit;
838 }
839
2d21ac55
A
840 /* Update vp/cp for attribute btree */
841 lck_mtx_lock(&hfsmp->hfs_mutex);
842 hfsmp->hfs_attribute_cp = VTOC(vp);
91447636 843 hfsmp->hfs_attribute_vp = vp;
2d21ac55 844 lck_mtx_unlock(&hfsmp->hfs_mutex);
91447636 845
2d21ac55 846 (void) hfs_flushvolumeheader(hfsmp, MNT_WAIT, HFS_ALTFLUSH);
91447636 847exit:
2d21ac55
A
848 if (vp) {
849 hfs_unlock(VTOC(vp));
850 }
91447636
A
851 if (result) {
852 if (btcb) {
853 FREE (btcb, M_TEMP);
854 }
2d21ac55
A
855 if (vp) {
856 vnode_put(vp);
857 }
858 /* XXX need to give back blocks ? */
859 }
860 if (intrans) {
861 hfs_end_transaction(hfsmp);
91447636 862 }
91447636 863
2d21ac55
A
864 /*
865 * All done, clear HFS_CREATING_BTREE, and wake up any sleepers.
866 */
867 lck_mtx_lock(&hfsmp->hfs_mutex);
868 hfsmp->hfs_flags &= ~HFS_CREATING_BTREE;
869 wakeup((caddr_t)hfsmp->hfs_attribute_cp);
870 lck_mtx_unlock(&hfsmp->hfs_mutex);
91447636 871
2d21ac55
A
872 return (result);
873}
91447636 874