lint, minor cleanup
[unix-history] / usr / src / sys / ufs / lfs / lfs_inode.c
... / ...
CommitLineData
1/*
2 * Copyright (c) 1982, 1986, 1989 Regents of the University of California.
3 * All rights reserved.
4 *
5 * %sccs.include.redist.c%
6 *
7 * @(#)lfs_inode.c 7.34 (Berkeley) %G%
8 */
9
10#include "param.h"
11#include "systm.h"
12#include "mount.h"
13#include "user.h"
14#include "proc.h"
15#include "file.h"
16#include "buf.h"
17#include "cmap.h"
18#include "vnode.h"
19#include "../ufs/quota.h"
20#include "../ufs/inode.h"
21#include "../ufs/fs.h"
22#include "../ufs/ufsmount.h"
23#include "kernel.h"
24#include "malloc.h"
25
26#define INOHSZ 512
27#if ((INOHSZ&(INOHSZ-1)) == 0)
28#define INOHASH(dev,ino) (((dev)+(ino))&(INOHSZ-1))
29#else
30#define INOHASH(dev,ino) (((unsigned)((dev)+(ino)))%INOHSZ)
31#endif
32
33union ihead {
34 union ihead *ih_head[2];
35 struct inode *ih_chain[2];
36} ihead[INOHSZ];
37
38int prtactive; /* 1 => print out reclaim of active vnodes */
39
40/*
41 * Initialize hash links for inodes.
42 */
43ufs_init()
44{
45 register int i;
46 register union ihead *ih = ihead;
47
48#ifndef lint
49 if (VN_MAXPRIVATE < sizeof(struct inode))
50 panic("ihinit: too small");
51#endif /* not lint */
52 for (i = INOHSZ; --i >= 0; ih++) {
53 ih->ih_head[0] = ih;
54 ih->ih_head[1] = ih;
55 }
56#ifdef QUOTA
57 dqinit();
58#endif /* QUOTA */
59}
60
61/*
62 * Look up an vnode/inode by device,inumber.
63 * If it is in core (in the inode structure),
64 * honor the locking protocol.
65 * If it is not in core, read it in from the
66 * specified device.
67 * Callers must check for mount points!!
68 * In all cases, a pointer to a locked
69 * inode structure is returned.
70 */
71iget(xp, ino, ipp)
72 struct inode *xp;
73 ino_t ino;
74 struct inode **ipp;
75{
76 dev_t dev = xp->i_dev;
77 struct mount *mntp = ITOV(xp)->v_mount;
78 register struct fs *fs = VFSTOUFS(mntp)->um_fs;
79 extern struct vnodeops ufs_vnodeops, spec_inodeops;
80 register struct inode *ip, *iq;
81 register struct vnode *vp;
82 struct vnode *nvp;
83 struct buf *bp;
84 struct dinode *dp;
85 union ihead *ih;
86 int i, error;
87
88 ih = &ihead[INOHASH(dev, ino)];
89loop:
90 for (ip = ih->ih_chain[0]; ip != (struct inode *)ih; ip = ip->i_forw) {
91 if (ino != ip->i_number || dev != ip->i_dev)
92 continue;
93 if ((ip->i_flag&ILOCKED) != 0) {
94 ip->i_flag |= IWANT;
95 sleep((caddr_t)ip, PINOD);
96 goto loop;
97 }
98 if (vget(ITOV(ip)))
99 goto loop;
100 *ipp = ip;
101 return(0);
102 }
103 /*
104 * Allocate a new inode.
105 */
106 if (error = getnewvnode(VT_UFS, mntp, &ufs_vnodeops, &nvp)) {
107 *ipp = 0;
108 return (error);
109 }
110 ip = VTOI(nvp);
111 ip->i_vnode = nvp;
112 ip->i_flag = 0;
113 ip->i_devvp = 0;
114 ip->i_mode = 0;
115 ip->i_diroff = 0;
116#ifdef QUOTA
117 for (i = 0; i < MAXQUOTAS; i++)
118 ip->i_dquot[i] = NODQUOT;
119#endif
120 /*
121 * Put it onto its hash chain and lock it so that other requests for
122 * this inode will block if they arrive while we are sleeping waiting
123 * for old data structures to be purged or for the contents of the
124 * disk portion of this inode to be read.
125 */
126 ip->i_dev = dev;
127 ip->i_number = ino;
128 insque(ip, ih);
129 ILOCK(ip);
130 /*
131 * Read in the disk contents for the inode.
132 */
133 if (error = bread(VFSTOUFS(mntp)->um_devvp, fsbtodb(fs, itod(fs, ino)),
134 (int)fs->fs_bsize, NOCRED, &bp)) {
135 /*
136 * The inode does not contain anything useful, so it would
137 * be misleading to leave it on its hash chain.
138 * Iput() will take care of putting it back on the free list.
139 */
140 remque(ip);
141 ip->i_forw = ip;
142 ip->i_back = ip;
143 /*
144 * Unlock and discard unneeded inode.
145 */
146 iput(ip);
147 brelse(bp);
148 *ipp = 0;
149 return (error);
150 }
151 dp = bp->b_un.b_dino;
152 dp += itoo(fs, ino);
153 ip->i_din = *dp;
154 brelse(bp);
155 /*
156 * Initialize the associated vnode
157 */
158 vp = ITOV(ip);
159 vp->v_type = IFTOVT(ip->i_mode);
160 if (vp->v_type == VFIFO) {
161#ifdef FIFO
162 extern struct vnodeops fifo_inodeops;
163 vp->v_op = &fifo_inodeops;
164#else
165 iput(ip);
166 *ipp = 0;
167 return (EOPNOTSUPP);
168#endif /* FIFO */
169 }
170 if (vp->v_type == VCHR || vp->v_type == VBLK) {
171 vp->v_op = &spec_inodeops;
172 if (nvp = checkalias(vp, ip->i_rdev, mntp)) {
173 /*
174 * Reinitialize aliased inode.
175 */
176 vp = nvp;
177 iq = VTOI(vp);
178 iq->i_vnode = vp;
179 iq->i_flag = 0;
180 ILOCK(iq);
181 iq->i_din = ip->i_din;
182 iq->i_dev = dev;
183 iq->i_number = ino;
184 insque(iq, ih);
185 /*
186 * Discard unneeded vnode
187 */
188 ip->i_mode = 0;
189 iput(ip);
190 ip = iq;
191 }
192 }
193 if (ino == ROOTINO)
194 vp->v_flag |= VROOT;
195 /*
196 * Finish inode initialization.
197 */
198 ip->i_fs = fs;
199 ip->i_devvp = VFSTOUFS(mntp)->um_devvp;
200 VREF(ip->i_devvp);
201 /*
202 * Set up a generation number for this inode if it does not
203 * already have one. This should only happen on old filesystems.
204 */
205 if (ip->i_gen == 0) {
206 if (++nextgennumber < (u_long)time.tv_sec)
207 nextgennumber = time.tv_sec;
208 ip->i_gen = nextgennumber;
209 if ((vp->v_mount->mnt_flag & MNT_RDONLY) == 0)
210 ip->i_flag |= IMOD;
211 }
212 *ipp = ip;
213 return (0);
214}
215
216/*
217 * Unlock and decrement the reference count of an inode structure.
218 */
219iput(ip)
220 register struct inode *ip;
221{
222
223 if ((ip->i_flag & ILOCKED) == 0)
224 panic("iput");
225 IUNLOCK(ip);
226 vrele(ITOV(ip));
227}
228
229/*
230 * Last reference to an inode, write the inode out and if necessary,
231 * truncate and deallocate the file.
232 */
233ufs_inactive(vp)
234 struct vnode *vp;
235{
236 register struct inode *ip = VTOI(vp);
237 int mode, error = 0;
238
239 if (prtactive && vp->v_usecount != 0)
240 vprint("ufs_inactive: pushing active", vp);
241 /*
242 * Get rid of inodes related to stale file handles.
243 */
244 if (ip->i_mode == 0) {
245 if ((vp->v_flag & VXLOCK) == 0)
246 vgone(vp);
247 return (0);
248 }
249 ILOCK(ip);
250 if (ip->i_nlink <= 0 && (vp->v_mount->mnt_flag & MNT_RDONLY) == 0) {
251#ifdef QUOTA
252 if (!getinoquota(ip))
253 (void) chkiq(ip, -1, NOCRED, 0);
254#endif
255 error = itrunc(ip, (u_long)0, 0);
256 mode = ip->i_mode;
257 ip->i_mode = 0;
258 ip->i_rdev = 0;
259 ip->i_flag |= IUPD|ICHG;
260 ifree(ip, ip->i_number, mode);
261 }
262 IUPDAT(ip, &time, &time, 0);
263 IUNLOCK(ip);
264 ip->i_flag = 0;
265 /*
266 * If we are done with the inode, reclaim it
267 * so that it can be reused immediately.
268 */
269 if (vp->v_usecount == 0 && ip->i_mode == 0)
270 vgone(vp);
271 return (error);
272}
273
274/*
275 * Reclaim an inode so that it can be used for other purposes.
276 */
277ufs_reclaim(vp)
278 register struct vnode *vp;
279{
280 register struct inode *ip = VTOI(vp);
281 int i;
282
283 if (prtactive && vp->v_usecount != 0)
284 vprint("ufs_reclaim: pushing active", vp);
285 /*
286 * Remove the inode from its hash chain.
287 */
288 remque(ip);
289 ip->i_forw = ip;
290 ip->i_back = ip;
291 /*
292 * Purge old data structures associated with the inode.
293 */
294 cache_purge(vp);
295 if (ip->i_devvp) {
296 vrele(ip->i_devvp);
297 ip->i_devvp = 0;
298 }
299#ifdef QUOTA
300 for (i = 0; i < MAXQUOTAS; i++) {
301 if (ip->i_dquot[i] != NODQUOT) {
302 dqrele(vp, ip->i_dquot[i]);
303 ip->i_dquot[i] = NODQUOT;
304 }
305 }
306#endif
307 ip->i_flag = 0;
308 return (0);
309}
310
311/*
312 * Check accessed and update flags on an inode structure.
313 * If any is on, update the inode with the current time.
314 * If waitfor is given, then must ensure I/O order,
315 * so wait for write to complete.
316 */
317iupdat(ip, ta, tm, waitfor)
318 register struct inode *ip;
319 struct timeval *ta, *tm;
320 int waitfor;
321{
322 struct buf *bp;
323 struct vnode *vp = ITOV(ip);
324 struct dinode *dp;
325 register struct fs *fs;
326
327 fs = ip->i_fs;
328 if ((ip->i_flag & (IUPD|IACC|ICHG|IMOD)) == 0)
329 return (0);
330 if (vp->v_mount->mnt_flag & MNT_RDONLY)
331 return (0);
332 error = bread(ip->i_devvp, fsbtodb(fs, itod(fs, ip->i_number)),
333 (int)fs->fs_bsize, NOCRED, &bp);
334 if (error) {
335 brelse(bp);
336 return (error);
337 }
338 if (ip->i_flag&IACC)
339 ip->i_atime = ta->tv_sec;
340 if (ip->i_flag&IUPD)
341 ip->i_mtime = tm->tv_sec;
342 if (ip->i_flag&ICHG)
343 ip->i_ctime = time.tv_sec;
344 ip->i_flag &= ~(IUPD|IACC|ICHG|IMOD);
345 dp = bp->b_un.b_dino + itoo(fs, ip->i_number);
346 *dp = ip->i_din;
347 if (waitfor) {
348 return (bwrite(bp));
349 } else {
350 bdwrite(bp);
351 return (0);
352 }
353}
354
355#define SINGLE 0 /* index of single indirect block */
356#define DOUBLE 1 /* index of double indirect block */
357#define TRIPLE 2 /* index of triple indirect block */
358/*
359 * Truncate the inode ip to at most length size. Free affected disk
360 * blocks -- the blocks of the file are removed in reverse order.
361 *
362 * NB: triple indirect blocks are untested.
363 */
364itrunc(oip, length, flags)
365 register struct inode *oip;
366 u_long length;
367 int flags;
368{
369 register daddr_t lastblock;
370 daddr_t bn, lbn, lastiblock[NIADDR];
371 register struct fs *fs;
372 register struct inode *ip;
373 struct buf *bp;
374 int offset, osize, size, level;
375 long count, nblocks, blocksreleased = 0;
376 register int i;
377 int aflags, error, allerror;
378 struct inode tip;
379
380 if (oip->i_size <= length) {
381 oip->i_flag |= ICHG|IUPD;
382 error = iupdat(oip, &time, &time, 1);
383 return (error);
384 }
385 /*
386 * Calculate index into inode's block list of
387 * last direct and indirect blocks (if any)
388 * which we want to keep. Lastblock is -1 when
389 * the file is truncated to 0.
390 */
391 fs = oip->i_fs;
392 lastblock = lblkno(fs, length + fs->fs_bsize - 1) - 1;
393 lastiblock[SINGLE] = lastblock - NDADDR;
394 lastiblock[DOUBLE] = lastiblock[SINGLE] - NINDIR(fs);
395 lastiblock[TRIPLE] = lastiblock[DOUBLE] - NINDIR(fs) * NINDIR(fs);
396 nblocks = btodb(fs->fs_bsize);
397 /*
398 * Update the size of the file. If the file is not being
399 * truncated to a block boundry, the contents of the
400 * partial block following the end of the file must be
401 * zero'ed in case it ever become accessable again because
402 * of subsequent file growth.
403 */
404 osize = oip->i_size;
405 offset = blkoff(fs, length);
406 if (offset == 0) {
407 oip->i_size = length;
408 } else {
409 lbn = lblkno(fs, length);
410 aflags = B_CLRBUF;
411 if (flags & IO_SYNC)
412 aflags |= B_SYNC;
413#ifdef QUOTA
414 if (error = getinoquota(oip))
415 return (error);
416#endif
417 if (error = balloc(oip, lbn, offset, &bp, aflags))
418 return (error);
419 oip->i_size = length;
420 size = blksize(fs, oip, lbn);
421 bn = bp->b_blkno;
422 count = howmany(size, CLBYTES);
423 munhash(oip->i_devvp, bn + i * CLBYTES / DEV_BSIZE);
424 bzero(bp->b_un.b_addr + offset, (unsigned)(size - offset));
425 brealloc(bp, size);
426 if (flags & IO_SYNC)
427 bwrite(bp);
428 else
429 bdwrite(bp);
430 }
431 /*
432 * Update file and block pointers
433 * on disk before we start freeing blocks.
434 * If we crash before free'ing blocks below,
435 * the blocks will be returned to the free list.
436 * lastiblock values are also normalized to -1
437 * for calls to indirtrunc below.
438 */
439 tip = *oip;
440 tip.i_size = osize;
441 for (level = TRIPLE; level >= SINGLE; level--)
442 if (lastiblock[level] < 0) {
443 oip->i_ib[level] = 0;
444 lastiblock[level] = -1;
445 }
446 for (i = NDADDR - 1; i > lastblock; i--)
447 oip->i_db[i] = 0;
448 oip->i_flag |= ICHG|IUPD;
449 vinvalbuf(ITOV(oip), (length > 0));
450 allerror = iupdat(oip, &time, &time, MNT_WAIT);
451
452 /*
453 * Indirect blocks first.
454 */
455 ip = &tip;
456 for (level = TRIPLE; level >= SINGLE; level--) {
457 bn = ip->i_ib[level];
458 if (bn != 0) {
459 error = indirtrunc(ip, bn, lastiblock[level], level,
460 &count);
461 if (error)
462 allerror = error;
463 blocksreleased += count;
464 if (lastiblock[level] < 0) {
465 ip->i_ib[level] = 0;
466 blkfree(ip, bn, (off_t)fs->fs_bsize);
467 blocksreleased += nblocks;
468 }
469 }
470 if (lastiblock[level] >= 0)
471 goto done;
472 }
473
474 /*
475 * All whole direct blocks or frags.
476 */
477 for (i = NDADDR - 1; i > lastblock; i--) {
478 register off_t bsize;
479
480 bn = ip->i_db[i];
481 if (bn == 0)
482 continue;
483 ip->i_db[i] = 0;
484 bsize = (off_t)blksize(fs, ip, i);
485 blkfree(ip, bn, bsize);
486 blocksreleased += btodb(bsize);
487 }
488 if (lastblock < 0)
489 goto done;
490
491 /*
492 * Finally, look for a change in size of the
493 * last direct block; release any frags.
494 */
495 bn = ip->i_db[lastblock];
496 if (bn != 0) {
497 off_t oldspace, newspace;
498
499 /*
500 * Calculate amount of space we're giving
501 * back as old block size minus new block size.
502 */
503 oldspace = blksize(fs, ip, lastblock);
504 ip->i_size = length;
505 newspace = blksize(fs, ip, lastblock);
506 if (newspace == 0)
507 panic("itrunc: newspace");
508 if (oldspace - newspace > 0) {
509 /*
510 * Block number of space to be free'd is
511 * the old block # plus the number of frags
512 * required for the storage we're keeping.
513 */
514 bn += numfrags(fs, newspace);
515 blkfree(ip, bn, oldspace - newspace);
516 blocksreleased += btodb(oldspace - newspace);
517 }
518 }
519done:
520/* BEGIN PARANOIA */
521 for (level = SINGLE; level <= TRIPLE; level++)
522 if (ip->i_ib[level] != oip->i_ib[level])
523 panic("itrunc1");
524 for (i = 0; i < NDADDR; i++)
525 if (ip->i_db[i] != oip->i_db[i])
526 panic("itrunc2");
527/* END PARANOIA */
528 oip->i_blocks -= blocksreleased;
529 if (oip->i_blocks < 0) /* sanity */
530 oip->i_blocks = 0;
531 oip->i_flag |= ICHG;
532#ifdef QUOTA
533 if (!getinoquota(oip))
534 (void) chkdq(oip, -blocksreleased, NOCRED, 0);
535#endif
536 return (allerror);
537}
538
539/*
540 * Release blocks associated with the inode ip and
541 * stored in the indirect block bn. Blocks are free'd
542 * in LIFO order up to (but not including) lastbn. If
543 * level is greater than SINGLE, the block is an indirect
544 * block and recursive calls to indirtrunc must be used to
545 * cleanse other indirect blocks.
546 *
547 * NB: triple indirect blocks are untested.
548 */
549indirtrunc(ip, bn, lastbn, level, countp)
550 register struct inode *ip;
551 daddr_t bn, lastbn;
552 int level;
553 long *countp;
554{
555 register int i;
556 struct buf *bp;
557 register struct fs *fs = ip->i_fs;
558 register daddr_t *bap;
559 daddr_t *copy, nb, last;
560 long blkcount, factor;
561 int nblocks, blocksreleased = 0;
562 int error, allerror = 0;
563
564 /*
565 * Calculate index in current block of last
566 * block to be kept. -1 indicates the entire
567 * block so we need not calculate the index.
568 */
569 factor = 1;
570 for (i = SINGLE; i < level; i++)
571 factor *= NINDIR(fs);
572 last = lastbn;
573 if (lastbn > 0)
574 last /= factor;
575 nblocks = btodb(fs->fs_bsize);
576 /*
577 * Get buffer of block pointers, zero those
578 * entries corresponding to blocks to be free'd,
579 * and update on disk copy first.
580 */
581#ifdef SECSIZE
582 bp = bread(ip->i_dev, fsbtodb(fs, bn), (int)fs->fs_bsize,
583 fs->fs_dbsize);
584#else SECSIZE
585 error = bread(ip->i_devvp, fsbtodb(fs, bn), (int)fs->fs_bsize,
586 NOCRED, &bp);
587 if (error) {
588 brelse(bp);
589 *countp = 0;
590 return (error);
591 }
592 bap = bp->b_un.b_daddr;
593 MALLOC(copy, daddr_t *, fs->fs_bsize, M_TEMP, M_WAITOK);
594 bcopy((caddr_t)bap, (caddr_t)copy, (u_int)fs->fs_bsize);
595 bzero((caddr_t)&bap[last + 1],
596 (u_int)(NINDIR(fs) - (last + 1)) * sizeof (daddr_t));
597 if (last == -1)
598 bp->b_flags |= B_INVAL;
599 error = bwrite(bp);
600 if (error)
601 allerror = error;
602 bap = copy;
603
604 /*
605 * Recursively free totally unused blocks.
606 */
607 for (i = NINDIR(fs) - 1; i > last; i--) {
608 nb = bap[i];
609 if (nb == 0)
610 continue;
611 if (level > SINGLE) {
612 error = indirtrunc(ip, nb, (daddr_t)-1, level - 1,
613 &blkcount);
614 if (error)
615 allerror = error;
616 blocksreleased += blkcount;
617 }
618 blkfree(ip, nb, (off_t)fs->fs_bsize);
619 blocksreleased += nblocks;
620 }
621
622 /*
623 * Recursively free last partial block.
624 */
625 if (level > SINGLE && lastbn >= 0) {
626 last = lastbn % factor;
627 nb = bap[i];
628 if (nb != 0) {
629 error = indirtrunc(ip, nb, last, level - 1, &blkcount);
630 if (error)
631 allerror = error;
632 blocksreleased += blkcount;
633 }
634 }
635 FREE(copy, M_TEMP);
636 *countp = blocksreleased;
637 return (allerror);
638}
639
640/*
641 * Lock an inode. If its already locked, set the WANT bit and sleep.
642 */
643ilock(ip)
644 register struct inode *ip;
645{
646
647 while (ip->i_flag & ILOCKED) {
648 ip->i_flag |= IWANT;
649 if (ip->i_spare0 == u.u_procp->p_pid)
650 panic("locking against myself");
651 ip->i_spare1 = u.u_procp->p_pid;
652 (void) sleep((caddr_t)ip, PINOD);
653 }
654 ip->i_spare1 = 0;
655 ip->i_spare0 = u.u_procp->p_pid;
656 u.u_spare[0]++;
657 ip->i_flag |= ILOCKED;
658}
659
660/*
661 * Unlock an inode. If WANT bit is on, wakeup.
662 */
663iunlock(ip)
664 register struct inode *ip;
665{
666
667 if ((ip->i_flag & ILOCKED) == 0)
668 vprint("iunlock: unlocked inode", ITOV(ip));
669 ip->i_spare0 = 0;
670 u.u_spare[0]--;
671 ip->i_flag &= ~ILOCKED;
672 if (ip->i_flag&IWANT) {
673 ip->i_flag &= ~IWANT;
674 wakeup((caddr_t)ip);
675 }
676}