Bug Summary

File:kern/vfs_cache.c
Warning:line 376, column 7
Although the value stored to 'lncp' is used in the enclosing expression, the value is never actually read from 'lncp'

Annotated Source Code

Press '?' to see keyboard shortcuts

clang -cc1 -cc1 -triple amd64-unknown-openbsd7.4 -analyze -disable-free -clear-ast-before-backend -disable-llvm-verifier -discard-value-names -main-file-name vfs_cache.c -analyzer-checker=core -analyzer-checker=apiModeling -analyzer-checker=unix -analyzer-checker=deadcode -analyzer-checker=security.insecureAPI.UncheckedReturn -analyzer-checker=security.insecureAPI.getpw -analyzer-checker=security.insecureAPI.gets -analyzer-checker=security.insecureAPI.mktemp -analyzer-checker=security.insecureAPI.mkstemp -analyzer-checker=security.insecureAPI.vfork -analyzer-checker=nullability.NullPassedToNonnull -analyzer-checker=nullability.NullReturnedFromNonnull -analyzer-output plist -w -setup-static-analyzer -mrelocation-model static -mframe-pointer=all -relaxed-aliasing -ffp-contract=on -fno-rounding-math -mconstructor-aliases -ffreestanding -mcmodel=kernel -target-cpu x86-64 -target-feature +retpoline-indirect-calls -target-feature +retpoline-indirect-branches -target-feature -sse2 -target-feature -sse -target-feature -3dnow -target-feature -mmx -target-feature +save-args -target-feature +retpoline-external-thunk -disable-red-zone -no-implicit-float -tune-cpu generic -debugger-tuning=gdb -fcoverage-compilation-dir=/usr/src/sys/arch/amd64/compile/GENERIC.MP/obj -nostdsysteminc -nobuiltininc -resource-dir /usr/local/llvm16/lib/clang/16 -I /usr/src/sys -I /usr/src/sys/arch/amd64/compile/GENERIC.MP/obj -I /usr/src/sys/arch -I /usr/src/sys/dev/pci/drm/include -I /usr/src/sys/dev/pci/drm/include/uapi -I /usr/src/sys/dev/pci/drm/amd/include/asic_reg -I /usr/src/sys/dev/pci/drm/amd/include -I /usr/src/sys/dev/pci/drm/amd/amdgpu -I /usr/src/sys/dev/pci/drm/amd/display -I /usr/src/sys/dev/pci/drm/amd/display/include -I /usr/src/sys/dev/pci/drm/amd/display/dc -I /usr/src/sys/dev/pci/drm/amd/display/amdgpu_dm -I /usr/src/sys/dev/pci/drm/amd/pm/inc -I /usr/src/sys/dev/pci/drm/amd/pm/legacy-dpm -I /usr/src/sys/dev/pci/drm/amd/pm/swsmu -I /usr/src/sys/dev/pci/drm/amd/pm/swsmu/inc -I /usr/src/sys/dev/pci/drm/amd/pm/swsmu/smu11 -I /usr/src/sys/dev/pci/drm/amd/pm/swsmu/smu12 -I /usr/src/sys/dev/pci/drm/amd/pm/swsmu/smu13 -I /usr/src/sys/dev/pci/drm/amd/pm/powerplay/inc -I /usr/src/sys/dev/pci/drm/amd/pm/powerplay/hwmgr -I /usr/src/sys/dev/pci/drm/amd/pm/powerplay/smumgr -I /usr/src/sys/dev/pci/drm/amd/pm/swsmu/inc -I /usr/src/sys/dev/pci/drm/amd/pm/swsmu/inc/pmfw_if -I /usr/src/sys/dev/pci/drm/amd/display/dc/inc -I /usr/src/sys/dev/pci/drm/amd/display/dc/inc/hw -I /usr/src/sys/dev/pci/drm/amd/display/dc/clk_mgr -I /usr/src/sys/dev/pci/drm/amd/display/modules/inc -I /usr/src/sys/dev/pci/drm/amd/display/modules/hdcp -I /usr/src/sys/dev/pci/drm/amd/display/dmub/inc -I /usr/src/sys/dev/pci/drm/i915 -D DDB -D DIAGNOSTIC -D KTRACE -D ACCOUNTING -D KMEMSTATS -D PTRACE -D POOL_DEBUG -D CRYPTO -D SYSVMSG -D SYSVSEM -D SYSVSHM -D UVM_SWAP_ENCRYPT -D FFS -D FFS2 -D FFS_SOFTUPDATES -D UFS_DIRHASH -D QUOTA -D EXT2FS -D MFS -D NFSCLIENT -D NFSSERVER -D CD9660 -D UDF -D MSDOSFS -D FIFO -D FUSE -D SOCKET_SPLICE -D TCP_ECN -D TCP_SIGNATURE -D INET6 -D IPSEC -D PPP_BSDCOMP -D PPP_DEFLATE -D PIPEX -D MROUTING -D MPLS -D BOOT_CONFIG -D USER_PCICONF -D APERTURE -D MTRR -D NTFS -D SUSPEND -D HIBERNATE -D PCIVERBOSE -D USBVERBOSE -D WSDISPLAY_COMPAT_USL -D WSDISPLAY_COMPAT_RAWKBD -D WSDISPLAY_DEFAULTSCREENS=6 -D X86EMU -D ONEWIREVERBOSE -D MULTIPROCESSOR -D MAXUSERS=80 -D _KERNEL -O2 -Wno-pointer-sign -Wno-address-of-packed-member -Wno-constant-conversion -Wno-unused-but-set-variable -Wno-gnu-folding-constant -fdebug-compilation-dir=/usr/src/sys/arch/amd64/compile/GENERIC.MP/obj -ferror-limit 19 -fwrapv -D_RET_PROTECTOR -ret-protector -fcf-protection=branch -fgnuc-version=4.2.1 -vectorize-loops -vectorize-slp -fno-builtin-malloc -fno-builtin-calloc -fno-builtin-realloc -fno-builtin-valloc -fno-builtin-free -fno-builtin-strdup -fno-builtin-strndup -analyzer-output=html -faddrsig -o /home/ben/Projects/scan/2024-01-11-110808-61670-1 -x c /usr/src/sys/kern/vfs_cache.c
1/* $OpenBSD: vfs_cache.c,v 1.58 2022/08/14 01:58:28 jsg Exp $ */
2/* $NetBSD: vfs_cache.c,v 1.13 1996/02/04 02:18:09 christos Exp $ */
3
4/*
5 * Copyright (c) 1989, 1993
6 * The Regents of the University of California. All rights reserved.
7 *
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
16 * 3. Neither the name of the University nor the names of its contributors
17 * may be used to endorse or promote products derived from this software
18 * without specific prior written permission.
19 *
20 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
21 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
24 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
25 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
26 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
27 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
28 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
29 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
30 * SUCH DAMAGE.
31 *
32 * @(#)vfs_cache.c 8.3 (Berkeley) 8/22/94
33 */
34
35#include <sys/param.h>
36#include <sys/systm.h>
37#include <sys/vnode.h>
38#include <sys/lock.h>
39#include <sys/namei.h>
40#include <sys/errno.h>
41#include <sys/pool.h>
42
43/*
44 * TODO: namecache access should really be locked.
45 */
46
47/*
48 * For simplicity (and economy of storage), names longer than
49 * a maximum length of NAMECACHE_MAXLEN are not cached; they occur
50 * infrequently in any case, and are almost never of interest.
51 *
52 * Upon reaching the last segment of a path, if the reference
53 * is for DELETE, or NOCACHE is set (rewrite), and the
54 * name is located in the cache, it will be dropped.
55 */
56
57/*
58 * Structures associated with name caching.
59 */
60long numcache; /* total number of cache entries allocated */
61long numneg; /* number of negative cache entries */
62
63TAILQ_HEAD(, namecache)struct { struct namecache *tqh_first; struct namecache **tqh_last
; }
nclruhead; /* Regular Entry LRU chain */
64TAILQ_HEAD(, namecache)struct { struct namecache *tqh_first; struct namecache **tqh_last
; }
nclruneghead; /* Negative Entry LRU chain */
65struct nchstats nchstats; /* cache effectiveness statistics */
66
67int doingcache = 1; /* 1 => enable the cache */
68
69struct pool nch_pool;
70
71void cache_zap(struct namecache *);
72u_long nextvnodeid;
73
74static inline int
75namecache_compare(const struct namecache *n1, const struct namecache *n2)
76{
77 if (n1->nc_nlen == n2->nc_nlen)
78 return (memcmp(n1->nc_name, n2->nc_name, n1->nc_nlen)__builtin_memcmp((n1->nc_name), (n2->nc_name), (n1->
nc_nlen))
);
79 else
80 return (n1->nc_nlen - n2->nc_nlen);
81}
82
83RBT_PROTOTYPE(namecache_rb_cache, namecache, n_rbcache, namecache_compare)extern const struct rb_type *const namecache_rb_cache_RBT_TYPE
; __attribute__((__unused__)) static inline void namecache_rb_cache_RBT_INIT
(struct namecache_rb_cache *head) { _rb_init(&head->rbh_root
); } __attribute__((__unused__)) static inline struct namecache
* namecache_rb_cache_RBT_INSERT(struct namecache_rb_cache *head
, struct namecache *elm) { return _rb_insert(namecache_rb_cache_RBT_TYPE
, &head->rbh_root, elm); } __attribute__((__unused__))
static inline struct namecache * namecache_rb_cache_RBT_REMOVE
(struct namecache_rb_cache *head, struct namecache *elm) { return
_rb_remove(namecache_rb_cache_RBT_TYPE, &head->rbh_root
, elm); } __attribute__((__unused__)) static inline struct namecache
* namecache_rb_cache_RBT_FIND(struct namecache_rb_cache *head
, const struct namecache *key) { return _rb_find(namecache_rb_cache_RBT_TYPE
, &head->rbh_root, key); } __attribute__((__unused__))
static inline struct namecache * namecache_rb_cache_RBT_NFIND
(struct namecache_rb_cache *head, const struct namecache *key
) { return _rb_nfind(namecache_rb_cache_RBT_TYPE, &head->
rbh_root, key); } __attribute__((__unused__)) static inline struct
namecache * namecache_rb_cache_RBT_ROOT(struct namecache_rb_cache
*head) { return _rb_root(namecache_rb_cache_RBT_TYPE, &head
->rbh_root); } __attribute__((__unused__)) static inline int
namecache_rb_cache_RBT_EMPTY(struct namecache_rb_cache *head
) { return _rb_empty(&head->rbh_root); } __attribute__
((__unused__)) static inline struct namecache * namecache_rb_cache_RBT_MIN
(struct namecache_rb_cache *head) { return _rb_min(namecache_rb_cache_RBT_TYPE
, &head->rbh_root); } __attribute__((__unused__)) static
inline struct namecache * namecache_rb_cache_RBT_MAX(struct namecache_rb_cache
*head) { return _rb_max(namecache_rb_cache_RBT_TYPE, &head
->rbh_root); } __attribute__((__unused__)) static inline struct
namecache * namecache_rb_cache_RBT_NEXT(struct namecache *elm
) { return _rb_next(namecache_rb_cache_RBT_TYPE, elm); } __attribute__
((__unused__)) static inline struct namecache * namecache_rb_cache_RBT_PREV
(struct namecache *elm) { return _rb_prev(namecache_rb_cache_RBT_TYPE
, elm); } __attribute__((__unused__)) static inline struct namecache
* namecache_rb_cache_RBT_LEFT(struct namecache *elm) { return
_rb_left(namecache_rb_cache_RBT_TYPE, elm); } __attribute__(
(__unused__)) static inline struct namecache * namecache_rb_cache_RBT_RIGHT
(struct namecache *elm) { return _rb_right(namecache_rb_cache_RBT_TYPE
, elm); } __attribute__((__unused__)) static inline struct namecache
* namecache_rb_cache_RBT_PARENT(struct namecache *elm) { return
_rb_parent(namecache_rb_cache_RBT_TYPE, elm); } __attribute__
((__unused__)) static inline void namecache_rb_cache_RBT_SET_LEFT
(struct namecache *elm, struct namecache *left) { _rb_set_left
(namecache_rb_cache_RBT_TYPE, elm, left); } __attribute__((__unused__
)) static inline void namecache_rb_cache_RBT_SET_RIGHT(struct
namecache *elm, struct namecache *right) { _rb_set_right(namecache_rb_cache_RBT_TYPE
, elm, right); } __attribute__((__unused__)) static inline void
namecache_rb_cache_RBT_SET_PARENT(struct namecache *elm, struct
namecache *parent) { _rb_set_parent(namecache_rb_cache_RBT_TYPE
, elm, parent); } __attribute__((__unused__)) static inline void
namecache_rb_cache_RBT_POISON(struct namecache *elm, unsigned
long poison) { _rb_poison(namecache_rb_cache_RBT_TYPE, elm, poison
); } __attribute__((__unused__)) static inline int namecache_rb_cache_RBT_CHECK
(struct namecache *elm, unsigned long poison) { return _rb_check
(namecache_rb_cache_RBT_TYPE, elm, poison); }
;
84RBT_GENERATE(namecache_rb_cache, namecache, n_rbcache, namecache_compare)static int namecache_rb_cache_RBT_COMPARE(const void *lptr, const
void *rptr) { const struct namecache *l = lptr, *r = rptr; return
namecache_compare(l, r); } static const struct rb_type namecache_rb_cache_RBT_INFO
= { namecache_rb_cache_RBT_COMPARE, ((void *)0), __builtin_offsetof
(struct namecache, n_rbcache), }; const struct rb_type *const
namecache_rb_cache_RBT_TYPE = &namecache_rb_cache_RBT_INFO
;
85
86void
87cache_tree_init(struct namecache_rb_cache *tree)
88{
89 RBT_INIT(namecache_rb_cache, tree)namecache_rb_cache_RBT_INIT(tree);
90}
91
92/*
93 * blow away a namecache entry
94 */
95void
96cache_zap(struct namecache *ncp)
97{
98 struct vnode *dvp = NULL((void *)0);
99
100 if (ncp->nc_vp != NULL((void *)0)) {
101 TAILQ_REMOVE(&nclruhead, ncp, nc_lru)do { if (((ncp)->nc_lru.tqe_next) != ((void *)0)) (ncp)->
nc_lru.tqe_next->nc_lru.tqe_prev = (ncp)->nc_lru.tqe_prev
; else (&nclruhead)->tqh_last = (ncp)->nc_lru.tqe_prev
; *(ncp)->nc_lru.tqe_prev = (ncp)->nc_lru.tqe_next; ((ncp
)->nc_lru.tqe_prev) = ((void *)-1); ((ncp)->nc_lru.tqe_next
) = ((void *)-1); } while (0)
;
102 numcache--;
103 } else {
104 TAILQ_REMOVE(&nclruneghead, ncp, nc_neg)do { if (((ncp)->nc_neg.tqe_next) != ((void *)0)) (ncp)->
nc_neg.tqe_next->nc_neg.tqe_prev = (ncp)->nc_neg.tqe_prev
; else (&nclruneghead)->tqh_last = (ncp)->nc_neg.tqe_prev
; *(ncp)->nc_neg.tqe_prev = (ncp)->nc_neg.tqe_next; ((ncp
)->nc_neg.tqe_prev) = ((void *)-1); ((ncp)->nc_neg.tqe_next
) = ((void *)-1); } while (0)
;
105 numneg--;
106 }
107 if (ncp->nc_dvp) {
108 RBT_REMOVE(namecache_rb_cache, &ncp->nc_dvp->v_nc_tree, ncp)namecache_rb_cache_RBT_REMOVE(&ncp->nc_dvp->v_nc_tree
, ncp)
;
109 if (RBT_EMPTY(namecache_rb_cache, &ncp->nc_dvp->v_nc_tree)namecache_rb_cache_RBT_EMPTY(&ncp->nc_dvp->v_nc_tree
)
)
110 dvp = ncp->nc_dvp;
111 }
112 if (ncp->nc_vp && (ncp->nc_vpid == ncp->nc_vp->v_id)) {
113 if (ncp->nc_vp != ncp->nc_dvp &&
114 ncp->nc_vp->v_type == VDIR &&
115 (ncp->nc_nlen > 2 ||
116 (ncp->nc_nlen > 1 &&
117 ncp->nc_name[1] != '.') ||
118 (ncp->nc_nlen > 0 &&
119 ncp->nc_name[0] != '.'))) {
120 TAILQ_REMOVE(&ncp->nc_vp->v_cache_dst, ncp, nc_me)do { if (((ncp)->nc_me.tqe_next) != ((void *)0)) (ncp)->
nc_me.tqe_next->nc_me.tqe_prev = (ncp)->nc_me.tqe_prev;
else (&ncp->nc_vp->v_cache_dst)->tqh_last = (ncp
)->nc_me.tqe_prev; *(ncp)->nc_me.tqe_prev = (ncp)->nc_me
.tqe_next; ((ncp)->nc_me.tqe_prev) = ((void *)-1); ((ncp)->
nc_me.tqe_next) = ((void *)-1); } while (0)
;
121 }
122 }
123 pool_put(&nch_pool, ncp);
124 if (dvp)
125 vdrop(dvp);
126}
127
128/*
129 * Look for a name in the cache.
130 * dvp points to the directory to search. The componentname cnp holds
131 * the information on the entry being sought, such as its length
132 * and its name. If the lookup succeeds, vpp is set to point to the vnode
133 * and an error of 0 is returned. If the lookup determines the name does
134 * not exist (negative caching) an error of ENOENT is returned. If the
135 * lookup fails, an error of -1 is returned.
136 */
137int
138cache_lookup(struct vnode *dvp, struct vnode **vpp,
139 struct componentname *cnp)
140{
141 struct namecache *ncp;
142 struct namecache n;
143 struct vnode *vp;
144 u_long vpid;
145 int error;
146
147 *vpp = NULL((void *)0);
148
149 if (!doingcache) {
150 cnp->cn_flags &= ~MAKEENTRY0x004000;
151 return (-1);
152 }
153 if (cnp->cn_namelen > NAMECACHE_MAXLEN31) {
154 nchstats.ncs_long++;
155 cnp->cn_flags &= ~MAKEENTRY0x004000;
156 return (-1);
157 }
158
159 /* lookup in directory vnode's redblack tree */
160 n.nc_nlen = cnp->cn_namelen;
161 memcpy(n.nc_name, cnp->cn_nameptr, n.nc_nlen)__builtin_memcpy((n.nc_name), (cnp->cn_nameptr), (n.nc_nlen
))
;
162 ncp = RBT_FIND(namecache_rb_cache, &dvp->v_nc_tree, &n)namecache_rb_cache_RBT_FIND(&dvp->v_nc_tree, &n);
163
164 if (ncp == NULL((void *)0)) {
165 nchstats.ncs_miss++;
166 return (-1);
167 }
168 if ((cnp->cn_flags & MAKEENTRY0x004000) == 0) {
169 nchstats.ncs_badhits++;
170 goto remove;
171 } else if (ncp->nc_vp == NULL((void *)0)) {
172 if (cnp->cn_nameiop != CREATE1 ||
173 (cnp->cn_flags & ISLASTCN0x008000) == 0) {
174 nchstats.ncs_neghits++;
175 /*
176 * Move this slot to end of the negative LRU chain,
177 */
178 if (TAILQ_NEXT(ncp, nc_neg)((ncp)->nc_neg.tqe_next) != NULL((void *)0)) {
179 TAILQ_REMOVE(&nclruneghead, ncp, nc_neg)do { if (((ncp)->nc_neg.tqe_next) != ((void *)0)) (ncp)->
nc_neg.tqe_next->nc_neg.tqe_prev = (ncp)->nc_neg.tqe_prev
; else (&nclruneghead)->tqh_last = (ncp)->nc_neg.tqe_prev
; *(ncp)->nc_neg.tqe_prev = (ncp)->nc_neg.tqe_next; ((ncp
)->nc_neg.tqe_prev) = ((void *)-1); ((ncp)->nc_neg.tqe_next
) = ((void *)-1); } while (0)
;
180 TAILQ_INSERT_TAIL(&nclruneghead, ncp,do { (ncp)->nc_neg.tqe_next = ((void *)0); (ncp)->nc_neg
.tqe_prev = (&nclruneghead)->tqh_last; *(&nclruneghead
)->tqh_last = (ncp); (&nclruneghead)->tqh_last = &
(ncp)->nc_neg.tqe_next; } while (0)
181 nc_neg)do { (ncp)->nc_neg.tqe_next = ((void *)0); (ncp)->nc_neg
.tqe_prev = (&nclruneghead)->tqh_last; *(&nclruneghead
)->tqh_last = (ncp); (&nclruneghead)->tqh_last = &
(ncp)->nc_neg.tqe_next; } while (0)
;
182 }
183 return (ENOENT2);
184 } else {
185 nchstats.ncs_badhits++;
186 goto remove;
187 }
188 } else if (ncp->nc_vpid != ncp->nc_vp->v_id) {
189 nchstats.ncs_falsehits++;
190 goto remove;
191 }
192
193 /*
194 * Move this slot to end of the regular LRU chain.
195 */
196 if (TAILQ_NEXT(ncp, nc_lru)((ncp)->nc_lru.tqe_next) != NULL((void *)0)) {
197 TAILQ_REMOVE(&nclruhead, ncp, nc_lru)do { if (((ncp)->nc_lru.tqe_next) != ((void *)0)) (ncp)->
nc_lru.tqe_next->nc_lru.tqe_prev = (ncp)->nc_lru.tqe_prev
; else (&nclruhead)->tqh_last = (ncp)->nc_lru.tqe_prev
; *(ncp)->nc_lru.tqe_prev = (ncp)->nc_lru.tqe_next; ((ncp
)->nc_lru.tqe_prev) = ((void *)-1); ((ncp)->nc_lru.tqe_next
) = ((void *)-1); } while (0)
;
198 TAILQ_INSERT_TAIL(&nclruhead, ncp, nc_lru)do { (ncp)->nc_lru.tqe_next = ((void *)0); (ncp)->nc_lru
.tqe_prev = (&nclruhead)->tqh_last; *(&nclruhead)->
tqh_last = (ncp); (&nclruhead)->tqh_last = &(ncp)->
nc_lru.tqe_next; } while (0)
;
199 }
200
201 vp = ncp->nc_vp;
202 vpid = vp->v_id;
203 if (vp == dvp) { /* lookup on "." */
204 vref(dvp);
205 error = 0;
206 } else if (cnp->cn_flags & ISDOTDOT0x002000) {
207 VOP_UNLOCK(dvp);
208 cnp->cn_flags |= PDIRUNLOCK0x200000;
209 error = vget(vp, LK_EXCLUSIVE0x0001UL);
210 /*
211 * If the above vget() succeeded and both LOCKPARENT and
212 * ISLASTCN is set, lock the directory vnode as well.
213 */
214 if (!error && (~cnp->cn_flags & (LOCKPARENT0x0008|ISLASTCN0x008000)) == 0) {
215 if ((error = vn_lock(dvp, LK_EXCLUSIVE0x0001UL)) != 0) {
216 vput(vp);
217 return (error);
218 }
219 cnp->cn_flags &= ~PDIRUNLOCK0x200000;
220 }
221 } else {
222 error = vget(vp, LK_EXCLUSIVE0x0001UL);
223 /*
224 * If the above vget() failed or either of LOCKPARENT or
225 * ISLASTCN is set, unlock the directory vnode.
226 */
227 if (error || (~cnp->cn_flags & (LOCKPARENT0x0008|ISLASTCN0x008000)) != 0) {
228 VOP_UNLOCK(dvp);
229 cnp->cn_flags |= PDIRUNLOCK0x200000;
230 }
231 }
232
233 /*
234 * Check that the lock succeeded, and that the capability number did
235 * not change while we were waiting for the lock.
236 */
237 if (error || vpid != vp->v_id) {
238 if (!error) {
239 vput(vp);
240 nchstats.ncs_falsehits++;
241 } else
242 nchstats.ncs_badhits++;
243 /*
244 * The parent needs to be locked when we return to VOP_LOOKUP().
245 * The `.' case here should be extremely rare (if it can happen
246 * at all), so we don't bother optimizing out the unlock/relock.
247 */
248 if (vp == dvp || error ||
249 (~cnp->cn_flags & (LOCKPARENT0x0008|ISLASTCN0x008000)) != 0) {
250 if ((error = vn_lock(dvp, LK_EXCLUSIVE0x0001UL)) != 0)
251 return (error);
252 cnp->cn_flags &= ~PDIRUNLOCK0x200000;
253 }
254 return (-1);
255 }
256
257 nchstats.ncs_goodhits++;
258 *vpp = vp;
259 return (0);
260
261remove:
262 /*
263 * Last component and we are renaming or deleting,
264 * the cache entry is invalid, or otherwise don't
265 * want cache entry to exist.
266 */
267 cache_zap(ncp);
268 return (-1);
269}
270
271/*
272 * Scan cache looking for name of directory entry pointing at vp.
273 *
274 * Fill in dvpp.
275 *
276 * If bufp is non-NULL, also place the name in the buffer which starts
277 * at bufp, immediately before *bpp, and move bpp backwards to point
278 * at the start of it. (Yes, this is a little baroque, but it's done
279 * this way to cater to the whims of getcwd).
280 *
281 * Returns 0 on success, -1 on cache miss, positive errno on failure.
282 *
283 * TODO: should we return *dvpp locked?
284 */
285
286int
287cache_revlookup(struct vnode *vp, struct vnode **dvpp, char **bpp, char *bufp)
288{
289 struct namecache *ncp;
290 struct vnode *dvp = NULL((void *)0);
291 char *bp;
292
293 if (!doingcache)
294 goto out;
295 TAILQ_FOREACH(ncp, &vp->v_cache_dst, nc_me)for((ncp) = ((&vp->v_cache_dst)->tqh_first); (ncp) !=
((void *)0); (ncp) = ((ncp)->nc_me.tqe_next))
{
296 dvp = ncp->nc_dvp;
297 if (dvp && dvp != vp && ncp->nc_dvpid == dvp->v_id)
298 goto found;
299 }
300 goto miss;
301found:
302#ifdef DIAGNOSTIC1
303 if (ncp->nc_nlen == 1 &&
304 ncp->nc_name[0] == '.')
305 panic("cache_revlookup: found entry for .");
306 if (ncp->nc_nlen == 2 &&
307 ncp->nc_name[0] == '.' &&
308 ncp->nc_name[1] == '.')
309 panic("cache_revlookup: found entry for ..");
310#endif
311 nchstats.ncs_revhits++;
312
313 if (bufp != NULL((void *)0)) {
314 bp = *bpp;
315 bp -= ncp->nc_nlen;
316 if (bp <= bufp) {
317 *dvpp = NULL((void *)0);
318 return (ERANGE34);
319 }
320 memcpy(bp, ncp->nc_name, ncp->nc_nlen)__builtin_memcpy((bp), (ncp->nc_name), (ncp->nc_nlen));
321 *bpp = bp;
322 }
323
324 *dvpp = dvp;
325
326 /*
327 * XXX: Should we vget() here to have more
328 * consistent semantics with cache_lookup()?
329 */
330 return (0);
331
332miss:
333 nchstats.ncs_revmiss++;
334out:
335 *dvpp = NULL((void *)0);
336 return (-1);
337}
338
339/*
340 * Add an entry to the cache
341 */
342void
343cache_enter(struct vnode *dvp, struct vnode *vp, struct componentname *cnp)
344{
345 struct namecache *ncp, *lncp;
346
347 if (!doingcache || cnp->cn_namelen > NAMECACHE_MAXLEN31)
348 return;
349
350 /*
351 * allocate, or recycle (free and allocate) an ncp.
352 */
353 if (numcache >= initialvnodes) {
354 if ((ncp = TAILQ_FIRST(&nclruhead)((&nclruhead)->tqh_first)) != NULL((void *)0))
355 cache_zap(ncp);
356 else if ((ncp = TAILQ_FIRST(&nclruneghead)((&nclruneghead)->tqh_first)) != NULL((void *)0))
357 cache_zap(ncp);
358 else
359 panic("wtf? leak?");
360 }
361 ncp = pool_get(&nch_pool, PR_WAITOK0x0001|PR_ZERO0x0008);
362
363 /* grab the vnode we just found */
364 ncp->nc_vp = vp;
365 if (vp)
366 ncp->nc_vpid = vp->v_id;
367
368 /* fill in cache info */
369 ncp->nc_dvp = dvp;
370 ncp->nc_dvpid = dvp->v_id;
371 ncp->nc_nlen = cnp->cn_namelen;
372 memcpy(ncp->nc_name, cnp->cn_nameptr, ncp->nc_nlen)__builtin_memcpy((ncp->nc_name), (cnp->cn_nameptr), (ncp
->nc_nlen))
;
373 if (RBT_EMPTY(namecache_rb_cache, &dvp->v_nc_tree)namecache_rb_cache_RBT_EMPTY(&dvp->v_nc_tree)) {
374 vhold(dvp);
375 }
376 if ((lncp = RBT_INSERT(namecache_rb_cache, &dvp->v_nc_tree, ncp)namecache_rb_cache_RBT_INSERT(&dvp->v_nc_tree, ncp))
Although the value stored to 'lncp' is used in the enclosing expression, the value is never actually read from 'lncp'
377 != NULL((void *)0)) {
378 /* someone has raced us and added a different entry
379 * for the same vnode (different ncp) - we don't need
380 * this entry, so free it and we are done.
381 */
382 pool_put(&nch_pool, ncp);
383 /* we know now dvp->v_nc_tree is not empty, no need
384 * to vdrop here
385 */
386 goto done;
387 }
388 if (vp) {
389 TAILQ_INSERT_TAIL(&nclruhead, ncp, nc_lru)do { (ncp)->nc_lru.tqe_next = ((void *)0); (ncp)->nc_lru
.tqe_prev = (&nclruhead)->tqh_last; *(&nclruhead)->
tqh_last = (ncp); (&nclruhead)->tqh_last = &(ncp)->
nc_lru.tqe_next; } while (0)
;
390 numcache++;
391 /* don't put . or .. in the reverse map */
392 if (vp != dvp && vp->v_type == VDIR &&
393 (ncp->nc_nlen > 2 ||
394 (ncp->nc_nlen > 1 &&
395 ncp->nc_name[1] != '.') ||
396 (ncp->nc_nlen > 0 &&
397 ncp->nc_name[0] != '.')))
398 TAILQ_INSERT_TAIL(&vp->v_cache_dst, ncp,do { (ncp)->nc_me.tqe_next = ((void *)0); (ncp)->nc_me.
tqe_prev = (&vp->v_cache_dst)->tqh_last; *(&vp->
v_cache_dst)->tqh_last = (ncp); (&vp->v_cache_dst)->
tqh_last = &(ncp)->nc_me.tqe_next; } while (0)
399 nc_me)do { (ncp)->nc_me.tqe_next = ((void *)0); (ncp)->nc_me.
tqe_prev = (&vp->v_cache_dst)->tqh_last; *(&vp->
v_cache_dst)->tqh_last = (ncp); (&vp->v_cache_dst)->
tqh_last = &(ncp)->nc_me.tqe_next; } while (0)
;
400 } else {
401 TAILQ_INSERT_TAIL(&nclruneghead, ncp, nc_neg)do { (ncp)->nc_neg.tqe_next = ((void *)0); (ncp)->nc_neg
.tqe_prev = (&nclruneghead)->tqh_last; *(&nclruneghead
)->tqh_last = (ncp); (&nclruneghead)->tqh_last = &
(ncp)->nc_neg.tqe_next; } while (0)
;
402 numneg++;
403 }
404 if (numneg > initialvnodes) {
405 if ((ncp = TAILQ_FIRST(&nclruneghead)((&nclruneghead)->tqh_first))
406 != NULL((void *)0))
407 cache_zap(ncp);
408 }
409done:
410 return;
411}
412
413
414/*
415 * Name cache initialization, from vfs_init() when we are booting
416 */
417void
418nchinit(void)
419{
420 TAILQ_INIT(&nclruhead)do { (&nclruhead)->tqh_first = ((void *)0); (&nclruhead
)->tqh_last = &(&nclruhead)->tqh_first; } while
(0)
;
421 TAILQ_INIT(&nclruneghead)do { (&nclruneghead)->tqh_first = ((void *)0); (&nclruneghead
)->tqh_last = &(&nclruneghead)->tqh_first; } while
(0)
;
422 pool_init(&nch_pool, sizeof(struct namecache), 0, IPL_NONE0x0, PR_WAITOK0x0001,
423 "nchpl", NULL((void *)0));
424}
425
426/*
427 * Cache flush, a particular vnode; called when a vnode is renamed to
428 * hide entries that would now be invalid
429 */
430void
431cache_purge(struct vnode *vp)
432{
433 struct namecache *ncp;
434
435 /* We should never have destinations cached for a non-VDIR vnode. */
436 KASSERT(vp->v_type == VDIR || TAILQ_EMPTY(&vp->v_cache_dst))((vp->v_type == VDIR || (((&vp->v_cache_dst)->tqh_first
) == ((void *)0))) ? (void)0 : __assert("diagnostic ", "/usr/src/sys/kern/vfs_cache.c"
, 436, "vp->v_type == VDIR || TAILQ_EMPTY(&vp->v_cache_dst)"
))
;
437
438 while ((ncp = TAILQ_FIRST(&vp->v_cache_dst)((&vp->v_cache_dst)->tqh_first)))
439 cache_zap(ncp);
440 while ((ncp = RBT_ROOT(namecache_rb_cache, &vp->v_nc_tree)namecache_rb_cache_RBT_ROOT(&vp->v_nc_tree)))
441 cache_zap(ncp);
442
443 /* XXX this blows goats */
444 vp->v_id = ++nextvnodeid;
445 if (vp->v_id == 0)
446 vp->v_id = ++nextvnodeid;
447}
448
449/*
450 * Cache flush, a whole filesystem; called when filesys is umounted to
451 * remove entries that would now be invalid
452 */
453void
454cache_purgevfs(struct mount *mp)
455{
456 struct namecache *ncp, *nxtcp;
457
458 /* whack the regular entries */
459 TAILQ_FOREACH_SAFE(ncp, &nclruhead, nc_lru, nxtcp)for ((ncp) = ((&nclruhead)->tqh_first); (ncp) != ((void
*)0) && ((nxtcp) = ((ncp)->nc_lru.tqe_next), 1); (
ncp) = (nxtcp))
{
460 if (ncp->nc_dvp == NULL((void *)0) || ncp->nc_dvp->v_mount != mp)
461 continue;
462 /* free the resources we had */
463 cache_zap(ncp);
464 }
465 /* whack the negative entries */
466 TAILQ_FOREACH_SAFE(ncp, &nclruneghead, nc_neg, nxtcp)for ((ncp) = ((&nclruneghead)->tqh_first); (ncp) != ((
void *)0) && ((nxtcp) = ((ncp)->nc_neg.tqe_next), 1
); (ncp) = (nxtcp))
{
467 if (ncp->nc_dvp == NULL((void *)0) || ncp->nc_dvp->v_mount != mp)
468 continue;
469 /* free the resources we had */
470 cache_zap(ncp);
471 }
472}