| File: | nfs/nfs_bio.c |
| Warning: | line 545, column 2 Value stored to 'nmp' is never read |
Press '?' to see keyboard shortcuts
Keyboard shortcuts:
| 1 | /* $OpenBSD: nfs_bio.c,v 1.84 2019/07/25 01:43:21 cheloha Exp $ */ |
| 2 | /* $NetBSD: nfs_bio.c,v 1.25.4.2 1996/07/08 20:47:04 jtc Exp $ */ |
| 3 | |
| 4 | /* |
| 5 | * Copyright (c) 1989, 1993 |
| 6 | * The Regents of the University of California. All rights reserved. |
| 7 | * |
| 8 | * This code is derived from software contributed to Berkeley by |
| 9 | * Rick Macklem at The University of Guelph. |
| 10 | * |
| 11 | * Redistribution and use in source and binary forms, with or without |
| 12 | * modification, are permitted provided that the following conditions |
| 13 | * are met: |
| 14 | * 1. Redistributions of source code must retain the above copyright |
| 15 | * notice, this list of conditions and the following disclaimer. |
| 16 | * 2. Redistributions in binary form must reproduce the above copyright |
| 17 | * notice, this list of conditions and the following disclaimer in the |
| 18 | * documentation and/or other materials provided with the distribution. |
| 19 | * 3. Neither the name of the University nor the names of its contributors |
| 20 | * may be used to endorse or promote products derived from this software |
| 21 | * without specific prior written permission. |
| 22 | * |
| 23 | * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND |
| 24 | * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| 25 | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| 26 | * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE |
| 27 | * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL |
| 28 | * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS |
| 29 | * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) |
| 30 | * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT |
| 31 | * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY |
| 32 | * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF |
| 33 | * SUCH DAMAGE. |
| 34 | * |
| 35 | * @(#)nfs_bio.c 8.9 (Berkeley) 3/30/95 |
| 36 | */ |
| 37 | |
| 38 | #include <sys/param.h> |
| 39 | #include <sys/systm.h> |
| 40 | #include <sys/resourcevar.h> |
| 41 | #include <sys/signalvar.h> |
| 42 | #include <sys/proc.h> |
| 43 | #include <sys/buf.h> |
| 44 | #include <sys/vnode.h> |
| 45 | #include <sys/mount.h> |
| 46 | #include <sys/kernel.h> |
| 47 | #include <sys/namei.h> |
| 48 | #include <sys/queue.h> |
| 49 | #include <sys/time.h> |
| 50 | |
| 51 | #include <nfs/nfsproto.h> |
| 52 | #include <nfs/nfs.h> |
| 53 | #include <nfs/nfsmount.h> |
| 54 | #include <nfs/nfsnode.h> |
| 55 | #include <nfs/nfs_var.h> |
| 56 | |
| 57 | extern int nfs_numasync; |
| 58 | extern struct nfsstats nfsstats; |
| 59 | struct nfs_bufqhead nfs_bufq; |
| 60 | uint32_t nfs_bufqmax, nfs_bufqlen; |
| 61 | |
| 62 | struct buf *nfs_getcacheblk(struct vnode *, daddr_t, int, struct proc *); |
| 63 | |
| 64 | /* |
| 65 | * Vnode op for read using bio |
| 66 | * Any similarity to readip() is purely coincidental |
| 67 | */ |
| 68 | int |
| 69 | nfs_bioread(struct vnode *vp, struct uio *uio, int ioflag, struct ucred *cred) |
| 70 | { |
| 71 | struct nfsnode *np = VTONFS(vp)((struct nfsnode *)(vp)->v_data); |
| 72 | int biosize, diff; |
| 73 | struct buf *bp = NULL((void *)0), *rabp; |
| 74 | struct vattr vattr; |
| 75 | struct proc *p; |
| 76 | struct nfsmount *nmp = VFSTONFS(vp->v_mount)((struct nfsmount *)((vp->v_mount)->mnt_data)); |
| 77 | daddr_t lbn, bn, rabn; |
| 78 | caddr_t baddr; |
| 79 | int got_buf = 0, nra, error = 0, n = 0, on = 0, not_readin; |
| 80 | off_t offdiff; |
| 81 | |
| 82 | #ifdef DIAGNOSTIC1 |
| 83 | if (uio->uio_rw != UIO_READ) |
| 84 | panic("nfs_read mode"); |
| 85 | #endif |
| 86 | if (uio->uio_resid == 0) |
| 87 | return (0); |
| 88 | if (uio->uio_offset < 0) |
| 89 | return (EINVAL22); |
| 90 | p = uio->uio_procp; |
| 91 | if ((nmp->nm_flag & (NFSMNT_NFSV30x00000200 | NFSMNT_GOTFSINFO0x00100000)) == NFSMNT_NFSV30x00000200) |
| 92 | (void)nfs_fsinfo(nmp, vp, cred, p); |
| 93 | biosize = nmp->nm_rsize; |
| 94 | /* |
| 95 | * For nfs, cache consistency can only be maintained approximately. |
| 96 | * Although RFC1094 does not specify the criteria, the following is |
| 97 | * believed to be compatible with the reference port. |
| 98 | * For nfs: |
| 99 | * If the file's modify time on the server has changed since the |
| 100 | * last read rpc or you have written to the file, |
| 101 | * you may have lost data cache consistency with the |
| 102 | * server, so flush all of the file's data out of the cache. |
| 103 | * Then force a getattr rpc to ensure that you have up to date |
| 104 | * attributes. |
| 105 | */ |
| 106 | if (np->n_flag & NMODIFIED0x0004) { |
| 107 | NFS_INVALIDATE_ATTRCACHE(np)((np)->n_attrstamp = 0); |
| 108 | error = VOP_GETATTR(vp, &vattr, cred, p); |
| 109 | if (error) |
| 110 | return (error); |
| 111 | np->n_mtime = vattr.va_mtime; |
| 112 | } else { |
| 113 | error = VOP_GETATTR(vp, &vattr, cred, p); |
| 114 | if (error) |
| 115 | return (error); |
| 116 | if (timespeccmp(&np->n_mtime, &vattr.va_mtime, !=)(((&np->n_mtime)->tv_sec == (&vattr.va_mtime)-> tv_sec) ? ((&np->n_mtime)->tv_nsec != (&vattr.va_mtime )->tv_nsec) : ((&np->n_mtime)->tv_sec != (&vattr .va_mtime)->tv_sec))) { |
| 117 | error = nfs_vinvalbuf(vp, V_SAVE0x0001, cred, p); |
| 118 | if (error) |
| 119 | return (error); |
| 120 | np->n_mtime = vattr.va_mtime; |
| 121 | } |
| 122 | } |
| 123 | |
| 124 | /* |
| 125 | * update the cache read creds for this vnode |
| 126 | */ |
| 127 | if (np->n_rcred) |
| 128 | crfree(np->n_rcred); |
| 129 | np->n_rcred = cred; |
| 130 | crhold(cred); |
| 131 | |
| 132 | do { |
| 133 | if ((vp->v_flag & VROOT0x0001) && vp->v_type == VLNK) { |
| 134 | return (nfs_readlinkrpc(vp, uio, cred)); |
| 135 | } |
| 136 | baddr = NULL((void *)0); |
| 137 | switch (vp->v_type) { |
| 138 | case VREG: |
| 139 | nfsstats.biocache_reads++; |
| 140 | lbn = uio->uio_offset / biosize; |
| 141 | on = uio->uio_offset & (biosize - 1); |
| 142 | bn = lbn * (biosize / DEV_BSIZE(1 << 9)); |
| 143 | not_readin = 1; |
| 144 | |
| 145 | /* |
| 146 | * Start the read ahead(s), as required. |
| 147 | */ |
| 148 | if (nfs_numasync > 0 && nmp->nm_readahead > 0) { |
| 149 | for (nra = 0; nra < nmp->nm_readahead && |
| 150 | (lbn + 1 + nra) * biosize < np->n_size; nra++) { |
| 151 | rabn = (lbn + 1 + nra) * (biosize / DEV_BSIZE(1 << 9)); |
| 152 | if (!incore(vp, rabn)) { |
| 153 | rabp = nfs_getcacheblk(vp, rabn, biosize, p); |
| 154 | if (!rabp) |
| 155 | return (EINTR4); |
| 156 | if ((rabp->b_flags & (B_DELWRI0x00000080 | B_DONE0x00000100)) == 0) { |
| 157 | rabp->b_flags |= (B_READ0x00008000 | B_ASYNC0x00000004); |
| 158 | if (nfs_asyncio(rabp, 1)) { |
| 159 | rabp->b_flags |= B_INVAL0x00000800; |
| 160 | brelse(rabp); |
| 161 | } |
| 162 | } else |
| 163 | brelse(rabp); |
| 164 | } |
| 165 | } |
| 166 | } |
| 167 | |
| 168 | again: |
| 169 | bp = nfs_getcacheblk(vp, bn, biosize, p); |
| 170 | if (!bp) |
| 171 | return (EINTR4); |
| 172 | got_buf = 1; |
| 173 | if ((bp->b_flags & (B_DONE0x00000100 | B_DELWRI0x00000080)) == 0) { |
| 174 | bp->b_flags |= B_READ0x00008000; |
| 175 | not_readin = 0; |
| 176 | error = nfs_doio(bp, p); |
| 177 | if (error) { |
| 178 | brelse(bp); |
| 179 | return (error); |
| 180 | } |
| 181 | } |
| 182 | n = ulmin(biosize - on, uio->uio_resid); |
| 183 | offdiff = np->n_size - uio->uio_offset; |
| 184 | if (offdiff < (off_t)n) |
| 185 | n = (int)offdiff; |
| 186 | if (not_readin && n > 0) { |
| 187 | if (on < bp->b_validoff || (on + n) > bp->b_validend) { |
| 188 | bp->b_flags |= B_INVAFTERWRITE0x00000800; |
| 189 | if (bp->b_dirtyend > 0) { |
| 190 | if ((bp->b_flags & B_DELWRI0x00000080) == 0) |
| 191 | panic("nfsbioread"); |
| 192 | if (VOP_BWRITE(bp) == EINTR4) |
| 193 | return (EINTR4); |
| 194 | } else |
| 195 | brelse(bp); |
| 196 | goto again; |
| 197 | } |
| 198 | } |
| 199 | diff = (on >= bp->b_validend) ? 0 : (bp->b_validend - on); |
| 200 | if (diff < n) |
| 201 | n = diff; |
| 202 | break; |
| 203 | case VLNK: |
| 204 | nfsstats.biocache_readlinks++; |
| 205 | bp = nfs_getcacheblk(vp, 0, NFS_MAXPATHLEN1024, p); |
| 206 | if (!bp) |
| 207 | return (EINTR4); |
| 208 | if ((bp->b_flags & B_DONE0x00000100) == 0) { |
| 209 | bp->b_flags |= B_READ0x00008000; |
| 210 | error = nfs_doio(bp, p); |
| 211 | if (error) { |
| 212 | brelse(bp); |
| 213 | return (error); |
| 214 | } |
| 215 | } |
| 216 | n = ulmin(uio->uio_resid, NFS_MAXPATHLEN1024 - bp->b_resid); |
| 217 | got_buf = 1; |
| 218 | on = 0; |
| 219 | break; |
| 220 | default: |
| 221 | panic("nfsbioread: type %x unexpected", vp->v_type); |
| 222 | break; |
| 223 | } |
| 224 | |
| 225 | if (n > 0) { |
| 226 | if (!baddr) |
| 227 | baddr = bp->b_data; |
| 228 | error = uiomove(baddr + on, n, uio); |
| 229 | } |
| 230 | |
| 231 | if (vp->v_type == VLNK) |
| 232 | n = 0; |
| 233 | |
| 234 | if (got_buf) |
| 235 | brelse(bp); |
| 236 | } while (error == 0 && uio->uio_resid > 0 && n > 0); |
| 237 | return (error); |
| 238 | } |
| 239 | |
| 240 | /* |
| 241 | * Vnode op for write using bio |
| 242 | */ |
| 243 | int |
| 244 | nfs_write(void *v) |
| 245 | { |
| 246 | struct vop_write_args *ap = v; |
| 247 | int biosize; |
| 248 | struct uio *uio = ap->a_uio; |
| 249 | struct proc *p = uio->uio_procp; |
| 250 | struct vnode *vp = ap->a_vp; |
| 251 | struct nfsnode *np = VTONFS(vp)((struct nfsnode *)(vp)->v_data); |
| 252 | struct ucred *cred = ap->a_cred; |
| 253 | int ioflag = ap->a_ioflag; |
| 254 | struct buf *bp; |
| 255 | struct vattr vattr; |
| 256 | struct nfsmount *nmp = VFSTONFS(vp->v_mount)((struct nfsmount *)((vp->v_mount)->mnt_data)); |
| 257 | daddr_t lbn, bn; |
| 258 | int n, on, error = 0, extended = 0, wrotedta = 0, truncated = 0; |
| 259 | ssize_t overrun; |
| 260 | |
| 261 | #ifdef DIAGNOSTIC1 |
| 262 | if (uio->uio_rw != UIO_WRITE) |
| 263 | panic("nfs_write mode"); |
| 264 | if (uio->uio_segflg == UIO_USERSPACE && uio->uio_procp != curproc({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc) |
| 265 | panic("nfs_write proc"); |
| 266 | #endif |
| 267 | if (vp->v_type != VREG) |
| 268 | return (EIO5); |
| 269 | if (np->n_flag & NWRITEERR0x0008) { |
| 270 | np->n_flag &= ~NWRITEERR0x0008; |
| 271 | return (np->n_error); |
| 272 | } |
| 273 | if ((nmp->nm_flag & (NFSMNT_NFSV30x00000200 | NFSMNT_GOTFSINFO0x00100000)) == NFSMNT_NFSV30x00000200) |
| 274 | (void)nfs_fsinfo(nmp, vp, cred, p); |
| 275 | if (ioflag & (IO_APPEND0x02 | IO_SYNC0x04)) { |
| 276 | if (np->n_flag & NMODIFIED0x0004) { |
| 277 | NFS_INVALIDATE_ATTRCACHE(np)((np)->n_attrstamp = 0); |
| 278 | error = nfs_vinvalbuf(vp, V_SAVE0x0001, cred, p); |
| 279 | if (error) |
| 280 | return (error); |
| 281 | } |
| 282 | if (ioflag & IO_APPEND0x02) { |
| 283 | NFS_INVALIDATE_ATTRCACHE(np)((np)->n_attrstamp = 0); |
| 284 | error = VOP_GETATTR(vp, &vattr, cred, p); |
| 285 | if (error) |
| 286 | return (error); |
| 287 | uio->uio_offset = np->n_size; |
| 288 | } |
| 289 | } |
| 290 | if (uio->uio_offset < 0) |
| 291 | return (EINVAL22); |
| 292 | if (uio->uio_resid == 0) |
| 293 | return (0); |
| 294 | |
| 295 | /* do the filesize rlimit check */ |
| 296 | if ((error = vn_fsizechk(vp, uio, ioflag, &overrun))) |
| 297 | return (error); |
| 298 | |
| 299 | /* |
| 300 | * update the cache write creds for this node. |
| 301 | */ |
| 302 | if (np->n_wcred) |
| 303 | crfree(np->n_wcred); |
| 304 | np->n_wcred = cred; |
| 305 | crhold(cred); |
| 306 | |
| 307 | /* |
| 308 | * I use nm_rsize, not nm_wsize so that all buffer cache blocks |
| 309 | * will be the same size within a filesystem. nfs_writerpc will |
| 310 | * still use nm_wsize when sizing the rpc's. |
| 311 | */ |
| 312 | biosize = nmp->nm_rsize; |
| 313 | do { |
| 314 | |
| 315 | /* |
| 316 | * XXX make sure we aren't cached in the VM page cache |
| 317 | */ |
| 318 | uvm_vnp_uncache(vp); |
| 319 | |
| 320 | nfsstats.biocache_writes++; |
| 321 | lbn = uio->uio_offset / biosize; |
| 322 | on = uio->uio_offset & (biosize-1); |
| 323 | n = ulmin(biosize - on, uio->uio_resid); |
| 324 | bn = lbn * (biosize / DEV_BSIZE(1 << 9)); |
| 325 | again: |
| 326 | bp = nfs_getcacheblk(vp, bn, biosize, p); |
| 327 | if (!bp) { |
| 328 | error = EINTR4; |
| 329 | goto out; |
| 330 | } |
| 331 | np->n_flag |= NMODIFIED0x0004; |
| 332 | if (uio->uio_offset + n > np->n_size) { |
| 333 | np->n_size = uio->uio_offset + n; |
| 334 | uvm_vnp_setsize(vp, (u_long)np->n_size); |
| 335 | extended = 1; |
| 336 | } else if (uio->uio_offset + n < np->n_size) |
| 337 | truncated = 1; |
| 338 | |
| 339 | /* |
| 340 | * If the new write will leave a contiguous dirty |
| 341 | * area, just update the b_dirtyoff and b_dirtyend, |
| 342 | * otherwise force a write rpc of the old dirty area. |
| 343 | */ |
| 344 | if (bp->b_dirtyend > 0 && |
| 345 | (on > bp->b_dirtyend || (on + n) < bp->b_dirtyoff)) { |
| 346 | bp->b_proc = p; |
| 347 | if (VOP_BWRITE(bp) == EINTR4) { |
| 348 | error = EINTR4; |
| 349 | goto out; |
| 350 | } |
| 351 | goto again; |
| 352 | } |
| 353 | |
| 354 | error = uiomove((char *)bp->b_data + on, n, uio); |
| 355 | if (error) { |
| 356 | bp->b_flags |= B_ERROR0x00000400; |
| 357 | brelse(bp); |
| 358 | goto out; |
| 359 | } |
| 360 | if (bp->b_dirtyend > 0) { |
| 361 | bp->b_dirtyoff = min(on, bp->b_dirtyoff); |
| 362 | bp->b_dirtyend = max((on + n), bp->b_dirtyend); |
| 363 | } else { |
| 364 | bp->b_dirtyoff = on; |
| 365 | bp->b_dirtyend = on + n; |
| 366 | } |
| 367 | if (bp->b_validend == 0 || bp->b_validend < bp->b_dirtyoff || |
| 368 | bp->b_validoff > bp->b_dirtyend) { |
| 369 | bp->b_validoff = bp->b_dirtyoff; |
| 370 | bp->b_validend = bp->b_dirtyend; |
| 371 | } else { |
| 372 | bp->b_validoff = min(bp->b_validoff, bp->b_dirtyoff); |
| 373 | bp->b_validend = max(bp->b_validend, bp->b_dirtyend); |
| 374 | } |
| 375 | |
| 376 | wrotedta = 1; |
| 377 | |
| 378 | /* |
| 379 | * Since this block is being modified, it must be written |
| 380 | * again and not just committed. |
| 381 | */ |
| 382 | |
| 383 | if (NFS_ISV3(vp)(((struct nfsmount *)(((vp)->v_mount)->mnt_data))->nm_flag & 0x00000200)) { |
| 384 | rw_enter_write(&np->n_commitlock); |
| 385 | if (bp->b_flags & B_NEEDCOMMIT0x00000002) { |
| 386 | bp->b_flags &= ~B_NEEDCOMMIT0x00000002; |
| 387 | nfs_del_tobecommitted_range(vp, bp); |
| 388 | } |
| 389 | nfs_del_committed_range(vp, bp); |
| 390 | rw_exit_write(&np->n_commitlock); |
| 391 | } else |
| 392 | bp->b_flags &= ~B_NEEDCOMMIT0x00000002; |
| 393 | |
| 394 | if (ioflag & IO_SYNC0x04) { |
| 395 | bp->b_proc = p; |
| 396 | error = VOP_BWRITE(bp); |
| 397 | if (error) |
| 398 | goto out; |
| 399 | } else if ((n + on) == biosize) { |
| 400 | bp->b_proc = NULL((void *)0); |
| 401 | bp->b_flags |= B_ASYNC0x00000004; |
| 402 | (void)nfs_writebp(bp, 0); |
| 403 | } else { |
| 404 | bdwrite(bp); |
| 405 | } |
| 406 | } while (uio->uio_resid > 0 && n > 0); |
| 407 | |
| 408 | /*out: XXX belongs here??? */ |
| 409 | if (wrotedta) |
| 410 | VN_KNOTE(vp, NOTE_WRITE | (extended ? NOTE_EXTEND : 0) |do { struct klist *__list = (&vp->v_selectinfo.si_note ); if (__list != ((void *)0)) knote(__list, (0x0002 | (extended ? 0x0004 : 0) | (truncated ? 0x0080 : 0))); } while (0) |
| 411 | (truncated ? NOTE_TRUNCATE : 0))do { struct klist *__list = (&vp->v_selectinfo.si_note ); if (__list != ((void *)0)) knote(__list, (0x0002 | (extended ? 0x0004 : 0) | (truncated ? 0x0080 : 0))); } while (0); |
| 412 | |
| 413 | out: |
| 414 | /* correct the result for writes clamped by vn_fsizechk() */ |
| 415 | uio->uio_resid += overrun; |
| 416 | |
| 417 | return (error); |
| 418 | } |
| 419 | |
| 420 | /* |
| 421 | * Get an nfs cache block. |
| 422 | * Allocate a new one if the block isn't currently in the cache |
| 423 | * and return the block marked busy. If the calling process is |
| 424 | * interrupted by a signal for an interruptible mount point, return |
| 425 | * NULL. |
| 426 | */ |
| 427 | struct buf * |
| 428 | nfs_getcacheblk(struct vnode *vp, daddr_t bn, int size, struct proc *p) |
| 429 | { |
| 430 | struct buf *bp; |
| 431 | struct nfsmount *nmp = VFSTONFS(vp->v_mount)((struct nfsmount *)((vp->v_mount)->mnt_data)); |
| 432 | |
| 433 | if (nmp->nm_flag & NFSMNT_INT0x00000040) { |
| 434 | bp = getblk(vp, bn, size, PCATCH0x100, INFSLP0xffffffffffffffffULL); |
| 435 | while (bp == NULL((void *)0)) { |
| 436 | if (nfs_sigintr(nmp, NULL((void *)0), p)) |
| 437 | return (NULL((void *)0)); |
| 438 | bp = getblk(vp, bn, size, 0, SEC_TO_NSEC(2)); |
| 439 | } |
| 440 | } else |
| 441 | bp = getblk(vp, bn, size, 0, INFSLP0xffffffffffffffffULL); |
| 442 | return (bp); |
| 443 | } |
| 444 | |
| 445 | /* |
| 446 | * Flush and invalidate all dirty buffers. If another process is already |
| 447 | * doing the flush, just wait for completion. |
| 448 | */ |
| 449 | int |
| 450 | nfs_vinvalbuf(struct vnode *vp, int flags, struct ucred *cred, struct proc *p) |
| 451 | { |
| 452 | struct nfsmount *nmp= VFSTONFS(vp->v_mount)((struct nfsmount *)((vp->v_mount)->mnt_data)); |
| 453 | struct nfsnode *np = VTONFS(vp)((struct nfsnode *)(vp)->v_data); |
| 454 | uint64_t stimeo; |
| 455 | int error, sintr; |
| 456 | |
| 457 | stimeo = INFSLP0xffffffffffffffffULL; |
| 458 | error = sintr = 0; |
| 459 | |
| 460 | if (ISSET(nmp->nm_flag, NFSMNT_INT)((nmp->nm_flag) & (0x00000040))) { |
| 461 | sintr = PCATCH0x100; |
| 462 | stimeo = SEC_TO_NSEC(2); |
| 463 | } |
| 464 | |
| 465 | /* First wait for any other process doing a flush to complete. */ |
| 466 | while (np->n_flag & NFLUSHINPROG0x0002) { |
| 467 | np->n_flag |= NFLUSHWANT0x0001; |
| 468 | error = tsleep_nsec(&np->n_flag, PRIBIO16|sintr, "nfsvinval", |
| 469 | stimeo); |
| 470 | if (error && sintr && nfs_sigintr(nmp, NULL((void *)0), p)) |
| 471 | return (EINTR4); |
| 472 | } |
| 473 | |
| 474 | /* Now, flush as required. */ |
| 475 | np->n_flag |= NFLUSHINPROG0x0002; |
| 476 | error = vinvalbuf(vp, flags, cred, p, sintr, INFSLP0xffffffffffffffffULL); |
| 477 | while (error) { |
| 478 | if (sintr && nfs_sigintr(nmp, NULL((void *)0), p)) { |
| 479 | np->n_flag &= ~NFLUSHINPROG0x0002; |
| 480 | if (np->n_flag & NFLUSHWANT0x0001) { |
| 481 | np->n_flag &= ~NFLUSHWANT0x0001; |
| 482 | wakeup(&np->n_flag); |
| 483 | } |
| 484 | return (EINTR4); |
| 485 | } |
| 486 | error = vinvalbuf(vp, flags, cred, p, 0, stimeo); |
| 487 | } |
| 488 | np->n_flag &= ~(NMODIFIED0x0004 | NFLUSHINPROG0x0002); |
| 489 | if (np->n_flag & NFLUSHWANT0x0001) { |
| 490 | np->n_flag &= ~NFLUSHWANT0x0001; |
| 491 | wakeup(&np->n_flag); |
| 492 | } |
| 493 | return (0); |
| 494 | } |
| 495 | |
| 496 | /* |
| 497 | * Initiate asynchronous I/O. Return an error if no nfsiods are available. |
| 498 | * This is mainly to avoid queueing async I/O requests when the nfsiods |
| 499 | * are all hung on a dead server. |
| 500 | */ |
| 501 | int |
| 502 | nfs_asyncio(struct buf *bp, int readahead) |
| 503 | { |
| 504 | if (nfs_numasync == 0) |
| 505 | goto out; |
| 506 | |
| 507 | while (nfs_bufqlen > nfs_bufqmax) |
| 508 | if (readahead) |
| 509 | goto out; |
| 510 | else |
| 511 | tsleep_nsec(&nfs_bufqlen, PRIBIO16, "nfs_bufq", INFSLP0xffffffffffffffffULL); |
| 512 | |
| 513 | if ((bp->b_flags & B_READ0x00008000) == 0) { |
| 514 | bp->b_flags |= B_WRITEINPROG0x00020000; |
| 515 | } |
| 516 | |
| 517 | TAILQ_INSERT_TAIL(&nfs_bufq, bp, b_freelist)do { (bp)->b_freelist.tqe_next = ((void *)0); (bp)->b_freelist .tqe_prev = (&nfs_bufq)->tqh_last; *(&nfs_bufq)-> tqh_last = (bp); (&nfs_bufq)->tqh_last = &(bp)-> b_freelist.tqe_next; } while (0); |
| 518 | nfs_bufqlen++; |
| 519 | |
| 520 | wakeup_one(&nfs_bufq)wakeup_n((&nfs_bufq), 1); |
| 521 | return (0); |
| 522 | |
| 523 | out: |
| 524 | nfsstats.forcedsync++; |
| 525 | return (EIO5); |
| 526 | } |
| 527 | |
| 528 | /* |
| 529 | * Do an I/O operation to/from a cache block. This may be called |
| 530 | * synchronously or from an nfsiod. |
| 531 | */ |
| 532 | int |
| 533 | nfs_doio(struct buf *bp, struct proc *p) |
| 534 | { |
| 535 | struct uio *uiop; |
| 536 | struct vnode *vp; |
| 537 | struct nfsnode *np; |
| 538 | struct nfsmount *nmp; |
| 539 | int s, error = 0, diff, len, iomode, must_commit = 0; |
| 540 | struct uio uio; |
| 541 | struct iovec io; |
| 542 | |
| 543 | vp = bp->b_vp; |
| 544 | np = VTONFS(vp)((struct nfsnode *)(vp)->v_data); |
| 545 | nmp = VFSTONFS(vp->v_mount)((struct nfsmount *)((vp->v_mount)->mnt_data)); |
Value stored to 'nmp' is never read | |
| 546 | uiop = &uio; |
| 547 | uiop->uio_iov = &io; |
| 548 | uiop->uio_iovcnt = 1; |
| 549 | uiop->uio_segflg = UIO_SYSSPACE; |
| 550 | uiop->uio_procp = p; |
| 551 | |
| 552 | /* |
| 553 | * Historically, paging was done with physio, but no more. |
| 554 | */ |
| 555 | if (bp->b_flags & B_PHYS0x00002000) { |
| 556 | io.iov_len = uiop->uio_resid = bp->b_bcount; |
| 557 | /* mapping was done by vmapbuf() */ |
| 558 | io.iov_base = bp->b_data; |
| 559 | uiop->uio_offset = ((off_t)bp->b_blkno) << DEV_BSHIFT9; |
| 560 | if (bp->b_flags & B_READ0x00008000) { |
| 561 | uiop->uio_rw = UIO_READ; |
| 562 | nfsstats.read_physios++; |
| 563 | error = nfs_readrpc(vp, uiop); |
| 564 | } else { |
| 565 | iomode = NFSV3WRITE_DATASYNC1; |
| 566 | uiop->uio_rw = UIO_WRITE; |
| 567 | nfsstats.write_physios++; |
| 568 | error = nfs_writerpc(vp, uiop, &iomode, &must_commit); |
| 569 | } |
| 570 | if (error) { |
| 571 | bp->b_flags |= B_ERROR0x00000400; |
| 572 | bp->b_error = error; |
| 573 | } |
| 574 | } else if (bp->b_flags & B_READ0x00008000) { |
| 575 | io.iov_len = uiop->uio_resid = bp->b_bcount; |
| 576 | io.iov_base = bp->b_data; |
| 577 | uiop->uio_rw = UIO_READ; |
| 578 | switch (vp->v_type) { |
| 579 | case VREG: |
| 580 | uiop->uio_offset = ((off_t)bp->b_blkno) << DEV_BSHIFT9; |
| 581 | nfsstats.read_bios++; |
| 582 | bcstats.pendingreads++; |
| 583 | bcstats.numreads++; |
| 584 | error = nfs_readrpc(vp, uiop); |
| 585 | if (!error) { |
| 586 | bp->b_validoff = 0; |
| 587 | if (uiop->uio_resid) { |
| 588 | /* |
| 589 | * If len > 0, there is a hole in the file and |
| 590 | * no writes after the hole have been pushed to |
| 591 | * the server yet. |
| 592 | * Just zero fill the rest of the valid area. |
| 593 | */ |
| 594 | diff = bp->b_bcount - uiop->uio_resid; |
| 595 | len = np->n_size - ((((off_t)bp->b_blkno) << DEV_BSHIFT9) |
| 596 | + diff); |
| 597 | if (len > 0) { |
| 598 | len = ulmin(len, uiop->uio_resid); |
| 599 | memset((char *)bp->b_data + diff, 0, len)__builtin_memset(((char *)bp->b_data + diff), (0), (len)); |
| 600 | bp->b_validend = diff + len; |
| 601 | } else |
| 602 | bp->b_validend = diff; |
| 603 | } else |
| 604 | bp->b_validend = bp->b_bcount; |
| 605 | } |
| 606 | if (p && (vp->v_flag & VTEXT0x0002) && |
| 607 | (timespeccmp(&np->n_mtime, &np->n_vattr.va_mtime, !=)(((&np->n_mtime)->tv_sec == (&np->n_vattr.va_mtime )->tv_sec) ? ((&np->n_mtime)->tv_nsec != (&np ->n_vattr.va_mtime)->tv_nsec) : ((&np->n_mtime)-> tv_sec != (&np->n_vattr.va_mtime)->tv_sec)))) { |
| 608 | uprintf("Process killed due to text file modification\n"); |
| 609 | psignal(p, SIGKILL9); |
| 610 | } |
| 611 | break; |
| 612 | case VLNK: |
| 613 | uiop->uio_offset = (off_t)0; |
| 614 | nfsstats.readlink_bios++; |
| 615 | bcstats.pendingreads++; |
| 616 | bcstats.numreads++; |
| 617 | error = nfs_readlinkrpc(vp, uiop, curproc({struct cpu_info *__ci; asm volatile("movq %%gs:%P1,%0" : "=r" (__ci) :"n" (__builtin_offsetof(struct cpu_info, ci_self))); __ci;})->ci_curproc->p_ucred); |
| 618 | break; |
| 619 | default: |
| 620 | panic("nfs_doio: type %x unexpected", vp->v_type); |
| 621 | break; |
| 622 | }; |
| 623 | if (error) { |
| 624 | bp->b_flags |= B_ERROR0x00000400; |
| 625 | bp->b_error = error; |
| 626 | } |
| 627 | } else { |
| 628 | io.iov_len = uiop->uio_resid = bp->b_dirtyend |
| 629 | - bp->b_dirtyoff; |
| 630 | uiop->uio_offset = ((off_t)bp->b_blkno) * DEV_BSIZE(1 << 9) |
| 631 | + bp->b_dirtyoff; |
| 632 | io.iov_base = (char *)bp->b_data + bp->b_dirtyoff; |
| 633 | uiop->uio_rw = UIO_WRITE; |
| 634 | nfsstats.write_bios++; |
| 635 | bcstats.pendingwrites++; |
| 636 | bcstats.numwrites++; |
| 637 | if ((bp->b_flags & (B_ASYNC0x00000004 | B_NEEDCOMMIT0x00000002 | B_NOCACHE0x00001000)) == B_ASYNC0x00000004) |
| 638 | iomode = NFSV3WRITE_UNSTABLE0; |
| 639 | else |
| 640 | iomode = NFSV3WRITE_FILESYNC2; |
| 641 | bp->b_flags |= B_WRITEINPROG0x00020000; |
| 642 | error = nfs_writerpc(vp, uiop, &iomode, &must_commit); |
| 643 | |
| 644 | rw_enter_write(&np->n_commitlock); |
| 645 | if (!error && iomode == NFSV3WRITE_UNSTABLE0) { |
| 646 | bp->b_flags |= B_NEEDCOMMIT0x00000002; |
| 647 | nfs_add_tobecommitted_range(vp, bp); |
| 648 | } else { |
| 649 | bp->b_flags &= ~B_NEEDCOMMIT0x00000002; |
| 650 | nfs_del_committed_range(vp, bp); |
| 651 | } |
| 652 | rw_exit_write(&np->n_commitlock); |
| 653 | |
| 654 | bp->b_flags &= ~B_WRITEINPROG0x00020000; |
| 655 | |
| 656 | /* |
| 657 | * For an interrupted write, the buffer is still valid and the |
| 658 | * write hasn't been pushed to the server yet, so we can't set |
| 659 | * B_ERROR and report the interruption by setting B_EINTR. For |
| 660 | * the B_ASYNC case, B_EINTR is not relevant, so the rpc attempt |
| 661 | * is essentially a noop. |
| 662 | * For the case of a V3 write rpc not being committed to stable |
| 663 | * storage, the block is still dirty and requires either a commit |
| 664 | * rpc or another write rpc with iomode == NFSV3WRITE_FILESYNC |
| 665 | * before the block is reused. This is indicated by setting the |
| 666 | * B_DELWRI and B_NEEDCOMMIT flags. |
| 667 | */ |
| 668 | if (error == EINTR4 || (!error && (bp->b_flags & B_NEEDCOMMIT0x00000002))) { |
| 669 | s = splbio()splraise(0x6); |
| 670 | buf_dirty(bp); |
| 671 | splx(s)spllower(s); |
| 672 | |
| 673 | if (!(bp->b_flags & B_ASYNC0x00000004) && error) |
| 674 | bp->b_flags |= B_EINTR0x00000200; |
| 675 | } else { |
| 676 | if (error) { |
| 677 | bp->b_flags |= B_ERROR0x00000400; |
| 678 | bp->b_error = np->n_error = error; |
| 679 | np->n_flag |= NWRITEERR0x0008; |
| 680 | } |
| 681 | bp->b_dirtyoff = bp->b_dirtyend = 0; |
| 682 | } |
| 683 | } |
| 684 | bp->b_resid = uiop->uio_resid; |
| 685 | if (must_commit) |
| 686 | nfs_clearcommit(vp->v_mount); |
| 687 | s = splbio()splraise(0x6); |
| 688 | biodone(bp); |
| 689 | splx(s)spllower(s); |
| 690 | return (error); |
| 691 | } |