xref: /csrg-svn/sys/kern/sysv_shm.c (revision 64833)
141490Smckusick /*
241490Smckusick  * Copyright (c) 1988 University of Utah.
363178Sbostic  * Copyright (c) 1990, 1993
463178Sbostic  *	The Regents of the University of California.  All rights reserved.
541490Smckusick  *
641490Smckusick  * This code is derived from software contributed to Berkeley by
741490Smckusick  * the Systems Programming Group of the University of Utah Computer
841490Smckusick  * Science Department. Originally from University of Wisconsin.
941490Smckusick  *
1064553Sbostic  * %sccs.include.proprietary.c%
1141490Smckusick  *
1254856Shibler  * from: Utah $Hdr: uipc_shm.c 1.11 92/04/23$
1341490Smckusick  *
14*64833Storek  *	@(#)sysv_shm.c	8.4 (Berkeley) 11/11/93
1541490Smckusick  */
1641490Smckusick 
1741490Smckusick /*
1841490Smckusick  * System V shared memory routines.
1943630Skarels  * TEMPORARY, until mmap is in place;
2043630Skarels  * needed now for HP-UX compatibility and X server (yech!).
2141490Smckusick  */
2241490Smckusick 
2341490Smckusick #ifdef SYSVSHM
2441490Smckusick 
2556517Sbostic #include <sys/param.h>
2656517Sbostic #include <sys/systm.h>
2756517Sbostic #include <sys/kernel.h>
2856517Sbostic #include <sys/proc.h>
2956517Sbostic #include <sys/shm.h>
3056517Sbostic #include <sys/malloc.h>
3156517Sbostic #include <sys/mman.h>
3264399Smckusick #include <sys/stat.h>
3341490Smckusick 
3456517Sbostic #include <vm/vm.h>
3556517Sbostic #include <vm/vm_kern.h>
3656517Sbostic #include <vm/vm_inherit.h>
3756517Sbostic #include <vm/vm_pager.h>
3856517Sbostic 
3941490Smckusick int	shmat(), shmctl(), shmdt(), shmget();
4041490Smckusick int	(*shmcalls[])() = { shmat, shmctl, shmdt, shmget };
4141490Smckusick int	shmtot = 0;
4241490Smckusick 
4345737Smckusick /*
4445737Smckusick  * Per process internal structure for managing segments.
4545737Smckusick  * Each process using shm will have an array of ``shmseg'' of these.
4645737Smckusick  */
4745737Smckusick struct	shmdesc {
4845737Smckusick 	vm_offset_t	shmd_uva;
4945737Smckusick 	int		shmd_id;
5045737Smckusick };
5141490Smckusick 
5245737Smckusick /*
5345737Smckusick  * Per segment internal structure (shm_handle).
5445737Smckusick  */
5545737Smckusick struct	shmhandle {
5645737Smckusick 	vm_offset_t	shmh_kva;
5745737Smckusick 	caddr_t		shmh_id;
5845737Smckusick };
5945737Smckusick 
6045737Smckusick vm_map_t shm_map;	/* address space for shared memory segments */
6145737Smckusick 
6241490Smckusick shminit()
6341490Smckusick {
6441490Smckusick 	register int i;
6545737Smckusick 	vm_offset_t whocares1, whocares2;
6641490Smckusick 
6745737Smckusick 	shm_map = kmem_suballoc(kernel_map, &whocares1, &whocares2,
6845737Smckusick 				shminfo.shmall * NBPG, FALSE);
6941490Smckusick 	if (shminfo.shmmni > SHMMMNI)
7041490Smckusick 		shminfo.shmmni = SHMMMNI;
7141490Smckusick 	for (i = 0; i < shminfo.shmmni; i++) {
7241490Smckusick 		shmsegs[i].shm_perm.mode = 0;
7341490Smckusick 		shmsegs[i].shm_perm.seq = 0;
7441490Smckusick 	}
7541490Smckusick }
7641490Smckusick 
7742961Smckusick /*
7842961Smckusick  * Entry point for all SHM calls
7942961Smckusick  */
8054932Storek struct shmsys_args {
8154932Storek 	u_int which;
8254932Storek };
8342961Smckusick shmsys(p, uap, retval)
8442961Smckusick 	struct proc *p;
8554932Storek 	struct shmsys_args *uap;
8642961Smckusick 	int *retval;
8741490Smckusick {
8841490Smckusick 
8942961Smckusick 	if (uap->which >= sizeof(shmcalls)/sizeof(shmcalls[0]))
9044405Skarels 		return (EINVAL);
9144405Skarels 	return ((*shmcalls[uap->which])(p, &uap[1], retval));
9241490Smckusick }
9341490Smckusick 
9442961Smckusick /*
9542961Smckusick  * Get a shared memory segment
9642961Smckusick  */
9754932Storek struct shmget_args {
9854932Storek 	key_t key;
9954932Storek 	int size;
10054932Storek 	int shmflg;
10154932Storek };
10242961Smckusick shmget(p, uap, retval)
10342961Smckusick 	struct proc *p;
10454932Storek 	register struct shmget_args *uap;
10542961Smckusick 	int *retval;
10642961Smckusick {
10741490Smckusick 	register struct shmid_ds *shp;
10847540Skarels 	register struct ucred *cred = p->p_ucred;
10941490Smckusick 	register int i;
11042961Smckusick 	int error, size, rval = 0;
11145737Smckusick 	register struct shmhandle *shmh;
11241490Smckusick 
11341490Smckusick 	/* look up the specified shm_id */
11441490Smckusick 	if (uap->key != IPC_PRIVATE) {
11541490Smckusick 		for (i = 0; i < shminfo.shmmni; i++)
11641490Smckusick 			if ((shmsegs[i].shm_perm.mode & SHM_ALLOC) &&
11741490Smckusick 			    shmsegs[i].shm_perm.key == uap->key) {
11841490Smckusick 				rval = i;
11941490Smckusick 				break;
12041490Smckusick 			}
12141490Smckusick 	} else
12241490Smckusick 		i = shminfo.shmmni;
12341490Smckusick 
12441490Smckusick 	/* create a new shared segment if necessary */
12541490Smckusick 	if (i == shminfo.shmmni) {
12642961Smckusick 		if ((uap->shmflg & IPC_CREAT) == 0)
12742961Smckusick 			return (ENOENT);
12842961Smckusick 		if (uap->size < shminfo.shmmin || uap->size > shminfo.shmmax)
12942961Smckusick 			return (EINVAL);
13041490Smckusick 		for (i = 0; i < shminfo.shmmni; i++)
13141490Smckusick 			if ((shmsegs[i].shm_perm.mode & SHM_ALLOC) == 0) {
13241490Smckusick 				rval = i;
13341490Smckusick 				break;
13441490Smckusick 			}
13542961Smckusick 		if (i == shminfo.shmmni)
13642961Smckusick 			return (ENOSPC);
13741490Smckusick 		size = clrnd(btoc(uap->size));
13842961Smckusick 		if (shmtot + size > shminfo.shmall)
13942961Smckusick 			return (ENOMEM);
14041490Smckusick 		shp = &shmsegs[rval];
14141490Smckusick 		/*
14241490Smckusick 		 * We need to do a couple of things to ensure consistency
14341490Smckusick 		 * in case we sleep in malloc().  We mark segment as
14441490Smckusick 		 * allocated so that other shmgets() will not allocate it.
14541490Smckusick 		 * We mark it as "destroyed" to insure that shmvalid() is
14641490Smckusick 		 * false making most operations fail (XXX).  We set the key,
14741490Smckusick 		 * so that other shmget()s will fail.
14841490Smckusick 		 */
14941490Smckusick 		shp->shm_perm.mode = SHM_ALLOC | SHM_DEST;
15041490Smckusick 		shp->shm_perm.key = uap->key;
15145737Smckusick 		shmh = (struct shmhandle *)
15245737Smckusick 			malloc(sizeof(struct shmhandle), M_SHM, M_WAITOK);
15345737Smckusick 		shmh->shmh_kva = 0;
15445737Smckusick 		shmh->shmh_id = (caddr_t)(0xc0000000|rval);	/* XXX */
15545737Smckusick 		error = vm_mmap(shm_map, &shmh->shmh_kva, ctob(size),
15658595Shibler 				VM_PROT_ALL, VM_PROT_ALL,
15758595Shibler 				MAP_ANON, shmh->shmh_id, 0);
15845737Smckusick 		if (error) {
15945737Smckusick 			free((caddr_t)shmh, M_SHM);
16041490Smckusick 			shp->shm_perm.mode = 0;
16145737Smckusick 			return(ENOMEM);
16241490Smckusick 		}
16345737Smckusick 		shp->shm_handle = (void *) shmh;
16441490Smckusick 		shmtot += size;
16542961Smckusick 		shp->shm_perm.cuid = shp->shm_perm.uid = cred->cr_uid;
16642961Smckusick 		shp->shm_perm.cgid = shp->shm_perm.gid = cred->cr_gid;
16764399Smckusick 		shp->shm_perm.mode = SHM_ALLOC | (uap->shmflg & ACCESSPERMS);
16841490Smckusick 		shp->shm_segsz = uap->size;
16942922Smckusick 		shp->shm_cpid = p->p_pid;
17041490Smckusick 		shp->shm_lpid = shp->shm_nattch = 0;
17141490Smckusick 		shp->shm_atime = shp->shm_dtime = 0;
17241490Smckusick 		shp->shm_ctime = time.tv_sec;
17341490Smckusick 	} else {
17441490Smckusick 		shp = &shmsegs[rval];
17541490Smckusick 		/* XXX: probably not the right thing to do */
17642961Smckusick 		if (shp->shm_perm.mode & SHM_DEST)
17742961Smckusick 			return (EBUSY);
17864399Smckusick 		if (error = ipcaccess(&shp->shm_perm, uap->shmflg & ACCESSPERMS,
17964399Smckusick 			    cred))
18042961Smckusick 			return (error);
18142961Smckusick 		if (uap->size && uap->size > shp->shm_segsz)
18242961Smckusick 			return (EINVAL);
18342961Smckusick 		if ((uap->shmflg&IPC_CREAT) && (uap->shmflg&IPC_EXCL))
18442961Smckusick 			return (EEXIST);
18541490Smckusick 	}
18642961Smckusick 	*retval = shp->shm_perm.seq * SHMMMNI + rval;
18743408Shibler 	return (0);
18841490Smckusick }
18941490Smckusick 
19042961Smckusick /*
19142961Smckusick  * Shared memory control
19242961Smckusick  */
19354932Storek struct shmctl_args {
19454932Storek 	int shmid;
19554932Storek 	int cmd;
19654932Storek 	caddr_t buf;
19754932Storek };
19842961Smckusick /* ARGSUSED */
19942961Smckusick shmctl(p, uap, retval)
20042961Smckusick 	struct proc *p;
20154932Storek 	register struct shmctl_args *uap;
20242961Smckusick 	int *retval;
20342961Smckusick {
20441490Smckusick 	register struct shmid_ds *shp;
20547540Skarels 	register struct ucred *cred = p->p_ucred;
20641490Smckusick 	struct shmid_ds sbuf;
20742961Smckusick 	int error;
20841490Smckusick 
20942961Smckusick 	if (error = shmvalid(uap->shmid))
21042961Smckusick 		return (error);
21141490Smckusick 	shp = &shmsegs[uap->shmid % SHMMMNI];
21241490Smckusick 	switch (uap->cmd) {
21341490Smckusick 	case IPC_STAT:
21443408Shibler 		if (error = ipcaccess(&shp->shm_perm, IPC_R, cred))
21542961Smckusick 			return (error);
21642961Smckusick 		return (copyout((caddr_t)shp, uap->buf, sizeof(*shp)));
21741490Smckusick 
21841490Smckusick 	case IPC_SET:
21942961Smckusick 		if (cred->cr_uid && cred->cr_uid != shp->shm_perm.uid &&
22042961Smckusick 		    cred->cr_uid != shp->shm_perm.cuid)
22142961Smckusick 			return (EPERM);
22242961Smckusick 		if (error = copyin(uap->buf, (caddr_t)&sbuf, sizeof sbuf))
22342961Smckusick 			return (error);
22442961Smckusick 		shp->shm_perm.uid = sbuf.shm_perm.uid;
22542961Smckusick 		shp->shm_perm.gid = sbuf.shm_perm.gid;
22664399Smckusick 		shp->shm_perm.mode = (shp->shm_perm.mode & ~ACCESSPERMS)
22764399Smckusick 			| (sbuf.shm_perm.mode & ACCESSPERMS);
22842961Smckusick 		shp->shm_ctime = time.tv_sec;
22941490Smckusick 		break;
23041490Smckusick 
23141490Smckusick 	case IPC_RMID:
23242961Smckusick 		if (cred->cr_uid && cred->cr_uid != shp->shm_perm.uid &&
23342961Smckusick 		    cred->cr_uid != shp->shm_perm.cuid)
23442961Smckusick 			return (EPERM);
23541490Smckusick 		/* set ctime? */
23641490Smckusick 		shp->shm_perm.key = IPC_PRIVATE;
23741490Smckusick 		shp->shm_perm.mode |= SHM_DEST;
23841490Smckusick 		if (shp->shm_nattch <= 0)
23941490Smckusick 			shmfree(shp);
24041490Smckusick 		break;
24141490Smckusick 
24241490Smckusick 	default:
24342961Smckusick 		return (EINVAL);
24441490Smckusick 	}
24542961Smckusick 	return (0);
24641490Smckusick }
24741490Smckusick 
24842961Smckusick /*
24942961Smckusick  * Attach to shared memory segment.
25042961Smckusick  */
25154932Storek struct shmat_args {
25254932Storek 	int	shmid;
25354932Storek 	caddr_t	shmaddr;
25454932Storek 	int	shmflg;
25554932Storek };
25642961Smckusick shmat(p, uap, retval)
25742961Smckusick 	struct proc *p;
25854932Storek 	register struct shmat_args *uap;
25942961Smckusick 	int *retval;
26042961Smckusick {
26141490Smckusick 	register struct shmid_ds *shp;
26241490Smckusick 	register int size;
26341490Smckusick 	caddr_t uva;
26445737Smckusick 	int error;
26545737Smckusick 	int flags;
26645737Smckusick 	vm_prot_t prot;
26745737Smckusick 	struct shmdesc *shmd;
26841490Smckusick 
26945737Smckusick 	/*
27045737Smckusick 	 * Allocate descriptors now (before validity check)
27145737Smckusick 	 * in case malloc() blocks.
27245737Smckusick 	 */
27347540Skarels 	shmd = (struct shmdesc *)p->p_vmspace->vm_shm;
27445737Smckusick 	size = shminfo.shmseg * sizeof(struct shmdesc);
27545737Smckusick 	if (shmd == NULL) {
27645737Smckusick 		shmd = (struct shmdesc *)malloc(size, M_SHM, M_WAITOK);
27745737Smckusick 		bzero((caddr_t)shmd, size);
27847540Skarels 		p->p_vmspace->vm_shm = (caddr_t)shmd;
27945737Smckusick 	}
28042961Smckusick 	if (error = shmvalid(uap->shmid))
28142961Smckusick 		return (error);
28241490Smckusick 	shp = &shmsegs[uap->shmid % SHMMMNI];
28341490Smckusick 	if (shp->shm_handle == NULL)
28442349Smckusick 		panic("shmat NULL handle");
28543408Shibler 	if (error = ipcaccess(&shp->shm_perm,
28647540Skarels 	    (uap->shmflg&SHM_RDONLY) ? IPC_R : IPC_R|IPC_W, p->p_ucred))
28742961Smckusick 		return (error);
28841490Smckusick 	uva = uap->shmaddr;
28941490Smckusick 	if (uva && ((int)uva & (SHMLBA-1))) {
29041490Smckusick 		if (uap->shmflg & SHM_RND)
29141490Smckusick 			uva = (caddr_t) ((int)uva & ~(SHMLBA-1));
29242961Smckusick 		else
29342961Smckusick 			return (EINVAL);
29441490Smckusick 	}
29541490Smckusick 	/*
29641490Smckusick 	 * Make sure user doesn't use more than their fair share
29741490Smckusick 	 */
29845737Smckusick 	for (size = 0; size < shminfo.shmseg; size++) {
29945737Smckusick 		if (shmd->shmd_uva == 0)
30045737Smckusick 			break;
30145737Smckusick 		shmd++;
30245737Smckusick 	}
30342961Smckusick 	if (size >= shminfo.shmseg)
30442961Smckusick 		return (EMFILE);
30541490Smckusick 	size = ctob(clrnd(btoc(shp->shm_segsz)));
30645737Smckusick 	prot = VM_PROT_READ;
30745737Smckusick 	if ((uap->shmflg & SHM_RDONLY) == 0)
30845737Smckusick 		prot |= VM_PROT_WRITE;
30945737Smckusick 	flags = MAP_ANON|MAP_SHARED;
31045737Smckusick 	if (uva)
31145737Smckusick 		flags |= MAP_FIXED;
31245737Smckusick 	else
31345737Smckusick 		uva = (caddr_t)0x1000000;	/* XXX */
31453313Smckusick 	error = vm_mmap(&p->p_vmspace->vm_map, (vm_offset_t *)&uva,
31558595Shibler 			(vm_size_t)size, prot, VM_PROT_ALL, flags,
31658595Shibler 			((struct shmhandle *)shp->shm_handle)->shmh_id, 0);
31742961Smckusick 	if (error)
31845737Smckusick 		return(error);
31945737Smckusick 	shmd->shmd_uva = (vm_offset_t)uva;
32045737Smckusick 	shmd->shmd_id = uap->shmid;
32141490Smckusick 	/*
32241490Smckusick 	 * Fill in the remaining fields
32341490Smckusick 	 */
32442922Smckusick 	shp->shm_lpid = p->p_pid;
32541490Smckusick 	shp->shm_atime = time.tv_sec;
32641490Smckusick 	shp->shm_nattch++;
32742961Smckusick 	*retval = (int) uva;
32843408Shibler 	return (0);
32941490Smckusick }
33041490Smckusick 
33142961Smckusick /*
33242961Smckusick  * Detach from shared memory segment.
33342961Smckusick  */
33454932Storek struct shmdt_args {
33554932Storek 	caddr_t	shmaddr;
33654932Storek };
33742961Smckusick /* ARGSUSED */
33842961Smckusick shmdt(p, uap, retval)
33942961Smckusick 	struct proc *p;
34054932Storek 	struct shmdt_args *uap;
34142961Smckusick 	int *retval;
34241490Smckusick {
34345737Smckusick 	register struct shmdesc *shmd;
34445737Smckusick 	register int i;
34541490Smckusick 
34647540Skarels 	shmd = (struct shmdesc *)p->p_vmspace->vm_shm;
34745737Smckusick 	for (i = 0; i < shminfo.shmseg; i++, shmd++)
34845737Smckusick 		if (shmd->shmd_uva &&
34945737Smckusick 		    shmd->shmd_uva == (vm_offset_t)uap->shmaddr)
35041490Smckusick 			break;
35145737Smckusick 	if (i == shminfo.shmseg)
352*64833Storek 		return (EINVAL);
35345737Smckusick 	shmufree(p, shmd);
35445737Smckusick 	shmsegs[shmd->shmd_id % SHMMMNI].shm_lpid = p->p_pid;
355*64833Storek 	return (0);
35641490Smckusick }
35741490Smckusick 
35847540Skarels shmfork(p1, p2, isvfork)
35947540Skarels 	struct proc *p1, *p2;
36045737Smckusick 	int isvfork;
36141490Smckusick {
36245737Smckusick 	register struct shmdesc *shmd;
36345737Smckusick 	register int size;
36441490Smckusick 
36545737Smckusick 	/*
36645737Smckusick 	 * Copy parents descriptive information
36745737Smckusick 	 */
36845737Smckusick 	size = shminfo.shmseg * sizeof(struct shmdesc);
36945737Smckusick 	shmd = (struct shmdesc *)malloc(size, M_SHM, M_WAITOK);
37047540Skarels 	bcopy((caddr_t)p1->p_vmspace->vm_shm, (caddr_t)shmd, size);
37147540Skarels 	p2->p_vmspace->vm_shm = (caddr_t)shmd;
37245737Smckusick 	/*
37345737Smckusick 	 * Increment reference counts
37445737Smckusick 	 */
37545737Smckusick 	for (size = 0; size < shminfo.shmseg; size++, shmd++)
37645737Smckusick 		if (shmd->shmd_uva)
37745737Smckusick 			shmsegs[shmd->shmd_id % SHMMMNI].shm_nattch++;
37841490Smckusick }
37941490Smckusick 
38045737Smckusick shmexit(p)
38145737Smckusick 	struct proc *p;
38241490Smckusick {
38345737Smckusick 	register struct shmdesc *shmd;
38445737Smckusick 	register int i;
38541490Smckusick 
38647540Skarels 	shmd = (struct shmdesc *)p->p_vmspace->vm_shm;
38745737Smckusick 	for (i = 0; i < shminfo.shmseg; i++, shmd++)
38845737Smckusick 		if (shmd->shmd_uva)
38945737Smckusick 			shmufree(p, shmd);
39047540Skarels 	free((caddr_t)p->p_vmspace->vm_shm, M_SHM);
39147540Skarels 	p->p_vmspace->vm_shm = NULL;
39241490Smckusick }
39341490Smckusick 
39441490Smckusick shmvalid(id)
39541490Smckusick 	register int id;
39641490Smckusick {
39741490Smckusick 	register struct shmid_ds *shp;
39841490Smckusick 
39941490Smckusick 	if (id < 0 || (id % SHMMMNI) >= shminfo.shmmni)
40042961Smckusick 		return(EINVAL);
40141490Smckusick 	shp = &shmsegs[id % SHMMMNI];
40241490Smckusick 	if (shp->shm_perm.seq == (id / SHMMMNI) &&
40341490Smckusick 	    (shp->shm_perm.mode & (SHM_ALLOC|SHM_DEST)) == SHM_ALLOC)
40442961Smckusick 		return(0);
40542961Smckusick 	return(EINVAL);
40641490Smckusick }
40741490Smckusick 
40841490Smckusick /*
40941490Smckusick  * Free user resources associated with a shared memory segment
41041490Smckusick  */
41145737Smckusick shmufree(p, shmd)
41242922Smckusick 	struct proc *p;
41345737Smckusick 	struct shmdesc *shmd;
41441490Smckusick {
41541490Smckusick 	register struct shmid_ds *shp;
41641490Smckusick 
41745737Smckusick 	shp = &shmsegs[shmd->shmd_id % SHMMMNI];
41849710Shibler 	(void) vm_deallocate(&p->p_vmspace->vm_map, shmd->shmd_uva,
41945737Smckusick 			     ctob(clrnd(btoc(shp->shm_segsz))));
42045737Smckusick 	shmd->shmd_id = 0;
42145737Smckusick 	shmd->shmd_uva = 0;
42241490Smckusick 	shp->shm_dtime = time.tv_sec;
42341490Smckusick 	if (--shp->shm_nattch <= 0 && (shp->shm_perm.mode & SHM_DEST))
42441490Smckusick 		shmfree(shp);
42541490Smckusick }
42641490Smckusick 
42741490Smckusick /*
42841490Smckusick  * Deallocate resources associated with a shared memory segment
42941490Smckusick  */
43041490Smckusick shmfree(shp)
43141490Smckusick 	register struct shmid_ds *shp;
43241490Smckusick {
43341490Smckusick 
43441490Smckusick 	if (shp->shm_handle == NULL)
43541490Smckusick 		panic("shmfree");
43645737Smckusick 	/*
43745737Smckusick 	 * Lose our lingering object reference by deallocating space
43845737Smckusick 	 * in kernel.  Pager will also be deallocated as a side-effect.
43945737Smckusick 	 */
44045737Smckusick 	vm_deallocate(shm_map,
44145737Smckusick 		      ((struct shmhandle *)shp->shm_handle)->shmh_kva,
44249668Shibler 		      ctob(clrnd(btoc(shp->shm_segsz))));
44345737Smckusick 	free((caddr_t)shp->shm_handle, M_SHM);
44441490Smckusick 	shp->shm_handle = NULL;
44541490Smckusick 	shmtot -= clrnd(btoc(shp->shm_segsz));
44641490Smckusick 	shp->shm_perm.mode = 0;
44741490Smckusick 	/*
44841490Smckusick 	 * Increment the sequence number to ensure that outstanding
44941490Smckusick 	 * shmids for this segment will be invalid in the event that
45041490Smckusick 	 * the segment is reallocated.  Note that shmids must be
45141490Smckusick 	 * positive as decreed by SVID.
45241490Smckusick 	 */
45341490Smckusick 	shp->shm_perm.seq++;
45441490Smckusick 	if ((int)(shp->shm_perm.seq * SHMMMNI) < 0)
45541490Smckusick 		shp->shm_perm.seq = 0;
45641490Smckusick }
45741490Smckusick 
45841490Smckusick /*
45941490Smckusick  * XXX This routine would be common to all sysV style IPC
46041490Smckusick  *     (if the others were implemented).
46141490Smckusick  */
46242961Smckusick ipcaccess(ipc, mode, cred)
46341490Smckusick 	register struct ipc_perm *ipc;
46442961Smckusick 	int mode;
46542961Smckusick 	register struct ucred *cred;
46641490Smckusick {
46741490Smckusick 	register int m;
46841490Smckusick 
46942961Smckusick 	if (cred->cr_uid == 0)
47041490Smckusick 		return(0);
47141490Smckusick 	/*
47241490Smckusick 	 * Access check is based on only one of owner, group, public.
47341490Smckusick 	 * If not owner, then check group.
47441490Smckusick 	 * If not a member of the group, then check public access.
47541490Smckusick 	 */
47641490Smckusick 	mode &= 0700;
47741490Smckusick 	m = ipc->mode;
47842961Smckusick 	if (cred->cr_uid != ipc->uid && cred->cr_uid != ipc->cuid) {
47941490Smckusick 		m <<= 3;
48042961Smckusick 		if (!groupmember(ipc->gid, cred) &&
48142961Smckusick 		    !groupmember(ipc->cgid, cred))
48241490Smckusick 			m <<= 3;
48341490Smckusick 	}
48441490Smckusick 	if ((mode&m) == mode)
48542961Smckusick 		return (0);
48642961Smckusick 	return (EACCES);
48741490Smckusick }
48841490Smckusick #endif /* SYSVSHM */
489