]> git.saurik.com Git - apple/xnu.git/blame_incremental - bsd/kern/sysv_shm.c
xnu-517.12.7.tar.gz
[apple/xnu.git] / bsd / kern / sysv_shm.c
... / ...
CommitLineData
1/*
2 * Copyright (c) 2000-2004 Apple Computer, Inc. All rights reserved.
3 *
4 * @APPLE_LICENSE_HEADER_START@
5 *
6 * The contents of this file constitute Original Code as defined in and
7 * are subject to the Apple Public Source License Version 1.1 (the
8 * "License"). You may not use this file except in compliance with the
9 * License. Please obtain a copy of the License at
10 * http://www.apple.com/publicsource and read it before using this file.
11 *
12 * This Original Code and all software distributed under the License are
13 * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
14 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
15 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the
17 * License for the specific language governing rights and limitations
18 * under the License.
19 *
20 * @APPLE_LICENSE_HEADER_END@
21 */
22/* $NetBSD: sysv_shm.c,v 1.23 1994/07/04 23:25:12 glass Exp $ */
23
24/*
25 * Copyright (c) 1994 Adam Glass and Charles Hannum. All rights reserved.
26 *
27 * Redistribution and use in source and binary forms, with or without
28 * modification, are permitted provided that the following conditions
29 * are met:
30 * 1. Redistributions of source code must retain the above copyright
31 * notice, this list of conditions and the following disclaimer.
32 * 2. Redistributions in binary form must reproduce the above copyright
33 * notice, this list of conditions and the following disclaimer in the
34 * documentation and/or other materials provided with the distribution.
35 * 3. All advertising materials mentioning features or use of this software
36 * must display the following acknowledgement:
37 * This product includes software developed by Adam Glass and Charles
38 * Hannum.
39 * 4. The names of the authors may not be used to endorse or promote products
40 * derived from this software without specific prior written permission.
41 *
42 * THIS SOFTWARE IS PROVIDED BY THE AUTHORS ``AS IS'' AND ANY EXPRESS OR
43 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
44 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
45 * IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY DIRECT, INDIRECT,
46 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
47 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
48 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
49 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
50 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
51 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
52 */
53
54
55#include <sys/appleapiopts.h>
56#include <sys/param.h>
57#include <sys/systm.h>
58#include <sys/kernel.h>
59#include <sys/shm.h>
60#include <sys/proc.h>
61#include <sys/malloc.h>
62#include <sys/mman.h>
63#include <sys/stat.h>
64#include <sys/sysctl.h>
65
66#include <bsm/audit_kernel.h>
67
68#include <mach/mach_types.h>
69#include <mach/vm_inherit.h>
70#include <vm/vm_map.h>
71
72struct shmat_args;
73extern int shmat __P((struct proc *p, struct shmat_args *uap, int *retval));
74struct shmctl_args;
75extern int shmctl __P((struct proc *p, struct shmctl_args *uap, int *retval));
76struct shmdt_args;
77extern int shmdt __P((struct proc *p, struct shmdt_args *uap, int *retval));
78struct shmget_args;
79extern int shmget __P((struct proc *p, struct shmget_args *uap, int *retval));
80
81#if 0
82static void shminit __P((void *));
83SYSINIT(sysv_shm, SI_SUB_SYSV_SHM, SI_ORDER_FIRST, shminit, NULL)
84#endif 0
85
86struct oshmctl_args;
87static int oshmctl __P((struct proc *p, struct oshmctl_args *uap, int * retval));
88static int shmget_allocate_segment __P((struct proc *p, struct shmget_args *uap, int mode, int * retval));
89static int shmget_existing __P((struct proc *p, struct shmget_args *uap, int mode, int segnum, int * retval));
90
91typedef int sy_call_t __P((struct proc *, void *, int *));
92
93/* XXX casting to (sy_call_t *) is bogus, as usual. */
94static sy_call_t *shmcalls[] = {
95 (sy_call_t *)shmat, (sy_call_t *)oshmctl,
96 (sy_call_t *)shmdt, (sy_call_t *)shmget,
97 (sy_call_t *)shmctl
98};
99
100#define SHMSEG_FREE 0x0200
101#define SHMSEG_REMOVED 0x0400
102#define SHMSEG_ALLOCATED 0x0800
103#define SHMSEG_WANTED 0x1000
104
105static int shm_last_free, shm_nused, shm_committed;
106struct shmid_ds *shmsegs;
107static int shm_inited = 0;
108
109struct shm_handle {
110 /* vm_offset_t kva; */
111 void * shm_object;
112};
113
114struct shmmap_state {
115 vm_offset_t va;
116 int shmid;
117};
118
119static void shm_deallocate_segment __P((struct shmid_ds *));
120static int shm_find_segment_by_key __P((key_t));
121static struct shmid_ds *shm_find_segment_by_shmid __P((int));
122static int shm_delete_mapping __P((struct proc *, struct shmmap_state *, int));
123
124#ifdef __APPLE_API_PRIVATE
125struct shminfo shminfo = {
126 -1, /* SHMMAX 4096 *1024 */
127 -1, /* SHMMIN = 1 */
128 -1, /* SHMMNI = 1 */
129 -1, /* SHMSEG = 8 */
130 -1 /* SHMALL = 1024 */
131};
132#endif /* __APPLE_API_PRIVATE */
133
134static int
135shm_find_segment_by_key(key)
136 key_t key;
137{
138 int i;
139
140 for (i = 0; i < shminfo.shmmni; i++)
141 if ((shmsegs[i].shm_perm.mode & SHMSEG_ALLOCATED) &&
142 shmsegs[i].shm_perm.key == key)
143 return i;
144 return -1;
145}
146
147static struct shmid_ds *
148shm_find_segment_by_shmid(shmid)
149 int shmid;
150{
151 int segnum;
152 struct shmid_ds *shmseg;
153
154 segnum = IPCID_TO_IX(shmid);
155 if (segnum < 0 || segnum >= shminfo.shmmni)
156 return NULL;
157 shmseg = &shmsegs[segnum];
158 if ((shmseg->shm_perm.mode & (SHMSEG_ALLOCATED | SHMSEG_REMOVED))
159 != SHMSEG_ALLOCATED ||
160 shmseg->shm_perm.seq != IPCID_TO_SEQ(shmid))
161 return NULL;
162 return shmseg;
163}
164
165static void
166shm_deallocate_segment(shmseg)
167 struct shmid_ds *shmseg;
168{
169 struct shm_handle *shm_handle;
170 struct shmmap_state *shmmap_s=NULL;
171 size_t size;
172 char * ptr;
173
174 shm_handle = shmseg->shm_internal;
175 size = round_page_32(shmseg->shm_segsz);
176 mach_destroy_memory_entry(shm_handle->shm_object);
177 FREE((caddr_t)shm_handle, M_SHM);
178 shmseg->shm_internal = NULL;
179 shm_committed -= btoc(size);
180 shm_nused--;
181 shmseg->shm_perm.mode = SHMSEG_FREE;
182}
183
184static int
185shm_delete_mapping(p, shmmap_s, deallocate)
186 struct proc *p;
187 struct shmmap_state *shmmap_s;
188 int deallocate;
189{
190 struct shmid_ds *shmseg;
191 int segnum, result;
192 size_t size;
193
194 segnum = IPCID_TO_IX(shmmap_s->shmid);
195 shmseg = &shmsegs[segnum];
196 size = round_page_32(shmseg->shm_segsz);
197 if (deallocate) {
198 result = vm_deallocate(current_map(), shmmap_s->va, size);
199 if (result != KERN_SUCCESS)
200 return EINVAL;
201 }
202 shmmap_s->shmid = -1;
203 shmseg->shm_dtime = time_second;
204 if ((--shmseg->shm_nattch <= 0) &&
205 (shmseg->shm_perm.mode & SHMSEG_REMOVED)) {
206 shm_deallocate_segment(shmseg);
207 shm_last_free = segnum;
208 }
209 return 0;
210}
211
212struct shmdt_args {
213 void *shmaddr;
214};
215
216int
217shmdt(p, uap, retval)
218 struct proc *p;
219 struct shmdt_args *uap;
220 register_t *retval;
221{
222 struct shmmap_state *shmmap_s;
223 int i;
224
225 AUDIT_ARG(svipc_addr, uap->shmaddr);
226 if (!shm_inited)
227 return(EINVAL);
228 shmmap_s = (struct shmmap_state *)p->vm_shm;
229 if (shmmap_s == NULL)
230 return EINVAL;
231 for (i = 0; i < shminfo.shmseg; i++, shmmap_s++)
232 if (shmmap_s->shmid != -1 &&
233 shmmap_s->va == (vm_offset_t)uap->shmaddr)
234 break;
235 if (i == shminfo.shmseg)
236 return EINVAL;
237 return shm_delete_mapping(p, shmmap_s, 1);
238}
239
240#ifndef _SYS_SYSPROTO_H_
241struct shmat_args {
242 int shmid;
243 void *shmaddr;
244 int shmflg;
245};
246#endif
247
248int
249shmat(p, uap, retval)
250 struct proc *p;
251 struct shmat_args *uap;
252 register_t *retval;
253{
254 int error, i, flags;
255 struct ucred *cred = p->p_ucred;
256 struct shmid_ds *shmseg;
257 struct shmmap_state *shmmap_s = NULL;
258 struct shm_handle *shm_handle;
259 vm_offset_t attach_va;
260 vm_prot_t prot;
261 vm_size_t size;
262 kern_return_t rv;
263
264 AUDIT_ARG(svipc_id, uap->shmid);
265 AUDIT_ARG(svipc_addr, uap->shmaddr);
266 if (!shm_inited)
267 return(EINVAL);
268 shmmap_s = (struct shmmap_state *)p->vm_shm;
269 if (shmmap_s == NULL) {
270 size = shminfo.shmseg * sizeof(struct shmmap_state);
271 shmmap_s = (struct shmmap_state *)_MALLOC(size, M_SHM, M_WAITOK);
272 for (i = 0; i < shminfo.shmseg; i++)
273 shmmap_s[i].shmid = -1;
274 p->vm_shm = (caddr_t)shmmap_s;
275 }
276 shmseg = shm_find_segment_by_shmid(uap->shmid);
277 if (shmseg == NULL)
278 return EINVAL;
279
280 AUDIT_ARG(svipc_perm, &shmseg->shm_perm);
281 error = ipcperm(cred, &shmseg->shm_perm,
282 (uap->shmflg & SHM_RDONLY) ? IPC_R : IPC_R|IPC_W);
283 if (error)
284 return error;
285 for (i = 0; i < shminfo.shmseg; i++) {
286 if (shmmap_s->shmid == -1)
287 break;
288 shmmap_s++;
289 }
290 if (i >= shminfo.shmseg)
291 return EMFILE;
292 size = round_page_32(shmseg->shm_segsz);
293 prot = VM_PROT_READ;
294 if ((uap->shmflg & SHM_RDONLY) == 0)
295 prot |= VM_PROT_WRITE;
296 flags = MAP_ANON | MAP_SHARED;
297 if (uap->shmaddr) {
298 flags |= MAP_FIXED;
299 if (uap->shmflg & SHM_RND)
300 attach_va = (vm_offset_t)uap->shmaddr & ~(SHMLBA-1);
301 else if (((vm_offset_t)uap->shmaddr & (SHMLBA-1)) == 0)
302 attach_va = (vm_offset_t)uap->shmaddr;
303 else
304 return EINVAL;
305 } else {
306 attach_va = round_page_32((unsigned int)uap->shmaddr);
307 }
308
309 shm_handle = shmseg->shm_internal;
310 rv = vm_map(current_map(), &attach_va, size, 0, (flags & MAP_FIXED)? FALSE: TRUE,
311 shm_handle->shm_object, 0, FALSE, prot, prot, VM_INHERIT_DEFAULT);
312 if (rv != KERN_SUCCESS)
313 goto out;
314 rv = vm_inherit(current_map(), attach_va, size,
315 VM_INHERIT_SHARE);
316 if (rv != KERN_SUCCESS) {
317 (void) vm_deallocate(current_map(), attach_va, size);
318 goto out;
319 }
320
321 shmmap_s->va = attach_va;
322 shmmap_s->shmid = uap->shmid;
323 shmseg->shm_lpid = p->p_pid;
324 shmseg->shm_atime = time_second;
325 shmseg->shm_nattch++;
326 *retval = attach_va;
327 return( 0);
328out:
329 switch (rv) {
330 case KERN_INVALID_ADDRESS:
331 case KERN_NO_SPACE:
332 return (ENOMEM);
333 case KERN_PROTECTION_FAILURE:
334 return (EACCES);
335 default:
336 return (EINVAL);
337 }
338
339}
340
341struct oshmid_ds {
342 struct ipc_perm shm_perm; /* operation perms */
343 int shm_segsz; /* size of segment (bytes) */
344 ushort shm_cpid; /* pid, creator */
345 ushort shm_lpid; /* pid, last operation */
346 short shm_nattch; /* no. of current attaches */
347 time_t shm_atime; /* last attach time */
348 time_t shm_dtime; /* last detach time */
349 time_t shm_ctime; /* last change time */
350 void *shm_handle; /* internal handle for shm segment */
351};
352
353struct oshmctl_args {
354 int shmid;
355 int cmd;
356 struct oshmid_ds *ubuf;
357};
358
359static int
360oshmctl(p, uap, retval)
361 struct proc *p;
362 struct oshmctl_args *uap;
363 register_t *retval;
364{
365#ifdef COMPAT_43
366 int error;
367 struct ucred *cred = p->p_ucred;
368 struct shmid_ds *shmseg;
369 struct oshmid_ds outbuf;
370
371 if (!shm_inited)
372 return(EINVAL);
373 shmseg = shm_find_segment_by_shmid(uap->shmid);
374 if (shmseg == NULL)
375 return EINVAL;
376 switch (uap->cmd) {
377 case IPC_STAT:
378 error = ipcperm(cred, &shmseg->shm_perm, IPC_R);
379 if (error)
380 return error;
381 outbuf.shm_perm = shmseg->shm_perm;
382 outbuf.shm_segsz = shmseg->shm_segsz;
383 outbuf.shm_cpid = shmseg->shm_cpid;
384 outbuf.shm_lpid = shmseg->shm_lpid;
385 outbuf.shm_nattch = shmseg->shm_nattch;
386 outbuf.shm_atime = shmseg->shm_atime;
387 outbuf.shm_dtime = shmseg->shm_dtime;
388 outbuf.shm_ctime = shmseg->shm_ctime;
389 outbuf.shm_handle = shmseg->shm_internal;
390 error = copyout((caddr_t)&outbuf, uap->ubuf, sizeof(outbuf));
391 if (error)
392 return error;
393 break;
394 default:
395 /* XXX casting to (sy_call_t *) is bogus, as usual. */
396 return ((sy_call_t *)shmctl)(p, uap, retval);
397 }
398 return 0;
399#else
400 return EINVAL;
401#endif
402}
403
404#ifndef _SYS_SYSPROTO_H_
405struct shmctl_args {
406 int shmid;
407 int cmd;
408 struct shmid_ds *buf;
409};
410#endif
411
412int
413shmctl(p, uap, retval)
414 struct proc *p;
415 struct shmctl_args *uap;
416 register_t *retval;
417{
418 int error;
419 struct ucred *cred = p->p_ucred;
420 struct shmid_ds inbuf;
421 struct shmid_ds *shmseg;
422
423 AUDIT_ARG(svipc_cmd, uap->cmd);
424 AUDIT_ARG(svipc_id, uap->shmid);
425 if (!shm_inited)
426 return(EINVAL);
427 shmseg = shm_find_segment_by_shmid(uap->shmid);
428 if (shmseg == NULL)
429 return EINVAL;
430 /* XXAUDIT: This is the perms BEFORE any change by this call. This
431 * may not be what is desired.
432 */
433 AUDIT_ARG(svipc_perm, &shmseg->shm_perm);
434
435 switch (uap->cmd) {
436 case IPC_STAT:
437 error = ipcperm(cred, &shmseg->shm_perm, IPC_R);
438 if (error)
439 return error;
440 error = copyout((caddr_t)shmseg, uap->buf, sizeof(inbuf));
441 if (error)
442 return error;
443 break;
444 case IPC_SET:
445 error = ipcperm(cred, &shmseg->shm_perm, IPC_M);
446 if (error)
447 return error;
448 error = copyin(uap->buf, (caddr_t)&inbuf, sizeof(inbuf));
449 if (error)
450 return error;
451 shmseg->shm_perm.uid = inbuf.shm_perm.uid;
452 shmseg->shm_perm.gid = inbuf.shm_perm.gid;
453 shmseg->shm_perm.mode =
454 (shmseg->shm_perm.mode & ~ACCESSPERMS) |
455 (inbuf.shm_perm.mode & ACCESSPERMS);
456 shmseg->shm_ctime = time_second;
457 break;
458 case IPC_RMID:
459 error = ipcperm(cred, &shmseg->shm_perm, IPC_M);
460 if (error)
461 return error;
462 shmseg->shm_perm.key = IPC_PRIVATE;
463 shmseg->shm_perm.mode |= SHMSEG_REMOVED;
464 if (shmseg->shm_nattch <= 0) {
465 shm_deallocate_segment(shmseg);
466 shm_last_free = IPCID_TO_IX(uap->shmid);
467 }
468 break;
469#if 0
470 case SHM_LOCK:
471 case SHM_UNLOCK:
472#endif
473 default:
474 return EINVAL;
475 }
476 return 0;
477}
478
479#ifndef _SYS_SYSPROTO_H_
480struct shmget_args {
481 key_t key;
482 size_t size;
483 int shmflg;
484};
485#endif
486
487static int
488shmget_existing(p, uap, mode, segnum, retval)
489 struct proc *p;
490 struct shmget_args *uap;
491 int mode;
492 int segnum;
493 int *retval;
494{
495 struct shmid_ds *shmseg;
496 struct ucred *cred = p->p_ucred;
497 int error;
498
499 shmseg = &shmsegs[segnum];
500 if (shmseg->shm_perm.mode & SHMSEG_REMOVED) {
501 /*
502 * This segment is in the process of being allocated. Wait
503 * until it's done, and look the key up again (in case the
504 * allocation failed or it was freed).
505 */
506 shmseg->shm_perm.mode |= SHMSEG_WANTED;
507 error = tsleep((caddr_t)shmseg, PLOCK | PCATCH, "shmget", 0);
508 if (error)
509 return error;
510 return EAGAIN;
511 }
512 error = ipcperm(cred, &shmseg->shm_perm, mode);
513 if (error)
514 return error;
515 if (uap->size && uap->size > shmseg->shm_segsz)
516 return EINVAL;
517 if ((uap->shmflg & (IPC_CREAT | IPC_EXCL)) == (IPC_CREAT | IPC_EXCL))
518 return EEXIST;
519 *retval = IXSEQ_TO_IPCID(segnum, shmseg->shm_perm);
520 return 0;
521}
522
523static int
524shmget_allocate_segment(p, uap, mode, retval)
525 struct proc *p;
526 struct shmget_args *uap;
527 int mode;
528 int * retval;
529{
530 int i, segnum, shmid, size;
531 struct ucred *cred = p->p_ucred;
532 struct shmid_ds *shmseg;
533 struct shm_handle *shm_handle;
534 kern_return_t kret;
535 vm_offset_t user_addr;
536 void * mem_object;
537
538 if (uap->size < shminfo.shmmin || uap->size > shminfo.shmmax)
539 return EINVAL;
540 if (shm_nused >= shminfo.shmmni) /* any shmids left? */
541 return ENOSPC;
542 size = round_page_32(uap->size);
543 if (shm_committed + btoc(size) > shminfo.shmall)
544 return ENOMEM;
545 if (shm_last_free < 0) {
546 for (i = 0; i < shminfo.shmmni; i++)
547 if (shmsegs[i].shm_perm.mode & SHMSEG_FREE)
548 break;
549 if (i == shminfo.shmmni)
550 panic("shmseg free count inconsistent");
551 segnum = i;
552 } else {
553 segnum = shm_last_free;
554 shm_last_free = -1;
555 }
556 shmseg = &shmsegs[segnum];
557 /*
558 * In case we sleep in malloc(), mark the segment present but deleted
559 * so that noone else tries to create the same key.
560 */
561 kret = vm_allocate(current_map(), &user_addr, size, TRUE);
562 if (kret != KERN_SUCCESS)
563 goto out;
564
565 kret = mach_make_memory_entry (current_map(), &size,
566 user_addr, VM_PROT_DEFAULT, &mem_object, 0);
567
568 if (kret != KERN_SUCCESS)
569 goto out;
570 shmseg->shm_perm.mode = SHMSEG_ALLOCATED | SHMSEG_REMOVED;
571 shmseg->shm_perm.key = uap->key;
572 shmseg->shm_perm.seq = (shmseg->shm_perm.seq + 1) & 0x7fff;
573 shm_handle = (struct shm_handle *)
574 _MALLOC(sizeof(struct shm_handle), M_SHM, M_WAITOK);
575 shm_handle->shm_object = mem_object;
576 shmid = IXSEQ_TO_IPCID(segnum, shmseg->shm_perm);
577
578 shmseg->shm_internal = shm_handle;
579 shmseg->shm_perm.cuid = shmseg->shm_perm.uid = cred->cr_uid;
580 shmseg->shm_perm.cgid = shmseg->shm_perm.gid = cred->cr_gid;
581 shmseg->shm_perm.mode = (shmseg->shm_perm.mode & SHMSEG_WANTED) |
582 (mode & ACCESSPERMS) | SHMSEG_ALLOCATED;
583 shmseg->shm_segsz = uap->size;
584 shmseg->shm_cpid = p->p_pid;
585 shmseg->shm_lpid = shmseg->shm_nattch = 0;
586 shmseg->shm_atime = shmseg->shm_dtime = 0;
587 shmseg->shm_ctime = time_second;
588 shm_committed += btoc(size);
589 shm_nused++;
590 AUDIT_ARG(svipc_perm, &shmseg->shm_perm);
591 if (shmseg->shm_perm.mode & SHMSEG_WANTED) {
592 /*
593 * Somebody else wanted this key while we were asleep. Wake
594 * them up now.
595 */
596 shmseg->shm_perm.mode &= ~SHMSEG_WANTED;
597 wakeup((caddr_t)shmseg);
598 }
599 *retval = shmid;
600 AUDIT_ARG(svipc_id, shmid);
601 return 0;
602out:
603 switch (kret) {
604 case KERN_INVALID_ADDRESS:
605 case KERN_NO_SPACE:
606 return (ENOMEM);
607 case KERN_PROTECTION_FAILURE:
608 return (EACCES);
609 default:
610 return (EINVAL);
611 }
612
613}
614
615int
616shmget(p, uap, retval)
617 struct proc *p;
618 struct shmget_args *uap;
619 register_t *retval;
620{
621 int segnum, mode, error;
622
623 /* Auditing is actually done in shmget_allocate_segment() */
624 if (!shm_inited)
625 return(EINVAL);
626
627 mode = uap->shmflg & ACCESSPERMS;
628 if (uap->key != IPC_PRIVATE) {
629 again:
630 segnum = shm_find_segment_by_key(uap->key);
631 if (segnum >= 0) {
632 error = shmget_existing(p, uap, mode, segnum, retval);
633 if (error == EAGAIN)
634 goto again;
635 return(error);
636 }
637 if ((uap->shmflg & IPC_CREAT) == 0)
638 return ENOENT;
639 }
640 return( shmget_allocate_segment(p, uap, mode, retval));;
641 /*NOTREACHED*/
642
643}
644
645struct shmsys_args {
646 u_int which;
647 int a2;
648 int a3;
649 int a4;
650};
651int
652shmsys(p, uap, retval)
653 struct proc *p;
654 /* XXX actually varargs. */
655 struct shmsys_args *uap;
656 register_t *retval;
657{
658
659 if (!shm_inited)
660 return(EINVAL);
661
662 if (uap->which >= sizeof(shmcalls)/sizeof(shmcalls[0]))
663 return EINVAL;
664 return ((*shmcalls[uap->which])(p, &uap->a2, retval));
665}
666
667void
668shmfork(p1, p2)
669 struct proc *p1, *p2;
670{
671 struct shmmap_state *shmmap_s;
672 size_t size;
673 int i;
674
675 if (!shm_inited)
676 return;
677 size = shminfo.shmseg * sizeof(struct shmmap_state);
678 shmmap_s = (struct shmmap_state *)_MALLOC(size, M_SHM, M_WAITOK);
679 bcopy((caddr_t)p1->vm_shm, (caddr_t)shmmap_s, size);
680 p2->vm_shm = (caddr_t)shmmap_s;
681 for (i = 0; i < shminfo.shmseg; i++, shmmap_s++)
682 if (shmmap_s->shmid != -1)
683 shmsegs[IPCID_TO_IX(shmmap_s->shmid)].shm_nattch++;
684}
685
686void
687shmexit(p)
688 struct proc *p;
689{
690 struct shmmap_state *shmmap_s;
691 int i;
692
693 shmmap_s = (struct shmmap_state *)p->vm_shm;
694 for (i = 0; i < shminfo.shmseg; i++, shmmap_s++)
695 if (shmmap_s->shmid != -1)
696 shm_delete_mapping(p, shmmap_s, 1);
697 FREE((caddr_t)p->vm_shm, M_SHM);
698 p->vm_shm = NULL;
699}
700
701/*
702 * shmexec() is like shmexit(), only it doesn't delete the mappings,
703 * since the old address space has already been destroyed and the new
704 * one instantiated. Instead, it just does the housekeeping work we
705 * need to do to keep the System V shared memory subsystem sane.
706 */
707__private_extern__ void
708shmexec(p)
709 struct proc *p;
710{
711 struct shmmap_state *shmmap_s;
712 int i;
713
714 shmmap_s = (struct shmmap_state *)p->vm_shm;
715 for (i = 0; i < shminfo.shmseg; i++, shmmap_s++)
716 if (shmmap_s->shmid != -1)
717 shm_delete_mapping(p, shmmap_s, 0);
718 FREE((caddr_t)p->vm_shm, M_SHM);
719 p->vm_shm = NULL;
720}
721
722void
723shminit(dummy)
724 void *dummy;
725{
726 int i;
727 int s;
728
729 if (!shm_inited) {
730 s = sizeof(struct shmid_ds) * shminfo.shmmni;
731
732 MALLOC(shmsegs, struct shmid_ds *, s,
733 M_SHM, M_WAITOK);
734 for (i = 0; i < shminfo.shmmni; i++) {
735 shmsegs[i].shm_perm.mode = SHMSEG_FREE;
736 shmsegs[i].shm_perm.seq = 0;
737 }
738 shm_last_free = 0;
739 shm_nused = 0;
740 shm_committed = 0;
741 shm_inited = 1;
742 }
743}
744
745/* (struct sysctl_oid *oidp, void *arg1, int arg2, \
746 struct sysctl_req *req) */
747static int
748sysctl_shminfo SYSCTL_HANDLER_ARGS
749{
750 int error = 0;
751
752 error = SYSCTL_OUT(req, arg1, sizeof(int));
753 if (error || !req->newptr)
754 return(error);
755
756 /* Set the values only if shared memory is not initialised */
757 if (!shm_inited) {
758 if (error = SYSCTL_IN(req, arg1, sizeof(int)))
759 return(error);
760 if (arg1 == &shminfo.shmmax) {
761 if (shminfo.shmmax & PAGE_MASK) {
762 shminfo.shmmax = -1;
763 return(EINVAL);
764 }
765 }
766
767 /* Initialize only when all values are set */
768 if ((shminfo.shmmax != -1) &&
769 (shminfo.shmmin != -1) &&
770 (shminfo.shmmni != -1) &&
771 (shminfo.shmseg != -1) &&
772 (shminfo.shmall != -1)) {
773 shminit(NULL);
774 }
775 }
776 return(0);
777}
778
779SYSCTL_NODE(_kern, KERN_SYSV, sysv, CTLFLAG_RW, 0, "SYSV");
780
781SYSCTL_PROC(_kern_sysv, KSYSV_SHMMAX, shmmax, CTLTYPE_INT | CTLFLAG_RW,
782 &shminfo.shmmax, 0, &sysctl_shminfo ,"I","shmmax");
783
784SYSCTL_PROC(_kern_sysv, KSYSV_SHMMIN, shmmin, CTLTYPE_INT | CTLFLAG_RW,
785 &shminfo.shmmin, 0, &sysctl_shminfo ,"I","shmmin");
786
787SYSCTL_PROC(_kern_sysv, KSYSV_SHMMNI, shmmni, CTLTYPE_INT | CTLFLAG_RW,
788 &shminfo.shmmni, 0, &sysctl_shminfo ,"I","shmmni");
789
790SYSCTL_PROC(_kern_sysv, KSYSV_SHMSEG, shmseg, CTLTYPE_INT | CTLFLAG_RW,
791 &shminfo.shmseg, 0, &sysctl_shminfo ,"I","shmseg");
792
793SYSCTL_PROC(_kern_sysv, KSYSV_SHMALL, shmall, CTLTYPE_INT | CTLFLAG_RW,
794 &shminfo.shmall, 0, &sysctl_shminfo ,"I","shmall");
795
796