]> git.saurik.com Git - apple/xnu.git/blame - bsd/kern/sysv_shm.c
xnu-2782.30.5.tar.gz
[apple/xnu.git] / bsd / kern / sysv_shm.c
CommitLineData
1c79356b 1/*
2d21ac55 2 * Copyright (c) 2000-2007 Apple Inc. All rights reserved.
5d5c5d0d 3 *
2d21ac55 4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
1c79356b 5 *
2d21ac55
A
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
8f6c56a5 14 *
2d21ac55
A
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
8f6c56a5
A
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
2d21ac55
A
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
8f6c56a5 25 *
2d21ac55 26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
1c79356b
A
27 */
28/* $NetBSD: sysv_shm.c,v 1.23 1994/07/04 23:25:12 glass Exp $ */
29
30/*
31 * Copyright (c) 1994 Adam Glass and Charles Hannum. All rights reserved.
32 *
33 * Redistribution and use in source and binary forms, with or without
34 * modification, are permitted provided that the following conditions
35 * are met:
36 * 1. Redistributions of source code must retain the above copyright
37 * notice, this list of conditions and the following disclaimer.
38 * 2. Redistributions in binary form must reproduce the above copyright
39 * notice, this list of conditions and the following disclaimer in the
40 * documentation and/or other materials provided with the distribution.
41 * 3. All advertising materials mentioning features or use of this software
42 * must display the following acknowledgement:
43 * This product includes software developed by Adam Glass and Charles
44 * Hannum.
45 * 4. The names of the authors may not be used to endorse or promote products
46 * derived from this software without specific prior written permission.
47 *
48 * THIS SOFTWARE IS PROVIDED BY THE AUTHORS ``AS IS'' AND ANY EXPRESS OR
49 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
50 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
51 * IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY DIRECT, INDIRECT,
52 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
53 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
54 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
55 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
56 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
57 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
58 */
2d21ac55
A
59/*
60 * NOTICE: This file was modified by McAfee Research in 2004 to introduce
61 * support for mandatory and extensible security protections. This notice
62 * is included in support of clause 2.2 (b) of the Apple Public License,
63 * Version 2.0.
64 * Copyright (c) 2005-2006 SPARTA, Inc.
65*/
1c79356b
A
66
67
9bccf70c 68#include <sys/appleapiopts.h>
1c79356b
A
69#include <sys/param.h>
70#include <sys/systm.h>
71#include <sys/kernel.h>
91447636
A
72#include <sys/shm_internal.h>
73#include <sys/proc_internal.h>
74#include <sys/kauth.h>
1c79356b
A
75#include <sys/malloc.h>
76#include <sys/mman.h>
77#include <sys/stat.h>
9bccf70c 78#include <sys/sysctl.h>
91447636
A
79#include <sys/ipcs.h>
80#include <sys/sysent.h>
81#include <sys/sysproto.h>
2d21ac55
A
82#if CONFIG_MACF
83#include <security/mac_framework.h>
84#endif
e5568f75 85
b0d623f7 86#include <security/audit/audit.h>
1c79356b
A
87
88#include <mach/mach_types.h>
89#include <mach/vm_inherit.h>
91447636
A
90#include <mach/vm_map.h>
91
92#include <mach/mach_vm.h>
93
1c79356b 94#include <vm/vm_map.h>
91447636 95#include <vm/vm_protos.h>
1c79356b 96
91447636 97#include <kern/locks.h>
1c79356b 98
2d21ac55
A
99/* Uncomment this line to see MAC debugging output. */
100/* #define MAC_DEBUG */
101#if CONFIG_MACF_DEBUG
102#define MPRINTF(a) printf a
103#else
104#define MPRINTF(a)
105#endif
106
107#if SYSV_SHM
91447636 108static void shminit(void *);
1c79356b 109#if 0
1c79356b 110SYSINIT(sysv_shm, SI_SUB_SYSV_SHM, SI_ORDER_FIRST, shminit, NULL)
b0d623f7 111#endif
1c79356b 112
91447636
A
113static lck_grp_t *sysv_shm_subsys_lck_grp;
114static lck_grp_attr_t *sysv_shm_subsys_lck_grp_attr;
115static lck_attr_t *sysv_shm_subsys_lck_attr;
116static lck_mtx_t sysv_shm_subsys_mutex;
1c79356b 117
91447636
A
118#define SYSV_SHM_SUBSYS_LOCK() lck_mtx_lock(&sysv_shm_subsys_mutex)
119#define SYSV_SHM_SUBSYS_UNLOCK() lck_mtx_unlock(&sysv_shm_subsys_mutex)
120
121static int oshmctl(void *p, void *uap, void *retval);
122static int shmget_allocate_segment(struct proc *p, struct shmget_args *uap, int mode, int * retval);
123static int shmget_existing(struct shmget_args *uap, int mode, int segnum, int * retval);
b0d623f7
A
124static void shmid_ds_64to32(struct user_shmid_ds *in, struct user32_shmid_ds *out);
125static void shmid_ds_32to64(struct user32_shmid_ds *in, struct user_shmid_ds *out);
1c79356b
A
126
127/* XXX casting to (sy_call_t *) is bogus, as usual. */
128static sy_call_t *shmcalls[] = {
129 (sy_call_t *)shmat, (sy_call_t *)oshmctl,
130 (sy_call_t *)shmdt, (sy_call_t *)shmget,
131 (sy_call_t *)shmctl
132};
133
134#define SHMSEG_FREE 0x0200
135#define SHMSEG_REMOVED 0x0400
136#define SHMSEG_ALLOCATED 0x0800
137#define SHMSEG_WANTED 0x1000
138
139static int shm_last_free, shm_nused, shm_committed;
2d21ac55 140struct shmid_kernel *shmsegs; /* 64 bit version */
9bccf70c 141static int shm_inited = 0;
1c79356b 142
b0d623f7
A
143/*
144 * Since anonymous memory chunks are limited to ANON_MAX_SIZE bytes,
145 * we have to keep a list of chunks when we want to handle a shared memory
146 * segment bigger than ANON_MAX_SIZE.
147 * Each chunk points to a VM named entry of up to ANON_MAX_SIZE bytes
148 * of anonymous memory.
149 */
1c79356b 150struct shm_handle {
b0d623f7
A
151 void * shm_object; /* named entry for this chunk*/
152 memory_object_size_t shm_handle_size; /* size of this chunk */
153 struct shm_handle *shm_handle_next; /* next chunk */
1c79356b
A
154};
155
156struct shmmap_state {
91447636
A
157 mach_vm_address_t va; /* user address */
158 int shmid; /* segment id */
1c79356b
A
159};
160
2d21ac55 161static void shm_deallocate_segment(struct shmid_kernel *);
91447636 162static int shm_find_segment_by_key(key_t);
2d21ac55 163static struct shmid_kernel *shm_find_segment_by_shmid(int);
91447636 164static int shm_delete_mapping(struct proc *, struct shmmap_state *, int);
1c79356b 165
9bccf70c 166#ifdef __APPLE_API_PRIVATE
2d21ac55
A
167#define DEFAULT_SHMMAX (4 * 1024 * 1024)
168#define DEFAULT_SHMMIN 1
169#define DEFAULT_SHMMNI 32
170#define DEFAULT_SHMSEG 8
171#define DEFAULT_SHMALL 1024
9bccf70c 172struct shminfo shminfo = {
2d21ac55
A
173 DEFAULT_SHMMAX,
174 DEFAULT_SHMMIN,
175 DEFAULT_SHMMNI,
176 DEFAULT_SHMSEG,
177 DEFAULT_SHMALL
9bccf70c
A
178};
179#endif /* __APPLE_API_PRIVATE */
180
91447636
A
181void sysv_shm_lock_init(void);
182
183static __inline__ time_t
184sysv_shmtime(void)
185{
186 struct timeval tv;
187 microtime(&tv);
188 return (tv.tv_sec);
189}
190
191/*
192 * This conversion is safe, since if we are converting for a 32 bit process,
193 * then it's value of (struct shmid_ds)->shm_segsz will never exceed 4G.
194 *
195 * NOTE: Source and target may *NOT* overlap! (target is smaller)
196 */
197static void
b0d623f7 198shmid_ds_64to32(struct user_shmid_ds *in, struct user32_shmid_ds *out)
91447636
A
199{
200 out->shm_perm = in->shm_perm;
b0d623f7 201 out->shm_segsz = in->shm_segsz;
91447636
A
202 out->shm_lpid = in->shm_lpid;
203 out->shm_cpid = in->shm_cpid;
204 out->shm_nattch = in->shm_nattch;
205 out->shm_atime = in->shm_atime;
206 out->shm_dtime = in->shm_dtime;
207 out->shm_ctime = in->shm_ctime;
b0d623f7 208 out->shm_internal = CAST_DOWN_EXPLICIT(int,in->shm_internal);
91447636
A
209}
210
211/*
212 * NOTE: Source and target may are permitted to overlap! (source is smaller);
213 * this works because we copy fields in order from the end of the struct to
214 * the beginning.
215 */
216static void
b0d623f7 217shmid_ds_32to64(struct user32_shmid_ds *in, struct user_shmid_ds *out)
91447636 218{
b0d623f7 219 out->shm_internal = in->shm_internal;
91447636
A
220 out->shm_ctime = in->shm_ctime;
221 out->shm_dtime = in->shm_dtime;
222 out->shm_atime = in->shm_atime;
223 out->shm_nattch = in->shm_nattch;
224 out->shm_cpid = in->shm_cpid;
225 out->shm_lpid = in->shm_lpid;
b0d623f7 226 out->shm_segsz = in->shm_segsz;
91447636
A
227 out->shm_perm = in->shm_perm;
228}
229
230
1c79356b 231static int
91447636 232shm_find_segment_by_key(key_t key)
1c79356b
A
233{
234 int i;
235
236 for (i = 0; i < shminfo.shmmni; i++)
2d21ac55
A
237 if ((shmsegs[i].u.shm_perm.mode & SHMSEG_ALLOCATED) &&
238 shmsegs[i].u.shm_perm._key == key)
1c79356b
A
239 return i;
240 return -1;
241}
242
2d21ac55 243static struct shmid_kernel *
91447636 244shm_find_segment_by_shmid(int shmid)
1c79356b
A
245{
246 int segnum;
2d21ac55 247 struct shmid_kernel *shmseg;
1c79356b
A
248
249 segnum = IPCID_TO_IX(shmid);
250 if (segnum < 0 || segnum >= shminfo.shmmni)
251 return NULL;
252 shmseg = &shmsegs[segnum];
2d21ac55 253 if ((shmseg->u.shm_perm.mode & (SHMSEG_ALLOCATED | SHMSEG_REMOVED))
1c79356b 254 != SHMSEG_ALLOCATED ||
2d21ac55 255 shmseg->u.shm_perm._seq != IPCID_TO_SEQ(shmid))
1c79356b
A
256 return NULL;
257 return shmseg;
258}
259
260static void
2d21ac55 261shm_deallocate_segment(struct shmid_kernel *shmseg)
1c79356b 262{
b0d623f7 263 struct shm_handle *shm_handle, *shm_handle_next;
91447636 264 mach_vm_size_t size;
1c79356b 265
b0d623f7
A
266 for (shm_handle = CAST_DOWN(void *,shmseg->u.shm_internal); /* tunnel */
267 shm_handle != NULL;
268 shm_handle = shm_handle_next) {
269 shm_handle_next = shm_handle->shm_handle_next;
270 mach_memory_entry_port_release(shm_handle->shm_object);
271 FREE((caddr_t) shm_handle, M_SHM);
272 }
2d21ac55 273 shmseg->u.shm_internal = USER_ADDR_NULL; /* tunnel */
b0d623f7 274 size = mach_vm_round_page(shmseg->u.shm_segsz);
1c79356b
A
275 shm_committed -= btoc(size);
276 shm_nused--;
2d21ac55
A
277 shmseg->u.shm_perm.mode = SHMSEG_FREE;
278#if CONFIG_MACF
279 /* Reset the MAC label */
280 mac_sysvshm_label_recycle(shmseg);
281#endif
1c79356b
A
282}
283
284static int
91447636
A
285shm_delete_mapping(__unused struct proc *p, struct shmmap_state *shmmap_s,
286 int deallocate)
1c79356b 287{
2d21ac55 288 struct shmid_kernel *shmseg;
1c79356b 289 int segnum, result;
91447636 290 mach_vm_size_t size;
1c79356b
A
291
292 segnum = IPCID_TO_IX(shmmap_s->shmid);
293 shmseg = &shmsegs[segnum];
2d21ac55 294 size = mach_vm_round_page(shmseg->u.shm_segsz); /* XXX done for us? */
55e303ae 295 if (deallocate) {
91447636 296 result = mach_vm_deallocate(current_map(), shmmap_s->va, size);
1c79356b
A
297 if (result != KERN_SUCCESS)
298 return EINVAL;
55e303ae 299 }
1c79356b 300 shmmap_s->shmid = -1;
2d21ac55
A
301 shmseg->u.shm_dtime = sysv_shmtime();
302 if ((--shmseg->u.shm_nattch <= 0) &&
303 (shmseg->u.shm_perm.mode & SHMSEG_REMOVED)) {
1c79356b
A
304 shm_deallocate_segment(shmseg);
305 shm_last_free = segnum;
306 }
307 return 0;
308}
309
1c79356b 310int
b0d623f7 311shmdt(struct proc *p, struct shmdt_args *uap, int32_t *retval)
1c79356b 312{
2d21ac55
A
313#if CONFIG_MACF
314 struct shmid_kernel *shmsegptr;
315#endif
1c79356b
A
316 struct shmmap_state *shmmap_s;
317 int i;
91447636
A
318 int shmdtret = 0;
319
2d21ac55 320 AUDIT_ARG(svipc_addr, uap->shmaddr);
91447636
A
321
322 SYSV_SHM_SUBSYS_LOCK();
1c79356b 323
91447636 324 if (!shm_inited) {
2d21ac55 325 shminit(NULL);
91447636 326 }
1c79356b 327 shmmap_s = (struct shmmap_state *)p->vm_shm;
91447636
A
328 if (shmmap_s == NULL) {
329 shmdtret = EINVAL;
330 goto shmdt_out;
331 }
332
1c79356b
A
333 for (i = 0; i < shminfo.shmseg; i++, shmmap_s++)
334 if (shmmap_s->shmid != -1 &&
91447636 335 shmmap_s->va == (mach_vm_offset_t)uap->shmaddr)
1c79356b 336 break;
91447636
A
337 if (i == shminfo.shmseg) {
338 shmdtret = EINVAL;
339 goto shmdt_out;
340 }
2d21ac55
A
341#if CONFIG_MACF
342 /*
343 * XXX: It might be useful to move this into the shm_delete_mapping
344 * function
345 */
346 shmsegptr = &shmsegs[IPCID_TO_IX(shmmap_s->shmid)];
347 shmdtret = mac_sysvshm_check_shmdt(kauth_cred_get(), shmsegptr);
348 if (shmdtret)
349 goto shmdt_out;
350#endif
91447636
A
351 i = shm_delete_mapping(p, shmmap_s, 1);
352
353 if (i == 0)
354 *retval = 0;
355 shmdtret = i;
356shmdt_out:
357 SYSV_SHM_SUBSYS_UNLOCK();
358 return shmdtret;
1c79356b
A
359}
360
1c79356b 361int
2d21ac55 362shmat(struct proc *p, struct shmat_args *uap, user_addr_t *retval)
1c79356b
A
363{
364 int error, i, flags;
2d21ac55 365 struct shmid_kernel *shmseg;
91447636
A
366 struct shmmap_state *shmmap_s = NULL;
367 struct shm_handle *shm_handle;
368 mach_vm_address_t attach_va; /* attach address in/out */
369 mach_vm_size_t map_size; /* size of map entry */
b0d623f7 370 mach_vm_size_t mapped_size;
91447636
A
371 vm_prot_t prot;
372 size_t size;
373 kern_return_t rv;
b0d623f7
A
374 int shmat_ret;
375 int vm_flags;
376
377 shmat_ret = 0;
1c79356b 378
55e303ae 379 AUDIT_ARG(svipc_id, uap->shmid);
2d21ac55 380 AUDIT_ARG(svipc_addr, uap->shmaddr);
91447636
A
381
382 SYSV_SHM_SUBSYS_LOCK();
383
384 if (!shm_inited) {
2d21ac55 385 shminit(NULL);
91447636
A
386 }
387
1c79356b 388 shmmap_s = (struct shmmap_state *)p->vm_shm;
91447636 389
1c79356b
A
390 if (shmmap_s == NULL) {
391 size = shminfo.shmseg * sizeof(struct shmmap_state);
91447636
A
392 MALLOC(shmmap_s, struct shmmap_state *, size, M_SHM, M_WAITOK);
393 if (shmmap_s == NULL) {
394 shmat_ret = ENOMEM;
395 goto shmat_out;
396 }
1c79356b
A
397 for (i = 0; i < shminfo.shmseg; i++)
398 shmmap_s[i].shmid = -1;
399 p->vm_shm = (caddr_t)shmmap_s;
400 }
401 shmseg = shm_find_segment_by_shmid(uap->shmid);
91447636
A
402 if (shmseg == NULL) {
403 shmat_ret = EINVAL;
404 goto shmat_out;
405 }
55e303ae 406
2d21ac55
A
407 AUDIT_ARG(svipc_perm, &shmseg->u.shm_perm);
408 error = ipcperm(kauth_cred_get(), &shmseg->u.shm_perm,
1c79356b 409 (uap->shmflg & SHM_RDONLY) ? IPC_R : IPC_R|IPC_W);
91447636
A
410 if (error) {
411 shmat_ret = error;
412 goto shmat_out;
413 }
414
2d21ac55
A
415#if CONFIG_MACF
416 error = mac_sysvshm_check_shmat(kauth_cred_get(), shmseg, uap->shmflg);
417 if (error) {
418 shmat_ret = error;
419 goto shmat_out;
420 }
421#endif
1c79356b
A
422 for (i = 0; i < shminfo.shmseg; i++) {
423 if (shmmap_s->shmid == -1)
424 break;
425 shmmap_s++;
426 }
91447636
A
427 if (i >= shminfo.shmseg) {
428 shmat_ret = EMFILE;
429 goto shmat_out;
430 }
431
2d21ac55 432 map_size = mach_vm_round_page(shmseg->u.shm_segsz);
1c79356b
A
433 prot = VM_PROT_READ;
434 if ((uap->shmflg & SHM_RDONLY) == 0)
435 prot |= VM_PROT_WRITE;
436 flags = MAP_ANON | MAP_SHARED;
91447636 437 if (uap->shmaddr)
1c79356b 438 flags |= MAP_FIXED;
91447636
A
439
440 attach_va = (mach_vm_address_t)uap->shmaddr;
441 if (uap->shmflg & SHM_RND)
442 attach_va &= ~(SHMLBA-1);
443 else if ((attach_va & (SHMLBA-1)) != 0) {
444 shmat_ret = EINVAL;
445 goto shmat_out;
446 }
447
b0d623f7
A
448 if (flags & MAP_FIXED) {
449 vm_flags = VM_FLAGS_FIXED;
450 } else {
451 vm_flags = VM_FLAGS_ANYWHERE;
452 }
453
454 mapped_size = 0;
455
456 /* first reserve enough space... */
457 rv = mach_vm_map(current_map(),
458 &attach_va,
459 map_size,
460 0,
461 vm_flags,
462 IPC_PORT_NULL,
463 0,
464 FALSE,
465 VM_PROT_NONE,
466 VM_PROT_NONE,
467 VM_INHERIT_NONE);
468 if (rv != KERN_SUCCESS) {
469 goto out;
470 }
91447636 471
b0d623f7
A
472 shmmap_s->va = attach_va;
473
474 /* ... then map the shared memory over the reserved space */
475 for (shm_handle = CAST_DOWN(void *, shmseg->u.shm_internal);/* tunnel */
476 shm_handle != NULL;
477 shm_handle = shm_handle->shm_handle_next) {
478
479 rv = vm_map_enter_mem_object(
480 current_map(), /* process map */
481 &attach_va, /* attach address */
482 shm_handle->shm_handle_size, /* segment size */
483 (mach_vm_offset_t)0, /* alignment mask */
484 VM_FLAGS_FIXED | VM_FLAGS_OVERWRITE,
91447636
A
485 shm_handle->shm_object,
486 (mach_vm_offset_t)0,
487 FALSE,
488 prot,
489 prot,
b0d623f7
A
490 VM_INHERIT_SHARE);
491 if (rv != KERN_SUCCESS)
1c79356b 492 goto out;
91447636 493
b0d623f7
A
494 mapped_size += shm_handle->shm_handle_size;
495 attach_va = attach_va + shm_handle->shm_handle_size;
1c79356b
A
496 }
497
1c79356b 498 shmmap_s->shmid = uap->shmid;
2d21ac55
A
499 shmseg->u.shm_lpid = p->p_pid;
500 shmseg->u.shm_atime = sysv_shmtime();
501 shmseg->u.shm_nattch++;
b0d623f7 502 *retval = shmmap_s->va; /* XXX return -1 on error */
91447636
A
503 shmat_ret = 0;
504 goto shmat_out;
1c79356b 505out:
b0d623f7
A
506 if (mapped_size > 0) {
507 (void) mach_vm_deallocate(current_map(),
508 shmmap_s->va,
509 mapped_size);
510 }
1c79356b
A
511 switch (rv) {
512 case KERN_INVALID_ADDRESS:
513 case KERN_NO_SPACE:
91447636 514 shmat_ret = ENOMEM;
b0d623f7 515 break;
1c79356b 516 case KERN_PROTECTION_FAILURE:
91447636 517 shmat_ret = EACCES;
b0d623f7 518 break;
1c79356b 519 default:
91447636 520 shmat_ret = EINVAL;
b0d623f7 521 break;
1c79356b 522 }
91447636
A
523shmat_out:
524 SYSV_SHM_SUBSYS_UNLOCK();
525 return shmat_ret;
1c79356b
A
526}
527
1c79356b 528static int
91447636 529oshmctl(__unused void *p, __unused void *uap, __unused void *retval)
1c79356b 530{
1c79356b 531 return EINVAL;
1c79356b
A
532}
533
2d21ac55
A
534/*
535 * Returns: 0 Success
536 * EINVAL
537 * copyout:EFAULT
538 * copyin:EFAULT
539 * ipcperm:EPERM
540 * ipcperm:EACCES
541 */
1c79356b 542int
b0d623f7 543shmctl(__unused struct proc *p, struct shmctl_args *uap, int32_t *retval)
1c79356b
A
544{
545 int error;
91447636
A
546 kauth_cred_t cred = kauth_cred_get();
547 struct user_shmid_ds inbuf;
2d21ac55 548 struct shmid_kernel *shmseg;
91447636
A
549
550 int shmctl_ret = 0;
1c79356b 551
55e303ae
A
552 AUDIT_ARG(svipc_cmd, uap->cmd);
553 AUDIT_ARG(svipc_id, uap->shmid);
91447636
A
554
555 SYSV_SHM_SUBSYS_LOCK();
556
557 if (!shm_inited) {
2d21ac55 558 shminit(NULL);
91447636
A
559 }
560
1c79356b 561 shmseg = shm_find_segment_by_shmid(uap->shmid);
91447636
A
562 if (shmseg == NULL) {
563 shmctl_ret = EINVAL;
564 goto shmctl_out;
565 }
566
55e303ae
A
567 /* XXAUDIT: This is the perms BEFORE any change by this call. This
568 * may not be what is desired.
569 */
2d21ac55 570 AUDIT_ARG(svipc_perm, &shmseg->u.shm_perm);
55e303ae 571
2d21ac55
A
572#if CONFIG_MACF
573 error = mac_sysvshm_check_shmctl(cred, shmseg, uap->cmd);
574 if (error) {
575 shmctl_ret = error;
576 goto shmctl_out;
577 }
578#endif
1c79356b
A
579 switch (uap->cmd) {
580 case IPC_STAT:
2d21ac55 581 error = ipcperm(cred, &shmseg->u.shm_perm, IPC_R);
91447636
A
582 if (error) {
583 shmctl_ret = error;
584 goto shmctl_out;
585 }
586
587 if (IS_64BIT_PROCESS(p)) {
22ba694c
A
588 struct user_shmid_ds shmid_ds;
589 memcpy(&shmid_ds, &shmseg->u, sizeof(struct user_shmid_ds));
590
591 /* Clear kernel reserved pointer before copying to user space */
592 shmid_ds.shm_internal = USER_ADDR_NULL;
593
594 error = copyout(&shmid_ds, uap->buf, sizeof(shmid_ds));
91447636 595 } else {
b0d623f7 596 struct user32_shmid_ds shmid_ds32;
2d21ac55 597 shmid_ds_64to32(&shmseg->u, &shmid_ds32);
22ba694c
A
598
599 /* Clear kernel reserved pointer before copying to user space */
600 shmid_ds32.shm_internal = (user32_addr_t)0;
601
b0d623f7 602 error = copyout(&shmid_ds32, uap->buf, sizeof(shmid_ds32));
91447636
A
603 }
604 if (error) {
605 shmctl_ret = error;
606 goto shmctl_out;
607 }
1c79356b
A
608 break;
609 case IPC_SET:
2d21ac55 610 error = ipcperm(cred, &shmseg->u.shm_perm, IPC_M);
91447636
A
611 if (error) {
612 shmctl_ret = error;
613 goto shmctl_out;
614 }
615 if (IS_64BIT_PROCESS(p)) {
616 error = copyin(uap->buf, &inbuf, sizeof(struct user_shmid_ds));
617 } else {
b0d623f7
A
618 struct user32_shmid_ds shmid_ds32;
619 error = copyin(uap->buf, &shmid_ds32, sizeof(shmid_ds32));
91447636 620 /* convert in place; ugly, but safe */
b0d623f7 621 shmid_ds_32to64(&shmid_ds32, &inbuf);
91447636
A
622 }
623 if (error) {
624 shmctl_ret = error;
625 goto shmctl_out;
626 }
2d21ac55
A
627 shmseg->u.shm_perm.uid = inbuf.shm_perm.uid;
628 shmseg->u.shm_perm.gid = inbuf.shm_perm.gid;
629 shmseg->u.shm_perm.mode =
630 (shmseg->u.shm_perm.mode & ~ACCESSPERMS) |
1c79356b 631 (inbuf.shm_perm.mode & ACCESSPERMS);
2d21ac55 632 shmseg->u.shm_ctime = sysv_shmtime();
1c79356b
A
633 break;
634 case IPC_RMID:
2d21ac55 635 error = ipcperm(cred, &shmseg->u.shm_perm, IPC_M);
91447636
A
636 if (error) {
637 shmctl_ret = error;
638 goto shmctl_out;
639 }
2d21ac55
A
640 shmseg->u.shm_perm._key = IPC_PRIVATE;
641 shmseg->u.shm_perm.mode |= SHMSEG_REMOVED;
642 if (shmseg->u.shm_nattch <= 0) {
1c79356b
A
643 shm_deallocate_segment(shmseg);
644 shm_last_free = IPCID_TO_IX(uap->shmid);
645 }
646 break;
647#if 0
648 case SHM_LOCK:
649 case SHM_UNLOCK:
650#endif
651 default:
91447636
A
652 shmctl_ret = EINVAL;
653 goto shmctl_out;
1c79356b 654 }
91447636
A
655 *retval = 0;
656 shmctl_ret = 0;
657shmctl_out:
658 SYSV_SHM_SUBSYS_UNLOCK();
659 return shmctl_ret;
1c79356b
A
660}
661
1c79356b 662static int
91447636 663shmget_existing(struct shmget_args *uap, int mode, int segnum, int *retval)
1c79356b 664{
2d21ac55
A
665 struct shmid_kernel *shmseg;
666 int error = 0;
1c79356b
A
667
668 shmseg = &shmsegs[segnum];
2d21ac55 669 if (shmseg->u.shm_perm.mode & SHMSEG_REMOVED) {
1c79356b
A
670 /*
671 * This segment is in the process of being allocated. Wait
672 * until it's done, and look the key up again (in case the
673 * allocation failed or it was freed).
674 */
2d21ac55 675 shmseg->u.shm_perm.mode |= SHMSEG_WANTED;
1c79356b
A
676 error = tsleep((caddr_t)shmseg, PLOCK | PCATCH, "shmget", 0);
677 if (error)
678 return error;
679 return EAGAIN;
680 }
2d21ac55
A
681
682 /*
683 * The low 9 bits of shmflag are the mode bits being requested, which
684 * are the actual mode bits desired on the segment, and not in IPC_R
685 * form; therefore it would be incorrect to call ipcperm() to validate
686 * them; instead, we AND the existing mode with the requested mode, and
687 * verify that it matches the requested mode; otherwise, we fail with
688 * EACCES (access denied).
689 */
690 if ((shmseg->u.shm_perm.mode & mode) != mode)
691 return EACCES;
692
693#if CONFIG_MACF
694 error = mac_sysvshm_check_shmget(kauth_cred_get(), shmseg, uap->shmflg);
695 if (error)
696 return (error);
697#endif
698
699 if (uap->size && uap->size > shmseg->u.shm_segsz)
1c79356b 700 return EINVAL;
2d21ac55 701
1c79356b
A
702 if ((uap->shmflg & (IPC_CREAT | IPC_EXCL)) == (IPC_CREAT | IPC_EXCL))
703 return EEXIST;
2d21ac55
A
704
705 *retval = IXSEQ_TO_IPCID(segnum, shmseg->u.shm_perm);
1c79356b
A
706 return 0;
707}
708
709static int
91447636
A
710shmget_allocate_segment(struct proc *p, struct shmget_args *uap, int mode,
711 int *retval)
1c79356b 712{
2d21ac55 713 int i, segnum, shmid;
91447636 714 kauth_cred_t cred = kauth_cred_get();
2d21ac55 715 struct shmid_kernel *shmseg;
1c79356b
A
716 struct shm_handle *shm_handle;
717 kern_return_t kret;
b0d623f7 718 mach_vm_size_t total_size, size, alloc_size;
1c79356b 719 void * mem_object;
b0d623f7 720 struct shm_handle *shm_handle_next, **shm_handle_next_p;
1c79356b 721
91447636
A
722 if (uap->size < (user_size_t)shminfo.shmmin ||
723 uap->size > (user_size_t)shminfo.shmmax)
1c79356b
A
724 return EINVAL;
725 if (shm_nused >= shminfo.shmmni) /* any shmids left? */
726 return ENOSPC;
b0d623f7
A
727 total_size = mach_vm_round_page(uap->size);
728 if ((user_ssize_t)(shm_committed + btoc(total_size)) > shminfo.shmall)
1c79356b
A
729 return ENOMEM;
730 if (shm_last_free < 0) {
731 for (i = 0; i < shminfo.shmmni; i++)
2d21ac55 732 if (shmsegs[i].u.shm_perm.mode & SHMSEG_FREE)
1c79356b
A
733 break;
734 if (i == shminfo.shmmni)
735 panic("shmseg free count inconsistent");
736 segnum = i;
737 } else {
738 segnum = shm_last_free;
739 shm_last_free = -1;
740 }
741 shmseg = &shmsegs[segnum];
b0d623f7 742
1c79356b
A
743 /*
744 * In case we sleep in malloc(), mark the segment present but deleted
745 * so that noone else tries to create the same key.
b0d623f7 746 * XXX but we don't release the global lock !?
1c79356b 747 */
2d21ac55
A
748 shmseg->u.shm_perm.mode = SHMSEG_ALLOCATED | SHMSEG_REMOVED;
749 shmseg->u.shm_perm._key = uap->key;
750 shmseg->u.shm_perm._seq = (shmseg->u.shm_perm._seq + 1) & 0x7fff;
b0d623f7
A
751
752 shm_handle_next_p = NULL;
753 for (alloc_size = 0;
754 alloc_size < total_size;
755 alloc_size += size) {
756 size = MIN(total_size - alloc_size, ANON_MAX_SIZE);
757 kret = mach_make_memory_entry_64(
758 VM_MAP_NULL,
759 (memory_object_size_t *) &size,
760 (memory_object_offset_t) 0,
761 MAP_MEM_NAMED_CREATE | VM_PROT_DEFAULT,
762 (ipc_port_t *) &mem_object, 0);
763 if (kret != KERN_SUCCESS)
764 goto out;
765
766 MALLOC(shm_handle, struct shm_handle *, sizeof(struct shm_handle), M_SHM, M_WAITOK);
767 if (shm_handle == NULL) {
768 kret = KERN_NO_SPACE;
769 mach_memory_entry_port_release(mem_object);
770 mem_object = NULL;
771 goto out;
772 }
773 shm_handle->shm_object = mem_object;
774 shm_handle->shm_handle_size = size;
775 shm_handle->shm_handle_next = NULL;
776 if (shm_handle_next_p == NULL) {
777 shmseg->u.shm_internal = CAST_USER_ADDR_T(shm_handle);/* tunnel */
778 } else {
779 *shm_handle_next_p = shm_handle;
780 }
781 shm_handle_next_p = &shm_handle->shm_handle_next;
91447636 782 }
b0d623f7 783
2d21ac55 784 shmid = IXSEQ_TO_IPCID(segnum, shmseg->u.shm_perm);
1c79356b 785
2d21ac55 786 shmseg->u.shm_perm.cuid = shmseg->u.shm_perm.uid = kauth_cred_getuid(cred);
6d2010ae 787 shmseg->u.shm_perm.cgid = shmseg->u.shm_perm.gid = kauth_cred_getgid(cred);
2d21ac55 788 shmseg->u.shm_perm.mode = (shmseg->u.shm_perm.mode & SHMSEG_WANTED) |
1c79356b 789 (mode & ACCESSPERMS) | SHMSEG_ALLOCATED;
2d21ac55
A
790 shmseg->u.shm_segsz = uap->size;
791 shmseg->u.shm_cpid = p->p_pid;
792 shmseg->u.shm_lpid = shmseg->u.shm_nattch = 0;
793 shmseg->u.shm_atime = shmseg->u.shm_dtime = 0;
794#if CONFIG_MACF
795 mac_sysvshm_label_associate(cred, shmseg);
796#endif
797 shmseg->u.shm_ctime = sysv_shmtime();
1c79356b
A
798 shm_committed += btoc(size);
799 shm_nused++;
2d21ac55
A
800 AUDIT_ARG(svipc_perm, &shmseg->u.shm_perm);
801 if (shmseg->u.shm_perm.mode & SHMSEG_WANTED) {
1c79356b
A
802 /*
803 * Somebody else wanted this key while we were asleep. Wake
804 * them up now.
805 */
2d21ac55 806 shmseg->u.shm_perm.mode &= ~SHMSEG_WANTED;
1c79356b
A
807 wakeup((caddr_t)shmseg);
808 }
809 *retval = shmid;
55e303ae 810 AUDIT_ARG(svipc_id, shmid);
1c79356b
A
811 return 0;
812out:
b0d623f7
A
813 if (kret != KERN_SUCCESS) {
814 for (shm_handle = CAST_DOWN(void *,shmseg->u.shm_internal); /* tunnel */
815 shm_handle != NULL;
816 shm_handle = shm_handle_next) {
817 shm_handle_next = shm_handle->shm_handle_next;
818 mach_memory_entry_port_release(shm_handle->shm_object);
819 FREE((caddr_t) shm_handle, M_SHM);
820 }
821 shmseg->u.shm_internal = USER_ADDR_NULL; /* tunnel */
822 }
823
1c79356b
A
824 switch (kret) {
825 case KERN_INVALID_ADDRESS:
826 case KERN_NO_SPACE:
827 return (ENOMEM);
828 case KERN_PROTECTION_FAILURE:
829 return (EACCES);
830 default:
831 return (EINVAL);
832 }
833
834}
835
836int
b0d623f7 837shmget(struct proc *p, struct shmget_args *uap, int32_t *retval)
1c79356b
A
838{
839 int segnum, mode, error;
91447636
A
840 int shmget_ret = 0;
841
55e303ae 842 /* Auditing is actually done in shmget_allocate_segment() */
91447636
A
843
844 SYSV_SHM_SUBSYS_LOCK();
845
846 if (!shm_inited) {
2d21ac55 847 shminit(NULL);
91447636 848 }
9bccf70c 849
1c79356b
A
850 mode = uap->shmflg & ACCESSPERMS;
851 if (uap->key != IPC_PRIVATE) {
852 again:
853 segnum = shm_find_segment_by_key(uap->key);
854 if (segnum >= 0) {
91447636 855 error = shmget_existing(uap, mode, segnum, retval);
1c79356b
A
856 if (error == EAGAIN)
857 goto again;
91447636
A
858 shmget_ret = error;
859 goto shmget_out;
860 }
861 if ((uap->shmflg & IPC_CREAT) == 0) {
862 shmget_ret = ENOENT;
863 goto shmget_out;
1c79356b 864 }
1c79356b 865 }
91447636
A
866 shmget_ret = shmget_allocate_segment(p, uap, mode, retval);
867shmget_out:
868 SYSV_SHM_SUBSYS_UNLOCK();
869 return shmget_ret;
1c79356b
A
870 /*NOTREACHED*/
871
872}
873
b0d623f7
A
874/*
875 * shmsys
876 *
877 * Entry point for all SHM calls: shmat, oshmctl, shmdt, shmget, shmctl
878 *
879 * Parameters: p Process requesting the call
880 * uap User argument descriptor (see below)
881 * retval Return value of the selected shm call
882 *
883 * Indirect parameters: uap->which msg call to invoke (index in array of shm calls)
884 * uap->a2 User argument descriptor
885 *
886 * Returns: 0 Success
887 * !0 Not success
888 *
889 * Implicit returns: retval Return value of the selected shm call
890 *
891 * DEPRECATED: This interface should not be used to call the other SHM
892 * functions (shmat, oshmctl, shmdt, shmget, shmctl). The correct
893 * usage is to call the other SHM functions directly.
894 */
1c79356b 895int
b0d623f7 896shmsys(struct proc *p, struct shmsys_args *uap, int32_t *retval)
1c79356b
A
897{
898
91447636 899 /* The routine that we are dispatching already does this */
9bccf70c 900
1c79356b
A
901 if (uap->which >= sizeof(shmcalls)/sizeof(shmcalls[0]))
902 return EINVAL;
903 return ((*shmcalls[uap->which])(p, &uap->a2, retval));
904}
905
91447636
A
906/*
907 * Return 0 on success, 1 on failure.
908 */
909int
910shmfork(struct proc *p1, struct proc *p2)
1c79356b
A
911{
912 struct shmmap_state *shmmap_s;
913 size_t size;
914 int i;
91447636 915 int shmfork_ret = 0;
1c79356b 916
91447636
A
917 SYSV_SHM_SUBSYS_LOCK();
918
919 if (!shm_inited) {
2d21ac55 920 shminit(NULL);
91447636
A
921 }
922
1c79356b 923 size = shminfo.shmseg * sizeof(struct shmmap_state);
91447636
A
924 MALLOC(shmmap_s, struct shmmap_state *, size, M_SHM, M_WAITOK);
925 if (shmmap_s != NULL) {
926 bcopy((caddr_t)p1->vm_shm, (caddr_t)shmmap_s, size);
927 p2->vm_shm = (caddr_t)shmmap_s;
928 for (i = 0; i < shminfo.shmseg; i++, shmmap_s++)
929 if (shmmap_s->shmid != -1)
2d21ac55 930 shmsegs[IPCID_TO_IX(shmmap_s->shmid)].u.shm_nattch++;
91447636
A
931 shmfork_ret = 0;
932 goto shmfork_out;
933 }
934
935 shmfork_ret = 1; /* failed to copy to child - ENOMEM */
936shmfork_out:
937 SYSV_SHM_SUBSYS_UNLOCK();
938 return shmfork_ret;
1c79356b
A
939}
940
941void
91447636 942shmexit(struct proc *p)
1c79356b
A
943{
944 struct shmmap_state *shmmap_s;
945 int i;
946
947 shmmap_s = (struct shmmap_state *)p->vm_shm;
91447636
A
948
949 SYSV_SHM_SUBSYS_LOCK();
1c79356b
A
950 for (i = 0; i < shminfo.shmseg; i++, shmmap_s++)
951 if (shmmap_s->shmid != -1)
2d21ac55
A
952 /*
953 * XXX: Should the MAC framework enforce
954 * check here as well.
955 */
55e303ae
A
956 shm_delete_mapping(p, shmmap_s, 1);
957 FREE((caddr_t)p->vm_shm, M_SHM);
958 p->vm_shm = NULL;
91447636 959 SYSV_SHM_SUBSYS_UNLOCK();
55e303ae
A
960}
961
962/*
963 * shmexec() is like shmexit(), only it doesn't delete the mappings,
964 * since the old address space has already been destroyed and the new
965 * one instantiated. Instead, it just does the housekeeping work we
966 * need to do to keep the System V shared memory subsystem sane.
967 */
968__private_extern__ void
91447636 969shmexec(struct proc *p)
55e303ae
A
970{
971 struct shmmap_state *shmmap_s;
972 int i;
973
974 shmmap_s = (struct shmmap_state *)p->vm_shm;
91447636 975 SYSV_SHM_SUBSYS_LOCK();
55e303ae
A
976 for (i = 0; i < shminfo.shmseg; i++, shmmap_s++)
977 if (shmmap_s->shmid != -1)
978 shm_delete_mapping(p, shmmap_s, 0);
1c79356b
A
979 FREE((caddr_t)p->vm_shm, M_SHM);
980 p->vm_shm = NULL;
91447636 981 SYSV_SHM_SUBSYS_UNLOCK();
1c79356b
A
982}
983
984void
91447636 985shminit(__unused void *dummy)
1c79356b
A
986{
987 int i;
988 int s;
989
9bccf70c 990 if (!shm_inited) {
91447636
A
991 /*
992 * we store internally 64 bit, since if we didn't, we would
993 * be unable to represent a segment size in excess of 32 bits
994 * with the (struct shmid_ds)->shm_segsz field; also, POSIX
995 * dictates this filed be a size_t, which is 64 bits when
996 * running 64 bit binaries.
997 */
2d21ac55 998 s = sizeof(struct shmid_kernel) * shminfo.shmmni;
9bccf70c 999
2d21ac55 1000 MALLOC(shmsegs, struct shmid_kernel *, s, M_SHM, M_WAITOK);
91447636
A
1001 if (shmsegs == NULL) {
1002 /* XXX fail safely: leave shared memory uninited */
1003 return;
1004 }
9bccf70c 1005 for (i = 0; i < shminfo.shmmni; i++) {
2d21ac55
A
1006 shmsegs[i].u.shm_perm.mode = SHMSEG_FREE;
1007 shmsegs[i].u.shm_perm._seq = 0;
1008#if CONFIG_MACF
1009 mac_sysvshm_label_init(&shmsegs[i]);
1010#endif
9bccf70c
A
1011 }
1012 shm_last_free = 0;
1013 shm_nused = 0;
1014 shm_committed = 0;
1015 shm_inited = 1;
1016 }
1017}
91447636
A
1018/* Initialize the mutex governing access to the SysV shm subsystem */
1019__private_extern__ void
1020sysv_shm_lock_init( void )
1021{
1022
1023 sysv_shm_subsys_lck_grp_attr = lck_grp_attr_alloc_init();
91447636
A
1024
1025 sysv_shm_subsys_lck_grp = lck_grp_alloc_init("sysv_shm_subsys_lock", sysv_shm_subsys_lck_grp_attr);
1026
1027 sysv_shm_subsys_lck_attr = lck_attr_alloc_init();
91447636
A
1028 lck_mtx_init(&sysv_shm_subsys_mutex, sysv_shm_subsys_lck_grp, sysv_shm_subsys_lck_attr);
1029}
9bccf70c
A
1030
1031/* (struct sysctl_oid *oidp, void *arg1, int arg2, \
1032 struct sysctl_req *req) */
1033static int
91447636
A
1034sysctl_shminfo(__unused struct sysctl_oid *oidp, void *arg1,
1035 __unused int arg2, struct sysctl_req *req)
9bccf70c
A
1036{
1037 int error = 0;
91447636 1038 int sysctl_shminfo_ret = 0;
2d21ac55 1039 uint64_t saved_shmmax;
9bccf70c 1040
0c530ab8 1041 error = SYSCTL_OUT(req, arg1, sizeof(int64_t));
91447636 1042 if (error || req->newptr == USER_ADDR_NULL)
9bccf70c 1043 return(error);
1c79356b 1044
91447636 1045 SYSV_SHM_SUBSYS_LOCK();
91447636 1046
2d21ac55
A
1047 /* shmmni can not be changed after SysV SHM has been initialized */
1048 if (shm_inited && arg1 == &shminfo.shmmni) {
1049 sysctl_shminfo_ret = EPERM;
1050 goto sysctl_shminfo_out;
1051 }
1052 saved_shmmax = shminfo.shmmax;
9bccf70c 1053
2d21ac55
A
1054 if ((error = SYSCTL_IN(req, arg1, sizeof(int64_t))) != 0) {
1055 sysctl_shminfo_ret = error;
1056 goto sysctl_shminfo_out;
1057 }
1058
1059 if (arg1 == &shminfo.shmmax) {
1060 /* shmmax needs to be page-aligned */
1061 if (shminfo.shmmax & PAGE_MASK_64) {
1062 shminfo.shmmax = saved_shmmax;
1063 sysctl_shminfo_ret = EINVAL;
1064 goto sysctl_shminfo_out;
9bccf70c 1065 }
1c79356b 1066 }
91447636
A
1067 sysctl_shminfo_ret = 0;
1068sysctl_shminfo_out:
1069 SYSV_SHM_SUBSYS_UNLOCK();
1070 return sysctl_shminfo_ret;
1071}
1072
1073static int
1074IPCS_shm_sysctl(__unused struct sysctl_oid *oidp, __unused void *arg1,
1075 __unused int arg2, struct sysctl_req *req)
1076{
1077 int error;
1078 int cursor;
1079 union {
b0d623f7 1080 struct user32_IPCS_command u32;
91447636
A
1081 struct user_IPCS_command u64;
1082 } ipcs;
b0d623f7 1083 struct user32_shmid_ds shmid_ds32; /* post conversion, 32 bit version */
22ba694c 1084 struct user_shmid_ds shmid_ds; /* 64 bit version */
91447636
A
1085 void *shmid_dsp;
1086 size_t ipcs_sz = sizeof(struct user_IPCS_command);
1087 size_t shmid_ds_sz = sizeof(struct user_shmid_ds);
1088 struct proc *p = current_proc();
1089
91447636
A
1090 SYSV_SHM_SUBSYS_LOCK();
1091
1092 if (!shm_inited) {
2d21ac55 1093 shminit(NULL);
91447636
A
1094 }
1095
1096 if (!IS_64BIT_PROCESS(p)) {
b0d623f7
A
1097 ipcs_sz = sizeof(struct user32_IPCS_command);
1098 shmid_ds_sz = sizeof(struct user32_shmid_ds);
91447636
A
1099 }
1100
1101 /* Copy in the command structure */
1102 if ((error = SYSCTL_IN(req, &ipcs, ipcs_sz)) != 0) {
1103 goto ipcs_shm_sysctl_out;
1104 }
1105
1106 if (!IS_64BIT_PROCESS(p)) /* convert in place */
1107 ipcs.u64.ipcs_data = CAST_USER_ADDR_T(ipcs.u32.ipcs_data);
1108
1109 /* Let us version this interface... */
1110 if (ipcs.u64.ipcs_magic != IPCS_MAGIC) {
1111 error = EINVAL;
1112 goto ipcs_shm_sysctl_out;
1113 }
1114
1115 switch(ipcs.u64.ipcs_op) {
1116 case IPCS_SHM_CONF: /* Obtain global configuration data */
1117 if (ipcs.u64.ipcs_datalen != sizeof(struct shminfo)) {
1118 if (ipcs.u64.ipcs_cursor != 0) { /* fwd. compat. */
1119 error = ENOMEM;
1120 break;
1121 }
1122 error = ERANGE;
1123 break;
1124 }
1125 error = copyout(&shminfo, ipcs.u64.ipcs_data, ipcs.u64.ipcs_datalen);
1126 break;
1127
1128 case IPCS_SHM_ITER: /* Iterate over existing segments */
1129 cursor = ipcs.u64.ipcs_cursor;
1130 if (cursor < 0 || cursor >= shminfo.shmmni) {
1131 error = ERANGE;
1132 break;
1133 }
1134 if (ipcs.u64.ipcs_datalen != (int)shmid_ds_sz) {
2d21ac55 1135 error = EINVAL;
91447636
A
1136 break;
1137 }
1138 for( ; cursor < shminfo.shmmni; cursor++) {
2d21ac55 1139 if (shmsegs[cursor].u.shm_perm.mode & SHMSEG_ALLOCATED)
91447636
A
1140 break;
1141 continue;
1142 }
1143 if (cursor == shminfo.shmmni) {
1144 error = ENOENT;
1145 break;
1146 }
1147
1148 shmid_dsp = &shmsegs[cursor]; /* default: 64 bit */
1149
1150 /*
1151 * If necessary, convert the 64 bit kernel segment
1152 * descriptor to a 32 bit user one.
1153 */
1154 if (!IS_64BIT_PROCESS(p)) {
1155 shmid_ds_64to32(shmid_dsp, &shmid_ds32);
22ba694c
A
1156
1157 /* Clear kernel reserved pointer before copying to user space */
1158 shmid_ds32.shm_internal = (user32_addr_t)0;
1159
91447636 1160 shmid_dsp = &shmid_ds32;
22ba694c
A
1161 } else {
1162 memcpy(&shmid_ds, shmid_dsp, sizeof(shmid_ds));
1163
1164 /* Clear kernel reserved pointer before copying to user space */
1165 shmid_ds.shm_internal = USER_ADDR_NULL;
1166
1167 shmid_dsp = &shmid_ds;
91447636
A
1168 }
1169 error = copyout(shmid_dsp, ipcs.u64.ipcs_data, ipcs.u64.ipcs_datalen);
1170 if (!error) {
1171 /* update cursor */
1172 ipcs.u64.ipcs_cursor = cursor + 1;
1173
1174 if (!IS_64BIT_PROCESS(p)) /* convert in place */
b0d623f7
A
1175 ipcs.u32.ipcs_data = CAST_DOWN_EXPLICIT(user32_addr_t,ipcs.u64.ipcs_data);
1176
1177 error = SYSCTL_OUT(req, &ipcs, ipcs_sz);
91447636
A
1178 }
1179 break;
1180
1181 default:
1182 error = EINVAL;
1183 break;
1184 }
1185ipcs_shm_sysctl_out:
1186 SYSV_SHM_SUBSYS_UNLOCK();
1187 return(error);
1c79356b 1188}
9bccf70c 1189
6d2010ae 1190SYSCTL_NODE(_kern, KERN_SYSV, sysv, CTLFLAG_RW | CTLFLAG_LOCKED | CTLFLAG_ANYBODY, 0, "SYSV");
9bccf70c 1191
6d2010ae 1192SYSCTL_PROC(_kern_sysv, OID_AUTO, shmmax, CTLTYPE_QUAD | CTLFLAG_RW | CTLFLAG_LOCKED,
91447636 1193 &shminfo.shmmax, 0, &sysctl_shminfo ,"Q","shmmax");
9bccf70c 1194
6d2010ae 1195SYSCTL_PROC(_kern_sysv, OID_AUTO, shmmin, CTLTYPE_QUAD | CTLFLAG_RW | CTLFLAG_LOCKED,
91447636 1196 &shminfo.shmmin, 0, &sysctl_shminfo ,"Q","shmmin");
9bccf70c 1197
6d2010ae 1198SYSCTL_PROC(_kern_sysv, OID_AUTO, shmmni, CTLTYPE_QUAD | CTLFLAG_RW | CTLFLAG_LOCKED,
91447636 1199 &shminfo.shmmni, 0, &sysctl_shminfo ,"Q","shmmni");
9bccf70c 1200
6d2010ae 1201SYSCTL_PROC(_kern_sysv, OID_AUTO, shmseg, CTLTYPE_QUAD | CTLFLAG_RW | CTLFLAG_LOCKED,
91447636 1202 &shminfo.shmseg, 0, &sysctl_shminfo ,"Q","shmseg");
9bccf70c 1203
6d2010ae 1204SYSCTL_PROC(_kern_sysv, OID_AUTO, shmall, CTLTYPE_QUAD | CTLFLAG_RW | CTLFLAG_LOCKED,
91447636 1205 &shminfo.shmall, 0, &sysctl_shminfo ,"Q","shmall");
9bccf70c 1206
6d2010ae 1207SYSCTL_NODE(_kern_sysv, OID_AUTO, ipcs, CTLFLAG_RW | CTLFLAG_LOCKED | CTLFLAG_ANYBODY, 0, "SYSVIPCS");
9bccf70c 1208
6d2010ae 1209SYSCTL_PROC(_kern_sysv_ipcs, OID_AUTO, shm, CTLFLAG_RW | CTLFLAG_ANYBODY | CTLFLAG_LOCKED,
91447636
A
1210 0, 0, IPCS_shm_sysctl,
1211 "S,IPCS_shm_command",
1212 "ipcs shm command interface");
2d21ac55
A
1213#endif /* SYSV_SHM */
1214
1215/* DSEP Review Done pl-20051108-v02 @2743,@2908,@2913,@3009 */