/* mmap.cc Copyright 1996, 1997, 1998, 2000, 2001, 2002, 2003, 2004, 2005 Red Hat, Inc. This file is part of Cygwin. This software is a copyrighted work licensed under the terms of the Cygwin license. Please consult the file "CYGWIN_LICENSE" for details. */ #include "winsup.h" #include #include #include #include #include "cygerrno.h" #include "security.h" #include "path.h" #include "fhandler.h" #include "dtable.h" #include "cygheap.h" #include "pinfo.h" #include "sys/cygwin.h" #define PAGE_CNT(bytes) howmany((bytes),getpagesize()) #define PGBITS (sizeof (DWORD)*8) #define MAPSIZE(pages) howmany ((pages), PGBITS) #define MAP_SET(n) (page_map_[(n)/PGBITS] |= (1L << ((n) % PGBITS))) #define MAP_CLR(n) (page_map_[(n)/PGBITS] &= ~(1L << ((n) % PGBITS))) #define MAP_ISSET(n) (page_map_[(n)/PGBITS] & (1L << ((n) % PGBITS))) /* Used for accessing the page file (anonymous mmaps). */ static fhandler_disk_file fh_paging_file; /* Class structure used to keep a record of all current mmap areas in a process. Needed for bookkeeping all mmaps in a process and for duplicating all mmaps after fork() since mmaps are not propagated to child processes by Windows. All information must be duplicated by hand, see fixup_mmaps_after_fork(). The class structure: One member of class map per process, global variable mmapped_areas. Contains a dynamic class list array. Each list entry represents all mapping to a file, keyed by file descriptor and file name hash. Each list entry contains a dynamic class mmap_record array. Each mmap_record represents exactly one mapping. For each mapping, there's an additional so called `page_map'. It's an array of bits, one bit per mapped memory page. The bit is set if the page is accessible, unset otherwise. */ class mmap_record { private: int fdesc_; HANDLE mapping_handle_; DWORD access_mode_; _off64_t offset_; DWORD size_to_map_; caddr_t base_address_; DWORD *page_map_; device dev; public: mmap_record (int fd, HANDLE h, DWORD ac, _off64_t o, DWORD s, caddr_t b) : fdesc_ (fd), mapping_handle_ (h), access_mode_ (ac), offset_ (o), size_to_map_ (s), base_address_ (b), page_map_ (NULL) { dev.devn = 0; if (fd >= 0 && !cygheap->fdtab.not_open (fd)) dev = cygheap->fdtab[fd]->dev (); } int get_fd () const { return fdesc_; } HANDLE get_handle () const { return mapping_handle_; } device& get_device () { return dev; } DWORD get_access () const { return access_mode_; } _off64_t get_offset () const { return offset_; } DWORD get_size () const { return size_to_map_; } caddr_t get_address () const { return base_address_; } bool alloc_page_map (_off64_t off, DWORD len); void free_page_map () { if (page_map_) cfree (page_map_); } void fixup_page_map (void); DWORD find_unused_pages (DWORD pages); _off64_t map_pages (_off64_t off, DWORD len); bool map_pages (caddr_t addr, DWORD len); bool unmap_pages (caddr_t addr, DWORD len); int access (caddr_t address); fhandler_base *alloc_fh (); void free_fh (fhandler_base *fh); }; class list { private: mmap_record *recs; int nrecs, maxrecs; int fd; DWORD hash; public: int get_fd () const { return fd; } DWORD get_hash () const { return hash; } mmap_record *get_record (int i) { return i >= nrecs ? NULL : recs + i; } void set (int nfd); mmap_record *add_record (mmap_record r, _off64_t off, DWORD len); bool del_record (int i); void free_recs () { if (recs) cfree (recs); } mmap_record *search_record (_off64_t off, DWORD len); long search_record (caddr_t addr, DWORD len, caddr_t &m_addr, DWORD &m_len, long start); }; class map { private: list *lists; int nlists, maxlists; public: list *get_list (int i) { return i >= nlists ? NULL : lists + i; } list *get_list_by_fd (int fd); list *add_list (int fd); void del_list (int i); }; /* This is the global map structure pointer. It's allocated once on the first call to mmap64(). */ static map *mmapped_areas; DWORD mmap_record::find_unused_pages (DWORD pages) { DWORD mapped_pages = PAGE_CNT (size_to_map_); DWORD start; if (pages > mapped_pages) return (DWORD)-1; for (start = 0; start <= mapped_pages - pages; ++start) if (!MAP_ISSET (start)) { DWORD cnt; for (cnt = 0; cnt < pages; ++cnt) if (MAP_ISSET (start + cnt)) break; if (cnt >= pages) return start; } return (DWORD)-1; } bool mmap_record::alloc_page_map (_off64_t off, DWORD len) { /* Allocate one bit per page */ if (!(page_map_ = (DWORD *) ccalloc (HEAP_MMAP, MAPSIZE (PAGE_CNT (size_to_map_)), sizeof (DWORD)))) return false; off -= offset_; len = PAGE_CNT (len); if (wincap.virtual_protect_works_on_shared_pages ()) { DWORD old_prot; DWORD vlen = len * getpagesize (); if (off > 0 && !VirtualProtect (base_address_, off, PAGE_NOACCESS, &old_prot)) syscall_printf ("VirtualProtect(%x,%D) failed, %E", base_address_, off); if (off + vlen < size_to_map_ && !VirtualProtect (base_address_ + off + vlen, size_to_map_ - vlen - off, PAGE_NOACCESS, &old_prot)) syscall_printf ("VirtualProtect(%x,%D) failed, %E", base_address_ + off + vlen, size_to_map_ - vlen - off); } off /= getpagesize (); while (len-- > 0) MAP_SET (off + len); return true; } _off64_t mmap_record::map_pages (_off64_t off, DWORD len) { /* Used ONLY if this mapping matches into the chunk of another already performed mapping in a special case of MAP_ANON|MAP_PRIVATE. Otherwise it's job is now done by alloc_page_map(). */ DWORD prot, old_prot; switch (access_mode_) { case FILE_MAP_WRITE: prot = PAGE_READWRITE; break; case FILE_MAP_READ: prot = PAGE_READONLY; break; default: prot = PAGE_WRITECOPY; break; } debug_printf ("map_pages (fd=%d, off=%D, len=%u)", fdesc_, off, len); len = PAGE_CNT (len); if ((off = find_unused_pages (len)) == (DWORD)-1) return 0L; if (wincap.virtual_protect_works_on_shared_pages () && !VirtualProtect (base_address_ + off * getpagesize (), len * getpagesize (), prot, &old_prot)) { __seterrno (); return (_off64_t)-1; } while (len-- > 0) MAP_SET (off + len); return off * getpagesize (); } bool mmap_record::map_pages (caddr_t addr, DWORD len) { debug_printf ("map_pages (addr=%x, len=%u)", addr, len); DWORD prot, old_prot; DWORD off = addr - base_address_; off /= getpagesize (); len = PAGE_CNT (len); /* First check if the area is unused right now. */ for (DWORD l = 0; l < len; ++l) if (MAP_ISSET (off + l)) { set_errno (EINVAL); return false; } switch (access_mode_) { case FILE_MAP_WRITE: prot = PAGE_READWRITE; break; case FILE_MAP_READ: prot = PAGE_READONLY; break; default: prot = PAGE_WRITECOPY; break; } if (wincap.virtual_protect_works_on_shared_pages () && !VirtualProtect (base_address_ + off * getpagesize (), len * getpagesize (), prot, &old_prot)) { __seterrno (); return false; } for (; len-- > 0; ++off) MAP_SET (off); return true; } bool mmap_record::unmap_pages (caddr_t addr, DWORD len) { DWORD old_prot; DWORD off = addr - base_address_; off /= getpagesize (); len = PAGE_CNT (len); if (wincap.virtual_protect_works_on_shared_pages () && !VirtualProtect (base_address_ + off * getpagesize (), len * getpagesize (), PAGE_NOACCESS, &old_prot)) syscall_printf ("-1 = unmap_pages (), %E"); for (; len-- > 0; ++off) MAP_CLR (off); /* Return TRUE if all pages are free'd which may result in unmapping the whole chunk. */ for (len = MAPSIZE (PAGE_CNT (size_to_map_)); len > 0; ) if (page_map_[--len]) return false; return true; } void mmap_record::fixup_page_map () { if (!wincap.virtual_protect_works_on_shared_pages ()) return; DWORD prot, old_prot; switch (access_mode_) { case FILE_MAP_WRITE: prot = PAGE_READWRITE; break; case FILE_MAP_READ: prot = PAGE_READONLY; break; default: prot = PAGE_WRITECOPY; break; } for (DWORD off = PAGE_CNT (size_to_map_); off > 0; --off) VirtualProtect (base_address_ + (off - 1) * getpagesize (), getpagesize (), MAP_ISSET (off - 1) ? prot : PAGE_NOACCESS, &old_prot); } int mmap_record::access (caddr_t address) { if (address < base_address_ || address >= base_address_ + size_to_map_) return 0; DWORD off = (address - base_address_) / getpagesize (); return MAP_ISSET (off); } fhandler_base * mmap_record::alloc_fh () { if (get_fd () == -1) { fh_paging_file.set_io_handle (INVALID_HANDLE_VALUE); return &fh_paging_file; } /* The file descriptor could have been closed or, even worse, could have been reused for another file before the call to fork(). This requires creating a fhandler of the correct type to be sure to call the method of the correct class. */ return build_fh_dev (get_device ()); } void mmap_record::free_fh (fhandler_base *fh) { if (get_fd () != -1) cfree (fh); } mmap_record * list::add_record (mmap_record r, _off64_t off, DWORD len) { if (nrecs == maxrecs) { mmap_record *new_recs; if (maxrecs == 0) new_recs = (mmap_record *) cmalloc (HEAP_MMAP, 5 * sizeof (mmap_record)); else new_recs = (mmap_record *) crealloc (recs, (maxrecs + 5) * sizeof (mmap_record)); if (!new_recs) return NULL; maxrecs += 5; recs = new_recs; } recs[nrecs] = r; if (!recs[nrecs].alloc_page_map (off, len)) return NULL; return recs + nrecs++; } /* Used in mmap() */ mmap_record * list::search_record (_off64_t off, DWORD len) { if (fd == -1 && !off) { len = PAGE_CNT (len); for (int i = 0; i < nrecs; ++i) if (recs[i].find_unused_pages (len) != (DWORD)-1) return recs + i; } else { for (int i = 0; i < nrecs; ++i) if (off >= recs[i].get_offset () && off + len <= recs[i].get_offset () + (PAGE_CNT (recs[i].get_size ()) * getpagesize ())) return recs + i; } return NULL; } /* Used in munmap() */ long list::search_record (caddr_t addr, DWORD len, caddr_t &m_addr, DWORD &m_len, long start) { caddr_t low, high; for (long i = start + 1; i < nrecs; ++i) { low = (addr >= recs[i].get_address ()) ? addr : recs[i].get_address (); high = recs[i].get_address () + (PAGE_CNT (recs[i].get_size ()) * getpagesize ()); high = (addr + len < high) ? addr + len : high; if (low < high) { m_addr = low; m_len = high - low; return i; } } return -1; } void list::set (int nfd) { if ((fd = nfd) != -1) hash = cygheap->fdtab[fd]->get_namehash (); nrecs = maxrecs = 0; recs = NULL; } bool list::del_record (int i) { if (i < nrecs) { recs[i].free_page_map (); for (; i < nrecs - 1; i++) recs[i] = recs[i + 1]; nrecs--; } /* Return true if the list is empty which allows the caller to remove this list from the list array. */ return !nrecs; } list * map::get_list_by_fd (int fd) { int i; for (i=0; ifdtab[fd]->get_namehash ())) return lists + i; return 0; } list * map::add_list (int fd) { if (nlists == maxlists) { list *new_lists; if (maxlists == 0) new_lists = (list *) cmalloc (HEAP_MMAP, 5 * sizeof (list)); else new_lists = (list *) crealloc (lists, (maxlists + 5) * sizeof (list)); if (!new_lists) return NULL; maxlists += 5; lists = new_lists; } lists[nlists].set (fd); return lists + nlists++; } void map::del_list (int i) { if (i < nlists) { lists[i].free_recs (); for (; i < nlists - 1; i++) lists[i] = lists[i + 1]; nlists--; } } extern "C" void * mmap64 (void *addr, size_t len, int prot, int flags, int fd, _off64_t off) { syscall_printf ("addr %x, len %u, prot %x, flags %x, fd %d, off %D", addr, len, prot, flags, fd, off); static DWORD granularity = getshmlba (); /* Error conditions according to SUSv2 */ if (off % getpagesize () || (!(flags & MAP_SHARED) && !(flags & MAP_PRIVATE)) || ((flags & MAP_SHARED) && (flags & MAP_PRIVATE)) || ((flags & MAP_FIXED) && ((DWORD)addr % getpagesize ())) || !len) { set_errno (EINVAL); syscall_printf ("-1 = mmap(): EINVAL"); return MAP_FAILED; } SetResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); if (mmapped_areas == NULL) { /* First mmap call, create STL map */ mmapped_areas = (map *) ccalloc (HEAP_MMAP, 1, sizeof (map)); if (mmapped_areas == NULL) { set_errno (ENOMEM); syscall_printf ("-1 = mmap(): ENOMEM"); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return MAP_FAILED; } } if (flags & MAP_ANONYMOUS) fd = -1; /* If MAP_FIXED is requested on a non-granularity boundary, change request so that this looks like a request with offset addr % granularity. */ if (fd == -1 && (flags & MAP_FIXED) && ((DWORD)addr % granularity) && !off) off = (DWORD)addr % granularity; /* Map always in multipliers of `granularity'-sized chunks. */ _off64_t gran_off = off & ~(granularity - 1); DWORD gran_len = howmany (off + len, granularity) * granularity - gran_off; fhandler_base *fh; if (fd != -1) { /* Ensure that fd is open */ cygheap_fdget cfd (fd); if (cfd < 0) { syscall_printf ("-1 = mmap(): EBADF"); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return MAP_FAILED; } fh = cfd; if (fh->get_device () == FH_FS) { DWORD high; DWORD low = GetFileSize (fh->get_handle (), &high); _off64_t fsiz = ((_off64_t)high << 32) + low; /* Don't allow mappings beginning beyond EOF since Windows can't handle that POSIX like. FIXME: Still looking for a good idea to allow that nevertheless. */ if (gran_off >= fsiz) { set_errno (ENXIO); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return MAP_FAILED; } fsiz -= gran_off; if (gran_len > fsiz) gran_len = fsiz; } else if (fh->get_device () == FH_ZERO) /* mmap /dev/zero is like MAP_ANONYMOUS. */ fd = -1; } if (fd == -1) { fh_paging_file.set_io_handle (INVALID_HANDLE_VALUE); fh = &fh_paging_file; } DWORD access = (prot & PROT_WRITE) ? FILE_MAP_WRITE : FILE_MAP_READ; /* copy-on-write doesn't work at all on 9x using anonymous maps. Workaround: Anonymous mappings always use normal READ or WRITE access and don't use named file mapping. copy-on-write doesn't also work properly on 9x with real files. While the changes are not propagated to the file, they are visible to other processes sharing the same file mapping object. Workaround: Don't use named file mapping. That should work since sharing file mappings only works reliable using named file mapping on 9x. */ if ((flags & MAP_PRIVATE) && (wincap.has_working_copy_on_write () || fd != -1)) access = FILE_MAP_COPY; list *map_list = mmapped_areas->get_list_by_fd (fd); /* First check if this mapping matches into the chunk of another already performed mapping. Only valid for MAP_ANON in a special case of MAP_PRIVATE. */ if (map_list && fd == -1 && off == 0 && !(flags & MAP_FIXED)) { mmap_record *rec; if ((rec = map_list->search_record (off, len)) != NULL && rec->get_access () == access) { if ((off = rec->map_pages (off, len)) == (_off64_t)-1) { syscall_printf ("-1 = mmap()"); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK|WRITE_LOCK, "mmap"); return MAP_FAILED; } caddr_t ret = rec->get_address () + off; syscall_printf ("%x = mmap() succeeded", ret); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return ret; } } if (map_list && fd == -1 && (flags & MAP_FIXED)) { caddr_t u_addr; DWORD u_len; long record_idx = -1; if ((record_idx = map_list->search_record ((caddr_t)addr, len, u_addr, u_len, record_idx)) >= 0) { mmap_record *rec = map_list->get_record (record_idx); if (u_addr > (caddr_t)addr || u_addr + len < (caddr_t)addr + len || rec->get_access () != access) { /* Partial match only, or access mode doesn't match. */ /* FIXME: Handle partial mappings gracefully if adjacent memory is available. */ set_errno (EINVAL); syscall_printf ("-1 = mmap()"); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return MAP_FAILED; } if (!rec->map_pages ((caddr_t)addr, len)) { syscall_printf ("-1 = mmap()"); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return MAP_FAILED; } caddr_t ret = (caddr_t)addr; syscall_printf ("%x = mmap() succeeded", ret); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return ret; } } caddr_t base = (caddr_t)addr; /* This shifts the base address to the next lower 64K boundary. The offset is re-added when evaluating the return value. */ if (base) base -= off - gran_off; HANDLE h = fh->mmap (&base, gran_len, access, flags, gran_off); if (h == INVALID_HANDLE_VALUE) { ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return MAP_FAILED; } /* At this point we should have a successfully mmapped area. Now it's time for bookkeeping stuff. */ if (fd == -1) gran_len = PAGE_CNT (gran_len) * getpagesize (); mmap_record mmap_rec (fd, h, access, gran_off, gran_len, base); /* Get list of mmapped areas for this fd, create a new one if one does not exist yet. */ if (!map_list) { /* Create a new one */ map_list = mmapped_areas->add_list (fd); if (!map_list) { fh->munmap (h, base, gran_len); set_errno (ENOMEM); syscall_printf ("-1 = mmap(): ENOMEM"); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return MAP_FAILED; } } /* Insert into the list */ mmap_record *rec = map_list->add_record (mmap_rec, off, len > gran_len ? gran_len : len); if (!rec) { fh->munmap (h, base, gran_len); set_errno (ENOMEM); syscall_printf ("-1 = mmap(): ENOMEM"); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return MAP_FAILED; } caddr_t ret = rec->get_address () + (off - gran_off); syscall_printf ("%x = mmap() succeeded", ret); ReleaseResourceLock (LOCK_MMAP_LIST, READ_LOCK | WRITE_LOCK, "mmap"); return ret; } extern "C" void * mmap (void *addr, size_t len, int prot, int flags, int fd, _off_t off) { return mmap64 (addr, len, prot, flags, fd, (_off64_t)off); } /* munmap () removes all mmapped pages between addr and addr+len. */ extern "C" int munmap (void *addr, size_t len) { syscall_printf ("munmap (addr %x, len %u)", addr, len); /* Error conditions according to SUSv3 */ if (!addr || ((DWORD)addr % getpagesize ()) || !len || check_invalid_virtual_addr (addr, len)) { set_errno (EINVAL); syscall_printf ("-1 = munmap(): Invalid parameters"); return -1; } SetResourceLock (LOCK_MMAP_LIST, WRITE_LOCK | READ_LOCK, "munmap"); if (mmapped_areas == NULL) { syscall_printf ("-1 = munmap(): mmapped_areas == NULL"); ReleaseResourceLock (LOCK_MMAP_LIST, WRITE_LOCK | READ_LOCK, "munmap"); return 0; } /* Iterate through the map, unmap pages between addr and addr+len in all maps. */ list *map_list; for (int list_idx = 0; (map_list = mmapped_areas->get_list (list_idx)); ++list_idx) { long record_idx = -1; caddr_t u_addr; DWORD u_len; while ((record_idx = map_list->search_record((caddr_t)addr, len, u_addr, u_len, record_idx)) >= 0) { mmap_record *rec = map_list->get_record (record_idx); if (rec->unmap_pages (u_addr, u_len)) { /* The whole record has been unmapped, so we now actually unmap it from the system in full length... */ fhandler_base *fh = rec->alloc_fh (); fh->munmap (rec->get_handle (), rec->get_address (), rec->get_size ()); rec->free_fh (fh); /* ...and delete the record. */ if (map_list->del_record (record_idx--)) { /* Yay, the last record has been removed from the list, we can remove the list now, too. */ mmapped_areas->del_list (list_idx--); break; } } } } ReleaseResourceLock (LOCK_MMAP_LIST, WRITE_LOCK | READ_LOCK, "munmap"); syscall_printf ("0 = munmap(): %x", addr); return 0; } /* Sync file with memory. Ignore flags for now. */ extern "C" int msync (void *addr, size_t len, int flags) { syscall_printf ("addr %x, len %u, flags %x", addr, len, flags); /* However, check flags for validity. */ if ((flags & ~(MS_ASYNC | MS_SYNC | MS_INVALIDATE)) || ((flags & MS_ASYNC) && (flags & MS_SYNC))) { syscall_printf ("-1 = msync(): Invalid flags"); set_errno (EINVAL); return -1; } SetResourceLock (LOCK_MMAP_LIST, WRITE_LOCK | READ_LOCK, "msync"); /* Check if a mmap'ed area was ever created */ if (mmapped_areas == NULL) { syscall_printf ("-1 = msync(): mmapped_areas == NULL"); set_errno (EINVAL); ReleaseResourceLock (LOCK_MMAP_LIST, WRITE_LOCK | READ_LOCK, "msync"); return -1; } /* Iterate through the map, looking for the mmapped area. Error if not found. */ list *map_list; for (int list_idx = 0; (map_list = mmapped_areas->get_list (list_idx)); ++list_idx) { mmap_record *rec; for (int record_idx = 0; (rec = map_list->get_record (record_idx)); ++record_idx) { if (rec->access ((caddr_t)addr)) { /* Check whole area given by len. */ for (DWORD i = getpagesize (); i < len; ++i) if (!rec->access ((caddr_t)addr + i)) goto invalid_address_range; fhandler_base *fh = rec->alloc_fh (); int ret = fh->msync (rec->get_handle (), (caddr_t)addr, len, flags); rec->free_fh (fh); if (ret) syscall_printf ("%d = msync(), %E", ret); else syscall_printf ("0 = msync()"); ReleaseResourceLock (LOCK_MMAP_LIST, WRITE_LOCK | READ_LOCK, "msync"); return 0; } } } invalid_address_range: /* SUSv2: Return code if indicated memory was not mapped is ENOMEM. */ set_errno (ENOMEM); syscall_printf ("-1 = msync(): ENOMEM"); ReleaseResourceLock (LOCK_MMAP_LIST, WRITE_LOCK | READ_LOCK, "msync"); return -1; } /* Set memory protection */ extern "C" int mprotect (void *addr, size_t len, int prot) { DWORD old_prot; DWORD new_prot = 0; syscall_printf ("mprotect (addr %x, len %u, prot %x)", addr, len, prot); if (!wincap.virtual_protect_works_on_shared_pages () && addr >= (caddr_t)0x80000000 && addr <= (caddr_t)0xBFFFFFFF) { syscall_printf ("0 = mprotect (9x: No VirtualProtect on shared memory)"); return 0; } /* If write protection is requested, check if the page was originally protected writecopy. In this case call VirtualProtect requesting PAGE_WRITECOPY, otherwise the VirtualProtect will fail on NT version >= 5.0 */ bool writecopy = false; if (prot & PROT_WRITE) { MEMORY_BASIC_INFORMATION mbi; if (VirtualQuery (addr, &mbi, sizeof mbi)) { if (mbi.AllocationProtect == PAGE_WRITECOPY || mbi.AllocationProtect == PAGE_EXECUTE_WRITECOPY) writecopy = true; } } switch (prot) { case PROT_READ | PROT_WRITE | PROT_EXEC: case PROT_WRITE | PROT_EXEC: new_prot = writecopy ? PAGE_EXECUTE_WRITECOPY : PAGE_EXECUTE_READWRITE; break; case PROT_READ | PROT_WRITE: case PROT_WRITE: new_prot = writecopy ? PAGE_WRITECOPY : PAGE_READWRITE; break; case PROT_READ | PROT_EXEC: new_prot = PAGE_EXECUTE_READ; break; case PROT_READ: new_prot = PAGE_READONLY; break; case PROT_EXEC: new_prot = PAGE_EXECUTE; break; case PROT_NONE: new_prot = PAGE_NOACCESS; break; default: syscall_printf ("-1 = mprotect (): invalid prot value"); set_errno (EINVAL); return -1; } if (VirtualProtect (addr, len, new_prot, &old_prot) == 0) { __seterrno (); syscall_printf ("-1 = mprotect (), %E"); return -1; } syscall_printf ("0 = mprotect ()"); return 0; } /* * Base implementation: * * `mmap' returns ENODEV as documented in SUSv2. * In contrast to the global function implementation, the member function * `mmap' has to return the mapped base address in `addr' and the handle to * the mapping object as return value. In case of failure, the fhandler * mmap has to close that handle by itself and return INVALID_HANDLE_VALUE. * * `munmap' and `msync' get the handle to the mapping object as first parameter * additionally. */ HANDLE fhandler_base::mmap (caddr_t *addr, size_t len, DWORD access, int flags, _off64_t off) { set_errno (ENODEV); return INVALID_HANDLE_VALUE; } int fhandler_base::munmap (HANDLE h, caddr_t addr, size_t len) { set_errno (ENODEV); return -1; } int fhandler_base::msync (HANDLE h, caddr_t addr, size_t len, int flags) { set_errno (ENODEV); return -1; } bool fhandler_base::fixup_mmap_after_fork (HANDLE h, DWORD access, DWORD offset, DWORD size, void *address) { set_errno (ENODEV); return -1; } /* Implementation for disk files. */ HANDLE fhandler_disk_file::mmap (caddr_t *addr, size_t len, DWORD access, int flags, _off64_t off) { DWORD protect; switch (access) { case FILE_MAP_WRITE: protect = PAGE_READWRITE; break; case FILE_MAP_READ: protect = PAGE_READONLY; break; default: protect = PAGE_WRITECOPY; break; } HANDLE h; /* On 9x/ME try first to open the mapping by name when opening a shared file object. This is needed since 9x/ME only shares objects between processes by name. What a mess... */ if (wincap.share_mmaps_only_by_name () && get_handle () != INVALID_HANDLE_VALUE && !(access & FILE_MAP_COPY)) { /* Grrr, the whole stuff is just needed to try to get a reliable mapping of the same file. Even that uprising isn't bullet proof but it does it's best... */ char namebuf[CYG_MAX_PATH]; cygwin_conv_to_full_posix_path (get_name (), namebuf); for (int i = strlen (namebuf) - 1; i >= 0; --i) namebuf[i] = cyg_tolower (namebuf [i]); debug_printf ("named sharing"); if (!(h = OpenFileMapping (access, TRUE, namebuf))) h = CreateFileMapping (get_handle (), &sec_none, protect, 0, 0, namebuf); } else h = CreateFileMapping (get_handle (), &sec_none, protect, 0, get_handle () == INVALID_HANDLE_VALUE ? len : 0, NULL); if (!h) { __seterrno (); syscall_printf ("-1 = mmap(): CreateFileMapping failed with %E"); return INVALID_HANDLE_VALUE; } DWORD high = off >> 32, low = off & UINT32_MAX; void *base = NULL; /* If a non-zero address is given, try mapping using the given address first. If it fails and flags is not MAP_FIXED, try again with NULL address. */ if (*addr) base = MapViewOfFileEx (h, access, high, low, len, *addr); if (!base && !(flags & MAP_FIXED)) base = MapViewOfFileEx (h, access, high, low, len, NULL); debug_printf ("%x = MapViewOfFileEx (h:%x, access:%x, 0, off:%D, " "len:%u, addr:%x)", base, h, access, off, len, *addr); if (!base || ((flags & MAP_FIXED) && base != *addr)) { if (!base) { __seterrno (); syscall_printf ("-1 = mmap(): MapViewOfFileEx failed with %E"); } else { set_errno (EINVAL); syscall_printf ("-1 = mmap(): address shift with MAP_FIXED given"); } CloseHandle (h); return INVALID_HANDLE_VALUE; } *addr = (caddr_t) base; return h; } int fhandler_disk_file::munmap (HANDLE h, caddr_t addr, size_t len) { UnmapViewOfFile (addr); CloseHandle (h); return 0; } int fhandler_disk_file::msync (HANDLE h, caddr_t addr, size_t len, int flags) { if (FlushViewOfFile (addr, len) == 0) { __seterrno (); return -1; } return 0; } bool fhandler_disk_file::fixup_mmap_after_fork (HANDLE h, DWORD access, DWORD offset, DWORD size, void *address) { /* Re-create the MapViewOfFileEx call */ void *base = MapViewOfFileEx (h, access, 0, offset, size, address); if (base != address) { MEMORY_BASIC_INFORMATION m; (void) VirtualQuery (address, &m, sizeof (m)); system_printf ("requested %p != %p mem alloc base %p, state %p, size %d, %E", address, base, m.AllocationBase, m.State, m.RegionSize); } return base == address; } /* * Call to re-create all the file mappings in a forked * child. Called from the child in initialization. At this * point we are passed a valid mmapped_areas map, and all the * HANDLE's are valid for the child, but none of the * mapped areas are in our address space. We need to iterate * through the map, doing the MapViewOfFile calls. */ int __stdcall fixup_mmaps_after_fork (HANDLE parent) { debug_printf ("recreate_mmaps_after_fork, mmapped_areas %p", mmapped_areas); /* Check if a mmapped area was ever created */ if (mmapped_areas == NULL) return 0; /* Iterate through the map */ list *map_list; for (int list_idx = 0; (map_list = mmapped_areas->get_list (list_idx)); ++list_idx) { mmap_record *rec; for (int record_idx = 0; (rec = map_list->get_record (record_idx)); ++record_idx) { debug_printf ("fd %d, h %x, access %x, offset %D, size %u, address %p", rec->get_fd (), rec->get_handle (), rec->get_access (), rec->get_offset (), rec->get_size (), rec->get_address ()); fhandler_base *fh = rec->alloc_fh (); bool ret = fh->fixup_mmap_after_fork (rec->get_handle (), rec->get_access (), rec->get_offset (), rec->get_size (), rec->get_address ()); rec->free_fh (fh); if (!ret) return -1; if (rec->get_access () == FILE_MAP_COPY) { for (char *address = rec->get_address (); address < rec->get_address () + rec->get_size (); address += getpagesize ()) if (rec->access (address) && !ReadProcessMemory (parent, address, address, getpagesize (), NULL)) { DWORD old_prot; DWORD last_error = GetLastError (); if (last_error != ERROR_PARTIAL_COPY && last_error != ERROR_NOACCESS || !wincap.virtual_protect_works_on_shared_pages ()) { system_printf ("ReadProcessMemory failed for " "MAP_PRIVATE address %p, %E", rec->get_address ()); return -1; } if (!VirtualProtectEx (parent, address, getpagesize (), PAGE_READONLY, &old_prot)) { system_printf ("VirtualProtectEx failed for " "MAP_PRIVATE address %p, %E", rec->get_address ()); return -1; } else { BOOL ret; DWORD dummy_prot; ret = ReadProcessMemory (parent, address, address, getpagesize (), NULL); if (!VirtualProtectEx(parent, address, getpagesize (), old_prot, &dummy_prot)) system_printf ("WARNING: VirtualProtectEx to " "return to previous state " "in parent failed for " "MAP_PRIVATE address %p, %E", rec->get_address ()); if (!VirtualProtect (address, getpagesize (), old_prot, &dummy_prot)) system_printf ("WARNING: VirtualProtect to copy " "protection to child failed for" "MAP_PRIVATE address %p, %E", rec->get_address ()); if (!ret) { system_printf ("ReadProcessMemory (2nd try) " "failed for " "MAP_PRIVATE address %p, %E", rec->get_address ()); return -1; } } } } rec->fixup_page_map (); } } debug_printf ("succeeded"); return 0; }