/***************************************************************************************** Copyright (c) 2002-2004,2005,2006,2007,2009 The UbixOS Project All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: Redistributions of source code must retain the above copyright notice, this list of conditions, the following disclaimer and the list of authors. Redistributions in binary form must reproduce the above copyright notice, this list of conditions, the following disclaimer and the list of authors in the documentation and/or other materials provided with the distribution. Neither the name of the UbixOS Project nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. $Id$ *****************************************************************************************/ #include <vmm/vmm.h> #include <vmm/paging.h> #include <lib/kprintf.h> #include <lib/kmalloc.h> #include <ubixos/types.h> #include <ubixos/kpanic.h> #include <ubixos/sched.h> #include <ubixos/spinlock.h> #include <sys/kern_descrip.h> #include <string.h> #include <assert.h> u_int32_t *kernelPageDirectory = 0x0; static spinLock_t fkpSpinLock = SPIN_LOCK_INITIALIZER; static spinLock_t rmpSpinLock = SPIN_LOCK_INITIALIZER; /***************************************************************************************** Function: int vmm_pagingInit(); Description: This Function Will Initialize The Operating Systems Paging Abilities. Notes: 02/20/2004 - Looked Over Code And Have Approved Its Quality 07/28/3004 - All pages are set for ring-0 only no more user accessable *****************************************************************************************/ int vmm_pagingInit() { u_int32_t i = 0x0; u_int32_t *pageTable = 0x0; /* Allocate A Page Of Memory For Kernels Page Directory */ kernelPageDirectory = (u_int32_t *) vmm_findFreePage(sysID); if (kernelPageDirectory == 0x0) { K_PANIC("Error: vmm_findFreePage Failed"); return (0x1); } /* end if */ /* Clear The Memory To Ensure There Is No Garbage */ //QUESTION Can't we do a memset here? for (i = 0x0; i < PAGE_ENTRIES; i++) { kernelPageDirectory[i] = (u_int32_t)0x0; } /* end for */ /* Allocate a page for the first 4MB of memory */ if ((pageTable = (u_int32_t *) vmm_findFreePage(sysID)) == 0x0) K_PANIC("Error: vmm_findFreePage Failed"); kernelPageDirectory[0x0] = (u_int32_t) ((u_int32_t)pageTable | KERNEL_PAGE_DEFAULT); /* Make Sure The Page Table Is Clean */ memset(pageTable,0x0,0x1000); /* * Map the first 1MB of Memory to the kernel MM space because our kernel starts * at 0x30000 * Do not map page at address 0x0 this is reserved for null... */ for (i = 0x1; i < (PAGE_ENTRIES / 0x4); i++) { pageTable[i] = (u_int32_t) ((i * 0x1000) | KERNEL_PAGE_DEFAULT); } /* end for */ /* * Create page tables for the top 1GB of VM space. This space is set aside * for kernel space and will be shared with each process */ for (i = 768; i < PAGE_ENTRIES; i++) { if ((pageTable = (u_int32_t *) vmm_findFreePage(sysID)) == 0x0) K_PANIC("Error: vmm_findFreePage Failed"); /* Make Sure The Page Table Is Clean */ memset(pageTable,0x0,0x1000); /* Map In The Page Directory */ kernelPageDirectory[i] = (u_int32_t) ((u_int32_t) (pageTable) | KERNEL_PAGE_DEFAULT); } /* end for */ /* Set Up Memory To Be All The Allocated Page Directories */ if ((pageTable = (u_int32_t *) vmm_findFreePage(sysID)) == 0x0) K_PANIC("Error: vmm_findFreePage Failed"); /* Clean Page Table */ memset(pageTable,0x0,0x1000); kernelPageDirectory[767] = ((u_int32_t) pageTable | KERNEL_PAGE_DEFAULT); for (i = 0; i < PAGE_ENTRIES; i++) { pageTable[i] = kernelPageDirectory[i]; } /* end for */ /* Also Set Up Page Directory To Be The The First Page In 0xE0400000 */ pageTable = (u_int32_t *) (kernelPageDirectory[0] & 0xFFFFF000); pageTable[256] = (u_int32_t) ((u_int32_t) (kernelPageDirectory) | KERNEL_PAGE_DEFAULT); /* Now Lets Turn On Paging With This Initial Page Table */ asm volatile( "movl %0,%%eax \n" "movl %%eax,%%cr3 \n" "movl %%cr0,%%eax \n" "orl $0x80010000,%%eax \n" /* Turn on memory protection */ "movl %%eax,%%cr0 \n" : : "d"((u_int32_t *) (kernelPageDirectory)) ); /* Remap The Memory List */ for (i = 0x101000; i <= (0x101000 + (numPages * sizeof(mMap))); i += 0x1000) { if ((vmm_remapPage(i, (vmmMemoryMapAddr + (i - 0x101000)),KERNEL_PAGE_DEFAULT)) == 0x0) K_PANIC("vmmRemapPage failed\n"); } /* end for */ /* Set New Address For Memory Map Since Its Relocation */ vmmMemoryMap = (mMap *) vmmMemoryMapAddr; /* Print information on paging */ kprintf("paging0 - Address: [0x%X], PagingISR Address: [0x%X]\n", kernelPageDirectory, &_vmm_pageFault); /* Return so we know everything went well */ return (0x0); } /* END func */ /***************************************************************************************** Function: int vmmRemapPage(Physical Source,Virtual Destination) Description: This Function Will Remap A Physical Page Into Virtual Space Notes: 07/29/02 - Rewrote This To Work With Our New Paging System 07/30/02 - Changed Address Of Page Tables And Page Directory 07/28/04 - If perms == 0x0 set to PAGE_DEFAULT 07/08/09 - Cleaned up *****************************************************************************************/ int vmm_remapPage(u_int32_t sourceAddr,u_int32_t destAddr,u_int16_t perms) { u_int16_t destPageDirectoryIndex = 0x0; u_int16_t destPageTableIndex = 0x0; u_int32_t *pageDir = 0x0; u_int32_t *pageTable = 0x0; short i = 0x0; #ifdef VMMDEBUG kprintf("vmm_remapPage"); #endif if (sourceAddr == 0x0) K_PANIC("source == 0x0"); if (destAddr == 0x0) K_PANIC("dest == 0x0"); /* Lock this function so to avoid race conditions */ spinLock(&rmpSpinLock); /* Set default permissions if not spedified */ if (perms == 0x0) perms = KERNEL_PAGE_DEFAULT; /* * Set Pointer pageDirectory To Point To The Virtual Mapping Of The Page * Directory */ pageDir = (u_int32_t *)PARENT_PAGEDIR_ADDR; /* Get Index Into The Page Directory */ destPageDirectoryIndex = PDI(destAddr); if ((pageDir[destPageDirectoryIndex] & PAGE_PRESENT) != PAGE_PRESENT) { /* If Page Table Is Non Existant Then Set It Up */ /* UBU Why does the page table need to be user writable? */ pageDir[destPageDirectoryIndex] = (u_int32_t) vmm_findFreePage(_current->id) | PAGE_DEFAULT; /* Also Add It To Virtual Space So We Can Make Changes Later */ pageTable = (u_int32_t *) (PAGE_TABLES_BASE_ADDR + 0x2FF000); pageTable[destPageDirectoryIndex] = pageDir[destPageDirectoryIndex]; /* Reload Page Directory */ asm volatile( "push %eax \n" "mov %cr3,%eax \n" "mov %eax,%cr3 \n" "pop %eax \n" ); pageTable = (u_int32_t *) (PAGE_TABLES_BASE_ADDR + (0x1000 * destPageDirectoryIndex)); /* Zero out the page table */ for (i = 0x0;i < PAGE_ENTRIES;i++) pageTable[i] = 0x0; } /* Set Address To Page Table */ pageTable = (u_int32_t *) (PAGE_TABLES_BASE_ADDR + (0x1000 * destPageDirectoryIndex)); /* Get The Index To The Page Table */ destPageTableIndex = PTI(destAddr); /* If The Page Is Mapped In Free It Before We Remap */ if ((pageTable[destPageTableIndex] & PAGE_PRESENT) == PAGE_PRESENT) { if ((pageTable[destPageTableIndex] & PAGE_STACK) == PAGE_STACK) kprintf("Stack Page: [0x%X]\n",destAddr); if ((pageTable[destPageTableIndex] & PAGE_COW) != PAGE_COW) { kprintf("Page NOT COW\n"); kprintf("Page Present: [0x%X][0x%X]",destAddr,pageTable[destPageTableIndex]); sourceAddr = 0x0; goto rmDone; } else { /* Clear The Page First For Security Reasons */ vmm_freePage(((u_int32_t) pageTable[destPageTableIndex] & 0xFFFFF000)); } } /* Set The Source Address In The Destination */ pageTable[destPageTableIndex] = (u_int32_t) (sourceAddr | perms); /* Reload The Page Table; */ asm volatile( "push %eax \n" "movl %cr3,%eax\n" "movl %eax,%cr3\n" "pop %eax \n" ); rmDone: /* Return */ spinUnlock(&rmpSpinLock); return (sourceAddr); } /* end func */ /************************************************************************ Function: int vmm_zeroVirtualPage(u_int32_t pageAddr); Description: This Will Null Out A Page Of Memory Notes: 07/07/2009 - Changed to void it will always return true 07/08/2009 - renamed and cleaned up function ************************************************************************/ int vmm_zeroVirtualPage(u_int32_t pageAddr) { u_int32_t *src = 0x0; int counter = 0x0; if ((pageAddr & 0xFFF) != 0x0) return(0x1); /* Set Source Pointer To Virtual Page Address */ src = (u_int32_t *) pageAddr; /* Clear Out The Page */ for (counter = 0x0; counter < PAGE_ENTRIES; counter++) { src[counter] = 0x0; } return(0x0); } void *vmm_mapFromTask(pidType pid,u_int32_t baseAddr,u_int32_t size) { kTask_t *child = 0x0; u_int32_t i = 0x0; u_int32_t x = 0x0; u_int32_t y = 0x0; u_int32_t count = ((size+4095)/0x1000); u_int32_t c = 0x0; u_int16_t dI = 0x0; u_int16_t tI = 0x0; u_int32_t offset = 0x0; u_int32_t *childPageDir = (u_int32_t *)0x5A00000; u_int32_t *childPageTable = 0x0; u_int32_t *pageTableSrc = 0x0; offset = baseAddr & 0xFFF; baseAddr &= 0xFFFFF000; child = schedFindTask(pid); //Calculate The Page Table Index And Page Directory Index dI = PDI(baseAddr); tI = PTI(baseAddr); if (vmm_remapPage(child->tss.cr3,0x5A00000,KERNEL_PAGE_DEFAULT) == 0x0) K_PANIC("vmmFailed"); for (i=0;i<0x1000;i++) { if (vmm_remapPage(childPageDir[i],0x5A01000 + (i * 0x1000),KERNEL_PAGE_DEFAULT) == 0x0) K_PANIC("Returned NULL"); } for (x=(_current->oInfo.vmStart/(1024*4096));x<1024;x++) { pageTableSrc = (u_int32_t *)(PAGE_TABLES_BASE_ADDR + (4096*x)); for (y=0;y<1024;y++) { //Loop Through The Page Table Find An UnAllocated Page if ((u_int32_t)pageTableSrc[y] == (u_int32_t)0x0) { if (count > 1) { for (c=0;((c<count) && (y+c < 1024));c++) { if ((u_int32_t)pageTableSrc[y+c] != (u_int32_t)0x0) { c = -1; break; } } if (c != -1) { for (c=0;c<count;c++) { if ((tI + c) >= 0x1000) { dI++; tI = 0-c; } childPageTable = (u_int32_t *)(0x5A01000 + (0x1000 * dI)); if (vmm_remapPage(childPageTable[tI+c],((x*(1024*4096))+((y+c)*4096)),KERNEL_PAGE_DEFAULT) == 0x0) K_PANIC("remap == NULL"); } vmm_unmapPages(0x5A00000,1,1); for (i=0;i<0x1000;i++) { vmm_unmapPages((0x5A01000 + (i*0x1000)),1,1); } return((void *)((x*(1024*4096))+(y*4096)+offset)); } } else { //Map A Physical Page To The Virtual Page childPageTable = (u_int32_t *)(0x5A01000 + (0x1000 * dI)); if (vmm_remapPage(childPageTable[tI],((x*(1024*4096))+(y*4096)),KERNEL_PAGE_DEFAULT) == 0x0) K_PANIC("remap Failed"); //Return The Address Of The Mapped In Memory vmm_unmapPages(0x5A00000,1,1); for (i=0;i<0x1000;i++) { vmm_unmapPages((0x5A01000 + (i*0x1000)),1,1); } return((void *)((x*(1024*4096))+(y*4096)+offset)); } } } } return(0x0); } void *vmm_getFreeMallocPage(u_int16_t count) { int c = 0x0; u_int32_t *pageTableSrc = 0x0; u_int16_t x = 0x0; u_int16_t y = 0x0; spinLock(&fkpSpinLock); /* Lets Search For A Free Page */ for (x = 960; x < 1024; x++) { /* Set Page Table Address */ pageTableSrc = (u_int32_t *) (PAGE_TABLES_BASE_ADDR + (0x1000 * x)); for (y = 0; y < 1024; y++) { /* Loop Through The Page Table Find An UnAllocated Page */ if ((u_int32_t) pageTableSrc[y] == (u_int32_t) 0x0) { if (count > 1) { for (c = 0; c < count; c++) { if (y + c < 1024) { if ((u_int32_t) pageTableSrc[y + c] != (u_int32_t) 0x0) { c = -1; break; } } } if (c != -1) { for (c = 0; c < count; c++) { if (vmm_remapPage((u_int32_t) vmm_findFreePage(sysID), ((x * 0x400000) + ((y + c) * 0x1000)),KERNEL_PAGE_DEFAULT) == 0x0) K_PANIC("remap Failed"); if (vmm_zeroVirtualPage((u_int32_t) ((x * 0x400000) + ((y + c) * 0x1000))) == 0x1) K_PANIC("vmm_zeroVirtualPage failed"); } spinUnlock(&fkpSpinLock); return ((void *)((x * 0x400000) + (y * 0x1000))); } } else { /* Map A Physical Page To The Virtual Page */ if (vmm_remapPage((u_int32_t) vmm_findFreePage(sysID), ((x * 0x400000) + (y * 0x1000)),KERNEL_PAGE_DEFAULT) == 0x0) K_PANIC("Failed"); /* Clear This Page So No Garbage Is There */ if (vmm_zeroVirtualPage((u_int32_t) ((x * 0x400000) + (y * 0x1000))) == 0x1) K_PANIC("vmm_zeroVirtualPage: Failed"); /* Return The Address Of The Newly Allocate Page */ spinUnlock(&fkpSpinLock); return ((void *)((x * 0x400000) + (y * 0x1000))); } } } } /* If No Free Page Was Found Return NULL */ spinUnlock(&fkpSpinLock); return (0x0); } /*! \brief sdf */ int mmap(struct thread *td,struct mmap_args *uap) { vm_offset_t addr = 0x0; char *tmp = 0x0; struct file *fd = 0x0; addr = (vm_offset_t) uap->addr; #ifdef VMMDEBUG if (uap->addr != 0x0) { kprintf("Address hints are not supported yet.\n"); } kprintf("uap->flags: [0x%X]\n",uap->flags); kprintf("uap->addr: [0x%X]\n",uap->addr); kprintf("uap->len: [0x%X]\n",uap->len); kprintf("uap->prot: [0x%X]\n",uap->prot); kprintf("uap->fd: [%i]\n",uap->fd); kprintf("uap->pad: [0x%X]\n",uap->pad); kprintf("uap->pos: [0x%X]\n",uap->pos); #endif if (uap->fd == -1) { /* NEED ROUND PAGE */ td->td_retval[0] = (int)vmmGetFreeVirtualPage(_current->id,(uap->len + 0xFFF)/0x1000,VM_TASK); } else { #ifdef VMMDEBUG kprintf("uap->flags: [0x%X]\n",uap->flags); kprintf("uap->addr: [0x%X]\n",uap->addr); kprintf("uap->len: [0x%X]\n",uap->len); kprintf("uap->prot: [0x%X]\n",uap->prot); kprintf("uap->fd: [%i]\n",uap->fd); kprintf("uap->pad: [0x%X]\n",uap->pad); kprintf("uap->pos: [0x%X]\n",uap->pos); #endif getfd(td,&fd,uap->fd); if (uap->addr == 0x0) tmp = (char *)vmmGetFreeVirtualPage(_current->id,(uap->len + 0xFFF)/0x1000,VM_TASK); else { tmp = (char *)vmmGetFreeVirtualPage_new(_current->id,(uap->len + 0xFFF)/0x1000,VM_TASK,uap->addr); } fd->offset = uap->pos; fread(tmp,uap->len,0x1,fd); td->td_retval[0] = (int)tmp; } return(0x0); } int obreak(struct thread *td,struct obreak_args *uap) { u_int32_t i = 0x0; vm_offset_t old = 0x0; vm_offset_t base = 0x0; vm_offset_t new = 0x0; #ifdef VMMDEBUG kprintf("obreak\n"); kprintf("vm_offset_t: [%i]\n",sizeof(vm_offset_t)); kprintf("nsize: [0x%X]\n",uap->nsize); kprintf("vm_daddr: [0x%X]\n",td->vm_daddr); kprintf("vm_dsize: [0x%X]\n",td->vm_dsize); kprintf("total: [0x%X]\n",td->vm_daddr + td->vm_dsize); #endif new = round_page((vm_offset_t)uap->nsize); base = round_page((vm_offset_t)td->vm_daddr); old = base + ctob(td->vm_dsize); if (new < base) K_PANIC("EINVAL"); if (new > old) { for (i = old;i < new;i+= 0x1000) { if (vmm_remapPage(vmm_findFreePage(_current->id),i,PAGE_DEFAULT) == 0x0) K_PANIC("remap Failed"); /* Clear Page */ } td->vm_dsize += btoc(new - old); } else if (new < old) { kprintf("0x%X < 0x%X\n",new,old); //K_PANIC("new < old"); td->vm_dsize -= btoc(old - new); } return(0x0); } int munmap(struct thread *td,struct munmap_args *uap) { /* HACK */ //kprintf("munmap: [0x%X:0x%X]",uap->addr,uap->len); return(0x0); } /*** END ***/