diff options
author | Suren A. Chilingaryan <csa@dside.dyndns.org> | 2011-02-13 03:07:11 +0100 |
---|---|---|
committer | Suren A. Chilingaryan <csa@dside.dyndns.org> | 2011-02-13 03:07:11 +0100 |
commit | b72e8567723c30f0e5a83675743e9380c48e5d55 (patch) | |
tree | f0dc052508543771fca08c358836507770864249 /driver/kmem.c | |
download | ipecamera-b72e8567723c30f0e5a83675743e9380c48e5d55.tar.gz ipecamera-b72e8567723c30f0e5a83675743e9380c48e5d55.tar.bz2 ipecamera-b72e8567723c30f0e5a83675743e9380c48e5d55.tar.xz ipecamera-b72e8567723c30f0e5a83675743e9380c48e5d55.zip |
Initial import
Diffstat (limited to 'driver/kmem.c')
-rw-r--r-- | driver/kmem.c | 321 |
1 files changed, 321 insertions, 0 deletions
diff --git a/driver/kmem.c b/driver/kmem.c new file mode 100644 index 0000000..737b74d --- /dev/null +++ b/driver/kmem.c @@ -0,0 +1,321 @@ +/** + * + * @file kmem.c + * @brief This file contains all functions dealing with kernel memory. + * @author Guillermo Marcus + * @date 2009-04-05 + * + */ +#include <linux/version.h> +#include <linux/string.h> +#include <linux/types.h> +#include <linux/list.h> +#include <linux/interrupt.h> +#include <linux/pci.h> +#include <linux/cdev.h> +#include <linux/wait.h> +#include <linux/mm.h> +#include <linux/pagemap.h> + +#include "config.h" /* compile-time configuration */ +#include "compat.h" /* compatibility definitions for older linux */ +#include "pciDriver.h" /* external interface for the driver */ +#include "common.h" /* internal definitions for all parts */ +#include "kmem.h" /* prototypes for kernel memory */ +#include "sysfs.h" /* prototypes for sysfs */ + +/** + * + * Allocates new kernel memory including the corresponding management structure, makes + * it available via sysfs if possible. + * + */ +int pcidriver_kmem_alloc(pcidriver_privdata_t *privdata, kmem_handle_t *kmem_handle) +{ + pcidriver_kmem_entry_t *kmem_entry; + void *retptr; + + /* First, allocate zeroed memory for the kmem_entry */ + if ((kmem_entry = kcalloc(1, sizeof(pcidriver_kmem_entry_t), GFP_KERNEL)) == NULL) + goto kmem_alloc_entry_fail; + + /* Initialize the kmem_entry */ + kmem_entry->id = atomic_inc_return(&privdata->kmem_count) - 1; + kmem_entry->size = kmem_handle->size; + kmem_handle->handle_id = kmem_entry->id; + + /* Initialize sysfs if possible */ + if (pcidriver_sysfs_initialize_kmem(privdata, kmem_entry->id, &(kmem_entry->sysfs_attr)) != 0) + goto kmem_alloc_mem_fail; + + /* ...and allocate the DMA memory */ + /* note this is a memory pair, referencing the same area: the cpu address (cpua) + * and the PCI bus address (pa). The CPU and PCI addresses may not be the same. + * The CPU sees only CPU addresses, while the device sees only PCI addresses. + * CPU address is used for the mmap (internal to the driver), and + * PCI address is the address passed to the DMA Controller in the device. + */ + retptr = pci_alloc_consistent( privdata->pdev, kmem_handle->size, &(kmem_entry->dma_handle) ); + if (retptr == NULL) + goto kmem_alloc_mem_fail; + kmem_entry->cpua = (unsigned long)retptr; + kmem_handle->pa = (unsigned long)(kmem_entry->dma_handle); + + set_pages_reserved_compat(kmem_entry->cpua, kmem_entry->size); + + /* Add the kmem_entry to the list of the device */ + spin_lock( &(privdata->kmemlist_lock) ); + list_add_tail( &(kmem_entry->list), &(privdata->kmem_list) ); + spin_unlock( &(privdata->kmemlist_lock) ); + + return 0; + +kmem_alloc_mem_fail: + kfree(kmem_entry); +kmem_alloc_entry_fail: + return -ENOMEM; +} + +/** + * + * Called via sysfs, frees kernel memory and the corresponding management structure + * + */ +int pcidriver_kmem_free( pcidriver_privdata_t *privdata, kmem_handle_t *kmem_handle ) +{ + pcidriver_kmem_entry_t *kmem_entry; + + /* Find the associated kmem_entry for this buffer */ + if ((kmem_entry = pcidriver_kmem_find_entry(privdata, kmem_handle)) == NULL) + return -EINVAL; /* kmem_handle is not valid */ + + return pcidriver_kmem_free_entry(privdata, kmem_entry); +} + +/** + * + * Called when cleaning up, frees all kernel memory and their corresponding management structure + * + */ +int pcidriver_kmem_free_all(pcidriver_privdata_t *privdata) +{ + struct list_head *ptr, *next; + pcidriver_kmem_entry_t *kmem_entry; + + /* iterate safely over the entries and delete them */ + list_for_each_safe(ptr, next, &(privdata->kmem_list)) { + kmem_entry = list_entry(ptr, pcidriver_kmem_entry_t, list); + pcidriver_kmem_free_entry(privdata, kmem_entry); /* spin lock inside! */ + } + + return 0; +} + +/** + * + * Synchronize memory to/from the device (or in both directions). + * + */ +int pcidriver_kmem_sync( pcidriver_privdata_t *privdata, kmem_sync_t *kmem_sync ) +{ + pcidriver_kmem_entry_t *kmem_entry; + + /* Find the associated kmem_entry for this buffer */ + if ((kmem_entry = pcidriver_kmem_find_entry(privdata, &(kmem_sync->handle))) == NULL) + return -EINVAL; /* kmem_handle is not valid */ + +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,11) + switch (kmem_sync->dir) { + case PCIDRIVER_DMA_TODEVICE: + pci_dma_sync_single_for_device( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_TODEVICE ); + break; + case PCIDRIVER_DMA_FROMDEVICE: + pci_dma_sync_single_for_cpu( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_FROMDEVICE ); + break; + case PCIDRIVER_DMA_BIDIRECTIONAL: + pci_dma_sync_single_for_device( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_BIDIRECTIONAL ); + pci_dma_sync_single_for_cpu( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_BIDIRECTIONAL ); + break; + default: + return -EINVAL; /* wrong direction parameter */ + } +#else + switch (kmem_sync->dir) { + case PCIDRIVER_DMA_TODEVICE: + pci_dma_sync_single( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_TODEVICE ); + break; + case PCIDRIVER_DMA_FROMDEVICE: + pci_dma_sync_single( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_FROMDEVICE ); + break; + case PCIDRIVER_DMA_BIDIRECTIONAL: + pci_dma_sync_single( privdata->pdev, kmem_entry->dma_handle, kmem_entry->size, PCI_DMA_BIDIRECTIONAL ); + break; + default: + return -EINVAL; /* wrong direction parameter */ + } +#endif + + return 0; /* success */ +} + +/** + * + * Free the given kmem_entry and its memory. + * + */ +int pcidriver_kmem_free_entry(pcidriver_privdata_t *privdata, pcidriver_kmem_entry_t *kmem_entry) +{ + pcidriver_sysfs_remove(privdata, &(kmem_entry->sysfs_attr)); + + /* Go over the pages of the kmem buffer, and mark them as not reserved */ +#if 0 +#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,15) + /* + * This code is DISABLED. + * Apparently, it is not needed to unreserve them. Doing so here + * hangs the machine. Why? + * + * Uhm.. see links: + * + * http://lwn.net/Articles/161204/ + * http://lists.openfabrics.org/pipermail/general/2007-March/034101.html + * + * I insist, this should be enabled, but doing so hangs the machine. + * Literature supports the point, and there is even a similar problem (see link) + * But this is not the case. It seems right to me. but obviously is not. + * + * Anyway, this goes away in kernel >=2.6.15. + */ + unsigned long start = __pa(kmem_entry->cpua) >> PAGE_SHIFT; + unsigned long end = __pa(kmem_entry->cpua + kmem_entry->size) >> PAGE_SHIFT; + unsigned long i; + for(i=start;i<end;i++) { + struct page *kpage = pfn_to_page(i); + ClearPageReserved(kpage); + } +#endif +#endif + + /* Release DMA memory */ + pci_free_consistent( privdata->pdev, kmem_entry->size, (void *)(kmem_entry->cpua), kmem_entry->dma_handle ); + + /* Remove the kmem list entry */ + spin_lock( &(privdata->kmemlist_lock) ); + list_del( &(kmem_entry->list) ); + spin_unlock( &(privdata->kmemlist_lock) ); + + /* Release kmem_entry memory */ + kfree(kmem_entry); + + return 0; +} + +/** + * + * Find the corresponding kmem_entry for the given kmem_handle. + * + */ +pcidriver_kmem_entry_t *pcidriver_kmem_find_entry(pcidriver_privdata_t *privdata, kmem_handle_t *kmem_handle) +{ + struct list_head *ptr; + pcidriver_kmem_entry_t *entry, *result = NULL; + + /* should I implement it better using the handle_id? */ + + spin_lock(&(privdata->kmemlist_lock)); + list_for_each(ptr, &(privdata->kmem_list)) { + entry = list_entry(ptr, pcidriver_kmem_entry_t, list); + + if (entry->dma_handle == kmem_handle->pa) { + result = entry; + break; + } + } + + spin_unlock(&(privdata->kmemlist_lock)); + return result; +} + +/** + * + * find the corresponding kmem_entry for the given id. + * + */ +pcidriver_kmem_entry_t *pcidriver_kmem_find_entry_id(pcidriver_privdata_t *privdata, int id) +{ + struct list_head *ptr; + pcidriver_kmem_entry_t *entry, *result = NULL; + + spin_lock(&(privdata->kmemlist_lock)); + list_for_each(ptr, &(privdata->kmem_list)) { + entry = list_entry(ptr, pcidriver_kmem_entry_t, list); + + if (entry->id == id) { + result = entry; + break; + } + } + + spin_unlock(&(privdata->kmemlist_lock)); + return result; +} + +/** + * + * mmap() kernel memory to userspace. + * + */ +int pcidriver_mmap_kmem(pcidriver_privdata_t *privdata, struct vm_area_struct *vma) +{ + unsigned long vma_size; + pcidriver_kmem_entry_t *kmem_entry; + int ret; + + mod_info_dbg("Entering mmap_kmem\n"); + + /* FIXME: Is this really right? Always just the latest one? Can't we identify one? */ + /* Get latest entry on the kmem_list */ + spin_lock(&(privdata->kmemlist_lock)); + if (list_empty(&(privdata->kmem_list))) { + spin_unlock(&(privdata->kmemlist_lock)); + mod_info("Trying to mmap a kernel memory buffer without creating it first!\n"); + return -EFAULT; + } + kmem_entry = list_entry(privdata->kmem_list.prev, pcidriver_kmem_entry_t, list); + spin_unlock(&(privdata->kmemlist_lock)); + + mod_info_dbg("Got kmem_entry with id: %d\n", kmem_entry->id); + + /* Check sizes */ + vma_size = (vma->vm_end - vma->vm_start); + if ((vma_size != kmem_entry->size) && + ((kmem_entry->size < PAGE_SIZE) && (vma_size != PAGE_SIZE))) { + mod_info("kem_entry size(%lu) and vma size do not match(%lu)\n", kmem_entry->size, vma_size); + return -EINVAL; + } + + vma->vm_flags |= (VM_RESERVED); + +#ifdef pgprot_noncached + // This is coherent memory, so it must not be cached. + vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); +#endif + + mod_info_dbg("Mapping address %08lx / PFN %08lx\n", + virt_to_phys((void*)kmem_entry->cpua), + page_to_pfn(virt_to_page((void*)kmem_entry->cpua))); + + ret = remap_pfn_range_cpua_compat( + vma, + vma->vm_start, + kmem_entry->cpua, + kmem_entry->size, + vma->vm_page_prot ); + + if (ret) { + mod_info("kmem remap failed: %d (%lx)\n", ret,kmem_entry->cpua); + return -EAGAIN; + } + + return ret; +} |