347 lines
8 KiB
C
347 lines
8 KiB
C
|
/* i915_mem.c -- Simple agp/fb memory manager for i915 -*- linux-c -*-
|
||
|
*/
|
||
|
/**************************************************************************
|
||
|
*
|
||
|
* Copyright 2003 Tungsten Graphics, Inc., Cedar Park, Texas.
|
||
|
* All Rights Reserved.
|
||
|
*
|
||
|
**************************************************************************/
|
||
|
|
||
|
#include "drmP.h"
|
||
|
#include "drm.h"
|
||
|
#include "i915_drm.h"
|
||
|
#include "i915_drv.h"
|
||
|
|
||
|
/* This memory manager is integrated into the global/local lru
|
||
|
* mechanisms used by the clients. Specifically, it operates by
|
||
|
* setting the 'in_use' fields of the global LRU to indicate whether
|
||
|
* this region is privately allocated to a client.
|
||
|
*
|
||
|
* This does require the client to actually respect that field.
|
||
|
*
|
||
|
* Currently no effort is made to allocate 'private' memory in any
|
||
|
* clever way - the LRU information isn't used to determine which
|
||
|
* block to allocate, and the ring is drained prior to allocations --
|
||
|
* in other words allocation is expensive.
|
||
|
*/
|
||
|
static void mark_block(drm_device_t * dev, struct mem_block *p, int in_use)
|
||
|
{
|
||
|
drm_i915_private_t *dev_priv = dev->dev_private;
|
||
|
drm_i915_sarea_t *sarea_priv = dev_priv->sarea_priv;
|
||
|
drm_tex_region_t *list;
|
||
|
unsigned shift, nr;
|
||
|
unsigned start;
|
||
|
unsigned end;
|
||
|
unsigned i;
|
||
|
int age;
|
||
|
|
||
|
shift = dev_priv->tex_lru_log_granularity;
|
||
|
nr = I915_NR_TEX_REGIONS;
|
||
|
|
||
|
start = p->start >> shift;
|
||
|
end = (p->start + p->size - 1) >> shift;
|
||
|
|
||
|
age = ++sarea_priv->texAge;
|
||
|
list = sarea_priv->texList;
|
||
|
|
||
|
/* Mark the regions with the new flag and update their age. Move
|
||
|
* them to head of list to preserve LRU semantics.
|
||
|
*/
|
||
|
for (i = start; i <= end; i++) {
|
||
|
list[i].in_use = in_use;
|
||
|
list[i].age = age;
|
||
|
|
||
|
/* remove_from_list(i)
|
||
|
*/
|
||
|
list[(unsigned)list[i].next].prev = list[i].prev;
|
||
|
list[(unsigned)list[i].prev].next = list[i].next;
|
||
|
|
||
|
/* insert_at_head(list, i)
|
||
|
*/
|
||
|
list[i].prev = nr;
|
||
|
list[i].next = list[nr].next;
|
||
|
list[(unsigned)list[nr].next].prev = i;
|
||
|
list[nr].next = i;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
/* Very simple allocator for agp memory, working on a static range
|
||
|
* already mapped into each client's address space.
|
||
|
*/
|
||
|
|
||
|
static struct mem_block *split_block(struct mem_block *p, int start, int size,
|
||
|
DRMFILE filp)
|
||
|
{
|
||
|
/* Maybe cut off the start of an existing block */
|
||
|
if (start > p->start) {
|
||
|
struct mem_block *newblock = drm_alloc(sizeof(*newblock), DRM_MEM_BUFLISTS);
|
||
|
if (!newblock)
|
||
|
goto out;
|
||
|
newblock->start = start;
|
||
|
newblock->size = p->size - (start - p->start);
|
||
|
newblock->filp = NULL;
|
||
|
newblock->next = p->next;
|
||
|
newblock->prev = p;
|
||
|
p->next->prev = newblock;
|
||
|
p->next = newblock;
|
||
|
p->size -= newblock->size;
|
||
|
p = newblock;
|
||
|
}
|
||
|
|
||
|
/* Maybe cut off the end of an existing block */
|
||
|
if (size < p->size) {
|
||
|
struct mem_block *newblock = drm_alloc(sizeof(*newblock), DRM_MEM_BUFLISTS);
|
||
|
if (!newblock)
|
||
|
goto out;
|
||
|
newblock->start = start + size;
|
||
|
newblock->size = p->size - size;
|
||
|
newblock->filp = NULL;
|
||
|
newblock->next = p->next;
|
||
|
newblock->prev = p;
|
||
|
p->next->prev = newblock;
|
||
|
p->next = newblock;
|
||
|
p->size = size;
|
||
|
}
|
||
|
|
||
|
out:
|
||
|
/* Our block is in the middle */
|
||
|
p->filp = filp;
|
||
|
return p;
|
||
|
}
|
||
|
|
||
|
static struct mem_block *alloc_block(struct mem_block *heap, int size,
|
||
|
int align2, DRMFILE filp)
|
||
|
{
|
||
|
struct mem_block *p;
|
||
|
int mask = (1 << align2) - 1;
|
||
|
|
||
|
for (p = heap->next; p != heap; p = p->next) {
|
||
|
int start = (p->start + mask) & ~mask;
|
||
|
if (p->filp == NULL && start + size <= p->start + p->size)
|
||
|
return split_block(p, start, size, filp);
|
||
|
}
|
||
|
|
||
|
return NULL;
|
||
|
}
|
||
|
|
||
|
static struct mem_block *find_block(struct mem_block *heap, int start)
|
||
|
{
|
||
|
struct mem_block *p;
|
||
|
|
||
|
for (p = heap->next; p != heap; p = p->next)
|
||
|
if (p->start == start)
|
||
|
return p;
|
||
|
|
||
|
return NULL;
|
||
|
}
|
||
|
|
||
|
static void free_block(struct mem_block *p)
|
||
|
{
|
||
|
p->filp = NULL;
|
||
|
|
||
|
/* Assumes a single contiguous range. Needs a special filp in
|
||
|
* 'heap' to stop it being subsumed.
|
||
|
*/
|
||
|
if (p->next->filp == NULL) {
|
||
|
struct mem_block *q = p->next;
|
||
|
p->size += q->size;
|
||
|
p->next = q->next;
|
||
|
p->next->prev = p;
|
||
|
drm_free(q, sizeof(*q), DRM_MEM_BUFLISTS);
|
||
|
}
|
||
|
|
||
|
if (p->prev->filp == NULL) {
|
||
|
struct mem_block *q = p->prev;
|
||
|
q->size += p->size;
|
||
|
q->next = p->next;
|
||
|
q->next->prev = q;
|
||
|
drm_free(p, sizeof(*q), DRM_MEM_BUFLISTS);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
/* Initialize. How to check for an uninitialized heap?
|
||
|
*/
|
||
|
static int init_heap(struct mem_block **heap, int start, int size)
|
||
|
{
|
||
|
struct mem_block *blocks = drm_alloc(sizeof(*blocks), DRM_MEM_BUFLISTS);
|
||
|
|
||
|
if (!blocks)
|
||
|
return -ENOMEM;
|
||
|
|
||
|
*heap = drm_alloc(sizeof(**heap), DRM_MEM_BUFLISTS);
|
||
|
if (!*heap) {
|
||
|
drm_free(blocks, sizeof(*blocks), DRM_MEM_BUFLISTS);
|
||
|
return -ENOMEM;
|
||
|
}
|
||
|
|
||
|
blocks->start = start;
|
||
|
blocks->size = size;
|
||
|
blocks->filp = NULL;
|
||
|
blocks->next = blocks->prev = *heap;
|
||
|
|
||
|
memset(*heap, 0, sizeof(**heap));
|
||
|
(*heap)->filp = (DRMFILE) - 1;
|
||
|
(*heap)->next = (*heap)->prev = blocks;
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
/* Free all blocks associated with the releasing file.
|
||
|
*/
|
||
|
void i915_mem_release(drm_device_t * dev, DRMFILE filp, struct mem_block *heap)
|
||
|
{
|
||
|
struct mem_block *p;
|
||
|
|
||
|
if (!heap || !heap->next)
|
||
|
return;
|
||
|
|
||
|
for (p = heap->next; p != heap; p = p->next) {
|
||
|
if (p->filp == filp) {
|
||
|
p->filp = NULL;
|
||
|
mark_block(dev, p, 0);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
/* Assumes a single contiguous range. Needs a special filp in
|
||
|
* 'heap' to stop it being subsumed.
|
||
|
*/
|
||
|
for (p = heap->next; p != heap; p = p->next) {
|
||
|
while (p->filp == NULL && p->next->filp == NULL) {
|
||
|
struct mem_block *q = p->next;
|
||
|
p->size += q->size;
|
||
|
p->next = q->next;
|
||
|
p->next->prev = p;
|
||
|
drm_free(q, sizeof(*q), DRM_MEM_BUFLISTS);
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
/* Shutdown.
|
||
|
*/
|
||
|
void i915_mem_takedown(struct mem_block **heap)
|
||
|
{
|
||
|
struct mem_block *p;
|
||
|
|
||
|
if (!*heap)
|
||
|
return;
|
||
|
|
||
|
for (p = (*heap)->next; p != *heap;) {
|
||
|
struct mem_block *q = p;
|
||
|
p = p->next;
|
||
|
drm_free(q, sizeof(*q), DRM_MEM_BUFLISTS);
|
||
|
}
|
||
|
|
||
|
drm_free(*heap, sizeof(**heap), DRM_MEM_BUFLISTS);
|
||
|
*heap = NULL;
|
||
|
}
|
||
|
|
||
|
static struct mem_block **get_heap(drm_i915_private_t * dev_priv, int region)
|
||
|
{
|
||
|
switch (region) {
|
||
|
case I915_MEM_REGION_AGP:
|
||
|
return &dev_priv->agp_heap;
|
||
|
default:
|
||
|
return NULL;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
/* IOCTL HANDLERS */
|
||
|
|
||
|
int i915_mem_alloc(DRM_IOCTL_ARGS)
|
||
|
{
|
||
|
DRM_DEVICE;
|
||
|
drm_i915_private_t *dev_priv = dev->dev_private;
|
||
|
drm_i915_mem_alloc_t alloc;
|
||
|
struct mem_block *block, **heap;
|
||
|
|
||
|
if (!dev_priv) {
|
||
|
DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
|
||
|
return DRM_ERR(EINVAL);
|
||
|
}
|
||
|
|
||
|
DRM_COPY_FROM_USER_IOCTL(alloc, (drm_i915_mem_alloc_t __user *) data,
|
||
|
sizeof(alloc));
|
||
|
|
||
|
heap = get_heap(dev_priv, alloc.region);
|
||
|
if (!heap || !*heap)
|
||
|
return DRM_ERR(EFAULT);
|
||
|
|
||
|
/* Make things easier on ourselves: all allocations at least
|
||
|
* 4k aligned.
|
||
|
*/
|
||
|
if (alloc.alignment < 12)
|
||
|
alloc.alignment = 12;
|
||
|
|
||
|
block = alloc_block(*heap, alloc.size, alloc.alignment, filp);
|
||
|
|
||
|
if (!block)
|
||
|
return DRM_ERR(ENOMEM);
|
||
|
|
||
|
mark_block(dev, block, 1);
|
||
|
|
||
|
if (DRM_COPY_TO_USER(alloc.region_offset, &block->start, sizeof(int))) {
|
||
|
DRM_ERROR("copy_to_user\n");
|
||
|
return DRM_ERR(EFAULT);
|
||
|
}
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
int i915_mem_free(DRM_IOCTL_ARGS)
|
||
|
{
|
||
|
DRM_DEVICE;
|
||
|
drm_i915_private_t *dev_priv = dev->dev_private;
|
||
|
drm_i915_mem_free_t memfree;
|
||
|
struct mem_block *block, **heap;
|
||
|
|
||
|
if (!dev_priv) {
|
||
|
DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
|
||
|
return DRM_ERR(EINVAL);
|
||
|
}
|
||
|
|
||
|
DRM_COPY_FROM_USER_IOCTL(memfree, (drm_i915_mem_free_t __user *) data,
|
||
|
sizeof(memfree));
|
||
|
|
||
|
heap = get_heap(dev_priv, memfree.region);
|
||
|
if (!heap || !*heap)
|
||
|
return DRM_ERR(EFAULT);
|
||
|
|
||
|
block = find_block(*heap, memfree.region_offset);
|
||
|
if (!block)
|
||
|
return DRM_ERR(EFAULT);
|
||
|
|
||
|
if (block->filp != filp)
|
||
|
return DRM_ERR(EPERM);
|
||
|
|
||
|
mark_block(dev, block, 0);
|
||
|
free_block(block);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
int i915_mem_init_heap(DRM_IOCTL_ARGS)
|
||
|
{
|
||
|
DRM_DEVICE;
|
||
|
drm_i915_private_t *dev_priv = dev->dev_private;
|
||
|
drm_i915_mem_init_heap_t initheap;
|
||
|
struct mem_block **heap;
|
||
|
|
||
|
if (!dev_priv) {
|
||
|
DRM_ERROR("%s called with no initialization\n", __FUNCTION__);
|
||
|
return DRM_ERR(EINVAL);
|
||
|
}
|
||
|
|
||
|
DRM_COPY_FROM_USER_IOCTL(initheap,
|
||
|
(drm_i915_mem_init_heap_t __user *) data,
|
||
|
sizeof(initheap));
|
||
|
|
||
|
heap = get_heap(dev_priv, initheap.region);
|
||
|
if (!heap)
|
||
|
return DRM_ERR(EFAULT);
|
||
|
|
||
|
if (*heap) {
|
||
|
DRM_ERROR("heap already initialized?");
|
||
|
return DRM_ERR(EFAULT);
|
||
|
}
|
||
|
|
||
|
return init_heap(heap, initheap.start, initheap.size);
|
||
|
}
|