c9636244f8
Exercise the new __sg_alloc_table_from_pages API (and through it also the old sg_alloc_table_from_pages), checking that the created table has the expected number of segments depending on the sequence of input pages and other conditions. v2: Move to data driven for readability. v3: Add some more testcases and -fsanitize=undefined. (Chris Wilson) Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> Cc: Chris Wilson <chris@chris-wilson.co.uk> Cc: linux-kernel@vger.kernel.org Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk> Link: https://patchwork.freedesktop.org/patch/msgid/20170906145506.14952-1-tvrtko.ursulin@linux.intel.com [tursulin: whitespace fixup]
125 lines
2.9 KiB
C
125 lines
2.9 KiB
C
#ifndef _LINUX_MM_H
|
|
#define _LINUX_MM_H
|
|
|
|
#include <assert.h>
|
|
#include <string.h>
|
|
#include <stdlib.h>
|
|
#include <errno.h>
|
|
#include <limits.h>
|
|
#include <stdio.h>
|
|
|
|
typedef unsigned long dma_addr_t;
|
|
|
|
#define unlikely
|
|
|
|
#define BUG_ON(x) assert(!(x))
|
|
|
|
#define WARN_ON(condition) ({ \
|
|
int __ret_warn_on = !!(condition); \
|
|
unlikely(__ret_warn_on); \
|
|
})
|
|
|
|
#define WARN_ON_ONCE(condition) ({ \
|
|
int __ret_warn_on = !!(condition); \
|
|
if (unlikely(__ret_warn_on)) \
|
|
assert(0); \
|
|
unlikely(__ret_warn_on); \
|
|
})
|
|
|
|
#define PAGE_SIZE (4096)
|
|
#define PAGE_SHIFT (12)
|
|
#define PAGE_MASK (~(PAGE_SIZE-1))
|
|
|
|
#define __ALIGN_KERNEL(x, a) __ALIGN_KERNEL_MASK(x, (typeof(x))(a) - 1)
|
|
#define __ALIGN_KERNEL_MASK(x, mask) (((x) + (mask)) & ~(mask))
|
|
#define ALIGN(x, a) __ALIGN_KERNEL((x), (a))
|
|
|
|
#define PAGE_ALIGN(addr) ALIGN(addr, PAGE_SIZE)
|
|
|
|
#define offset_in_page(p) ((unsigned long)(p) & ~PAGE_MASK)
|
|
|
|
#define virt_to_page(x) ((void *)x)
|
|
#define page_address(x) ((void *)x)
|
|
|
|
static inline unsigned long page_to_phys(struct page *page)
|
|
{
|
|
assert(0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#define page_to_pfn(page) ((unsigned long)(page) / PAGE_SIZE)
|
|
#define pfn_to_page(pfn) (void *)((pfn) * PAGE_SIZE)
|
|
#define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n))
|
|
|
|
#define __min(t1, t2, min1, min2, x, y) ({ \
|
|
t1 min1 = (x); \
|
|
t2 min2 = (y); \
|
|
(void) (&min1 == &min2); \
|
|
min1 < min2 ? min1 : min2; })
|
|
|
|
#define ___PASTE(a,b) a##b
|
|
#define __PASTE(a,b) ___PASTE(a,b)
|
|
|
|
#define __UNIQUE_ID(prefix) __PASTE(__PASTE(__UNIQUE_ID_, prefix), __COUNTER__)
|
|
|
|
#define min(x, y) \
|
|
__min(typeof(x), typeof(y), \
|
|
__UNIQUE_ID(min1_), __UNIQUE_ID(min2_), \
|
|
x, y)
|
|
|
|
#define min_t(type, x, y) \
|
|
__min(type, type, \
|
|
__UNIQUE_ID(min1_), __UNIQUE_ID(min2_), \
|
|
x, y)
|
|
|
|
#define preemptible() (1)
|
|
|
|
static inline void *kmap(struct page *page)
|
|
{
|
|
assert(0);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static inline void *kmap_atomic(struct page *page)
|
|
{
|
|
assert(0);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static inline void kunmap(void *addr)
|
|
{
|
|
assert(0);
|
|
}
|
|
|
|
static inline void kunmap_atomic(void *addr)
|
|
{
|
|
assert(0);
|
|
}
|
|
|
|
static inline unsigned long __get_free_page(unsigned int flags)
|
|
{
|
|
return (unsigned long)malloc(PAGE_SIZE);
|
|
}
|
|
|
|
static inline void free_page(unsigned long page)
|
|
{
|
|
free((void *)page);
|
|
}
|
|
|
|
static inline void *kmalloc(unsigned int size, unsigned int flags)
|
|
{
|
|
return malloc(size);
|
|
}
|
|
|
|
#define kfree(x) free(x)
|
|
|
|
#define kmemleak_alloc(a, b, c, d)
|
|
#define kmemleak_free(a)
|
|
|
|
#define PageSlab(p) (0)
|
|
#define flush_kernel_dcache_page(p)
|
|
|
|
#endif
|