/drivers/include/linux/bitmap.h |
---|
144,7 → 144,9 |
extern void bitmap_release_region(unsigned long *bitmap, int pos, int order); |
extern int bitmap_allocate_region(unsigned long *bitmap, int pos, int order); |
extern void bitmap_copy_le(void *dst, const unsigned long *src, int nbits); |
extern int bitmap_ord_to_pos(const unsigned long *bitmap, int n, int bits); |
#define BITMAP_FIRST_WORD_MASK(start) (~0UL << ((start) % BITS_PER_LONG)) |
#define BITMAP_LAST_WORD_MASK(nbits) \ |
( \ |
((nbits) % BITS_PER_LONG) ? \ |
/drivers/include/linux/ctype.h |
---|
52,4 → 52,13 |
#define tolower(c) __tolower(c) |
#define toupper(c) __toupper(c) |
/* |
* Fast implementation of tolower() for internal usage. Do not use in your |
* code. |
*/ |
static inline char _tolower(const char c) |
{ |
return c | 0x20; |
} |
#endif |
/drivers/include/linux/errno.h |
---|
111,4 → 111,6 |
#define ERFKILL 132 /* Operation not possible due to RF-kill */ |
#define ENOTSUPP 524 /* Operation is not supported */ |
#endif |
/drivers/include/linux/i2c.h |
---|
280,9 → 280,6 |
/* Internal numbers to terminate lists */ |
#define I2C_CLIENT_END 0xfffeU |
/* The numbers to use to set I2C bus address */ |
#define ANY_I2C_BUS 0xffff |
/* Construct an I2C_CLIENT_END-terminated array of i2c addresses */ |
#define I2C_ADDRS(addr, addrs...) \ |
((const unsigned short []){ addr, ## addrs, I2C_CLIENT_END }) |
289,6 → 286,7 |
#endif /* __KERNEL__ */ |
/** |
* struct i2c_msg - an I2C transaction segment beginning with START |
* @addr: Slave address, either seven or ten bits. When this is a ten |
/drivers/include/linux/jiffies.h |
---|
0,0 → 1,307 |
#ifndef _LINUX_JIFFIES_H |
#define _LINUX_JIFFIES_H |
//#include <linux/math64.h> |
#include <linux/kernel.h> |
#include <linux/types.h> |
//#include <linux/time.h> |
//#include <linux/timex.h> |
//#include <asm/param.h> /* for HZ */ |
#define HZ 100 |
#define CLOCK_TICK_RATE 1193182ul |
/* |
* The following defines establish the engineering parameters of the PLL |
* model. The HZ variable establishes the timer interrupt frequency, 100 Hz |
* for the SunOS kernel, 256 Hz for the Ultrix kernel and 1024 Hz for the |
* OSF/1 kernel. The SHIFT_HZ define expresses the same value as the |
* nearest power of two in order to avoid hardware multiply operations. |
*/ |
#if HZ >= 12 && HZ < 24 |
# define SHIFT_HZ 4 |
#elif HZ >= 24 && HZ < 48 |
# define SHIFT_HZ 5 |
#elif HZ >= 48 && HZ < 96 |
# define SHIFT_HZ 6 |
#elif HZ >= 96 && HZ < 192 |
# define SHIFT_HZ 7 |
#elif HZ >= 192 && HZ < 384 |
# define SHIFT_HZ 8 |
#elif HZ >= 384 && HZ < 768 |
# define SHIFT_HZ 9 |
#elif HZ >= 768 && HZ < 1536 |
# define SHIFT_HZ 10 |
#elif HZ >= 1536 && HZ < 3072 |
# define SHIFT_HZ 11 |
#elif HZ >= 3072 && HZ < 6144 |
# define SHIFT_HZ 12 |
#elif HZ >= 6144 && HZ < 12288 |
# define SHIFT_HZ 13 |
#else |
# error Invalid value of HZ. |
#endif |
/* LATCH is used in the interval timer and ftape setup. */ |
#define LATCH ((CLOCK_TICK_RATE + HZ/2) / HZ) /* For divider */ |
/* Suppose we want to divide two numbers NOM and DEN: NOM/DEN, then we can |
* improve accuracy by shifting LSH bits, hence calculating: |
* (NOM << LSH) / DEN |
* This however means trouble for large NOM, because (NOM << LSH) may no |
* longer fit in 32 bits. The following way of calculating this gives us |
* some slack, under the following conditions: |
* - (NOM / DEN) fits in (32 - LSH) bits. |
* - (NOM % DEN) fits in (32 - LSH) bits. |
*/ |
#define SH_DIV(NOM,DEN,LSH) ( (((NOM) / (DEN)) << (LSH)) \ |
+ ((((NOM) % (DEN)) << (LSH)) + (DEN) / 2) / (DEN)) |
/* HZ is the requested value. ACTHZ is actual HZ ("<< 8" is for accuracy) */ |
#define ACTHZ (SH_DIV (CLOCK_TICK_RATE, LATCH, 8)) |
/* TICK_NSEC is the time between ticks in nsec assuming real ACTHZ */ |
#define TICK_NSEC (SH_DIV (1000000UL * 1000, ACTHZ, 8)) |
/* TICK_USEC is the time between ticks in usec assuming fake USER_HZ */ |
#define TICK_USEC ((1000000UL + USER_HZ/2) / USER_HZ) |
/* TICK_USEC_TO_NSEC is the time between ticks in nsec assuming real ACTHZ and */ |
/* a value TUSEC for TICK_USEC (can be set bij adjtimex) */ |
#define TICK_USEC_TO_NSEC(TUSEC) (SH_DIV (TUSEC * USER_HZ * 1000, ACTHZ, 8)) |
#define jiffies GetTimerTicks() |
#if (BITS_PER_LONG < 64) |
u64 get_jiffies_64(void); |
#else |
static inline u64 get_jiffies_64(void) |
{ |
return (u64)jiffies; |
} |
#endif |
/* |
* These inlines deal with timer wrapping correctly. You are |
* strongly encouraged to use them |
* 1. Because people otherwise forget |
* 2. Because if the timer wrap changes in future you won't have to |
* alter your driver code. |
* |
* time_after(a,b) returns true if the time a is after time b. |
* |
* Do this with "<0" and ">=0" to only test the sign of the result. A |
* good compiler would generate better code (and a really good compiler |
* wouldn't care). Gcc is currently neither. |
*/ |
#define time_after(a,b) \ |
(typecheck(unsigned long, a) && \ |
typecheck(unsigned long, b) && \ |
((long)(b) - (long)(a) < 0)) |
#define time_before(a,b) time_after(b,a) |
#define time_after_eq(a,b) \ |
(typecheck(unsigned long, a) && \ |
typecheck(unsigned long, b) && \ |
((long)(a) - (long)(b) >= 0)) |
#define time_before_eq(a,b) time_after_eq(b,a) |
/* |
* Calculate whether a is in the range of [b, c]. |
*/ |
#define time_in_range(a,b,c) \ |
(time_after_eq(a,b) && \ |
time_before_eq(a,c)) |
/* |
* Calculate whether a is in the range of [b, c). |
*/ |
#define time_in_range_open(a,b,c) \ |
(time_after_eq(a,b) && \ |
time_before(a,c)) |
/* Same as above, but does so with platform independent 64bit types. |
* These must be used when utilizing jiffies_64 (i.e. return value of |
* get_jiffies_64() */ |
#define time_after64(a,b) \ |
(typecheck(__u64, a) && \ |
typecheck(__u64, b) && \ |
((__s64)(b) - (__s64)(a) < 0)) |
#define time_before64(a,b) time_after64(b,a) |
#define time_after_eq64(a,b) \ |
(typecheck(__u64, a) && \ |
typecheck(__u64, b) && \ |
((__s64)(a) - (__s64)(b) >= 0)) |
#define time_before_eq64(a,b) time_after_eq64(b,a) |
/* |
* These four macros compare jiffies and 'a' for convenience. |
*/ |
/* time_is_before_jiffies(a) return true if a is before jiffies */ |
#define time_is_before_jiffies(a) time_after(jiffies, a) |
/* time_is_after_jiffies(a) return true if a is after jiffies */ |
#define time_is_after_jiffies(a) time_before(jiffies, a) |
/* time_is_before_eq_jiffies(a) return true if a is before or equal to jiffies*/ |
#define time_is_before_eq_jiffies(a) time_after_eq(jiffies, a) |
/* time_is_after_eq_jiffies(a) return true if a is after or equal to jiffies*/ |
#define time_is_after_eq_jiffies(a) time_before_eq(jiffies, a) |
/* |
* Have the 32 bit jiffies value wrap 5 minutes after boot |
* so jiffies wrap bugs show up earlier. |
*/ |
#define INITIAL_JIFFIES ((unsigned long)(unsigned int) (-300*HZ)) |
/* |
* Change timeval to jiffies, trying to avoid the |
* most obvious overflows.. |
* |
* And some not so obvious. |
* |
* Note that we don't want to return LONG_MAX, because |
* for various timeout reasons we often end up having |
* to wait "jiffies+1" in order to guarantee that we wait |
* at _least_ "jiffies" - so "jiffies+1" had better still |
* be positive. |
*/ |
#define MAX_JIFFY_OFFSET ((LONG_MAX >> 1)-1) |
extern unsigned long preset_lpj; |
/* |
* We want to do realistic conversions of time so we need to use the same |
* values the update wall clock code uses as the jiffies size. This value |
* is: TICK_NSEC (which is defined in timex.h). This |
* is a constant and is in nanoseconds. We will use scaled math |
* with a set of scales defined here as SEC_JIFFIE_SC, USEC_JIFFIE_SC and |
* NSEC_JIFFIE_SC. Note that these defines contain nothing but |
* constants and so are computed at compile time. SHIFT_HZ (computed in |
* timex.h) adjusts the scaling for different HZ values. |
* Scaled math??? What is that? |
* |
* Scaled math is a way to do integer math on values that would, |
* otherwise, either overflow, underflow, or cause undesired div |
* instructions to appear in the execution path. In short, we "scale" |
* up the operands so they take more bits (more precision, less |
* underflow), do the desired operation and then "scale" the result back |
* by the same amount. If we do the scaling by shifting we avoid the |
* costly mpy and the dastardly div instructions. |
* Suppose, for example, we want to convert from seconds to jiffies |
* where jiffies is defined in nanoseconds as NSEC_PER_JIFFIE. The |
* simple math is: jiff = (sec * NSEC_PER_SEC) / NSEC_PER_JIFFIE; We |
* observe that (NSEC_PER_SEC / NSEC_PER_JIFFIE) is a constant which we |
* might calculate at compile time, however, the result will only have |
* about 3-4 bits of precision (less for smaller values of HZ). |
* |
* So, we scale as follows: |
* jiff = (sec) * (NSEC_PER_SEC / NSEC_PER_JIFFIE); |
* jiff = ((sec) * ((NSEC_PER_SEC * SCALE)/ NSEC_PER_JIFFIE)) / SCALE; |
* Then we make SCALE a power of two so: |
* jiff = ((sec) * ((NSEC_PER_SEC << SCALE)/ NSEC_PER_JIFFIE)) >> SCALE; |
* Now we define: |
* #define SEC_CONV = ((NSEC_PER_SEC << SCALE)/ NSEC_PER_JIFFIE)) |
* jiff = (sec * SEC_CONV) >> SCALE; |
* |
* Often the math we use will expand beyond 32-bits so we tell C how to |
* do this and pass the 64-bit result of the mpy through the ">> SCALE" |
* which should take the result back to 32-bits. We want this expansion |
* to capture as much precision as possible. At the same time we don't |
* want to overflow so we pick the SCALE to avoid this. In this file, |
* that means using a different scale for each range of HZ values (as |
* defined in timex.h). |
* |
* For those who want to know, gcc will give a 64-bit result from a "*" |
* operator if the result is a long long AND at least one of the |
* operands is cast to long long (usually just prior to the "*" so as |
* not to confuse it into thinking it really has a 64-bit operand, |
* which, buy the way, it can do, but it takes more code and at least 2 |
* mpys). |
* We also need to be aware that one second in nanoseconds is only a |
* couple of bits away from overflowing a 32-bit word, so we MUST use |
* 64-bits to get the full range time in nanoseconds. |
*/ |
/* |
* Here are the scales we will use. One for seconds, nanoseconds and |
* microseconds. |
* |
* Within the limits of cpp we do a rough cut at the SEC_JIFFIE_SC and |
* check if the sign bit is set. If not, we bump the shift count by 1. |
* (Gets an extra bit of precision where we can use it.) |
* We know it is set for HZ = 1024 and HZ = 100 not for 1000. |
* Haven't tested others. |
* Limits of cpp (for #if expressions) only long (no long long), but |
* then we only need the most signicant bit. |
*/ |
#define SEC_JIFFIE_SC (31 - SHIFT_HZ) |
#if !((((NSEC_PER_SEC << 2) / TICK_NSEC) << (SEC_JIFFIE_SC - 2)) & 0x80000000) |
#undef SEC_JIFFIE_SC |
#define SEC_JIFFIE_SC (32 - SHIFT_HZ) |
#endif |
#define NSEC_JIFFIE_SC (SEC_JIFFIE_SC + 29) |
#define USEC_JIFFIE_SC (SEC_JIFFIE_SC + 19) |
#define SEC_CONVERSION ((unsigned long)((((u64)NSEC_PER_SEC << SEC_JIFFIE_SC) +\ |
TICK_NSEC -1) / (u64)TICK_NSEC)) |
#define NSEC_CONVERSION ((unsigned long)((((u64)1 << NSEC_JIFFIE_SC) +\ |
TICK_NSEC -1) / (u64)TICK_NSEC)) |
#define USEC_CONVERSION \ |
((unsigned long)((((u64)NSEC_PER_USEC << USEC_JIFFIE_SC) +\ |
TICK_NSEC -1) / (u64)TICK_NSEC)) |
/* |
* USEC_ROUND is used in the timeval to jiffie conversion. See there |
* for more details. It is the scaled resolution rounding value. Note |
* that it is a 64-bit value. Since, when it is applied, we are already |
* in jiffies (albit scaled), it is nothing but the bits we will shift |
* off. |
*/ |
#define USEC_ROUND (u64)(((u64)1 << USEC_JIFFIE_SC) - 1) |
/* |
* The maximum jiffie value is (MAX_INT >> 1). Here we translate that |
* into seconds. The 64-bit case will overflow if we are not careful, |
* so use the messy SH_DIV macro to do it. Still all constants. |
*/ |
#if BITS_PER_LONG < 64 |
# define MAX_SEC_IN_JIFFIES \ |
(long)((u64)((u64)MAX_JIFFY_OFFSET * TICK_NSEC) / NSEC_PER_SEC) |
#else /* take care of overflow on 64 bits machines */ |
# define MAX_SEC_IN_JIFFIES \ |
(SH_DIV((MAX_JIFFY_OFFSET >> SEC_JIFFIE_SC) * TICK_NSEC, NSEC_PER_SEC, 1) - 1) |
#endif |
/* |
* Convert various time units to each other: |
*/ |
extern unsigned int jiffies_to_msecs(const unsigned long j); |
extern unsigned int jiffies_to_usecs(const unsigned long j); |
extern unsigned long msecs_to_jiffies(const unsigned int m); |
extern unsigned long usecs_to_jiffies(const unsigned int u); |
extern unsigned long timespec_to_jiffies(const struct timespec *value); |
extern void jiffies_to_timespec(const unsigned long jiffies, |
struct timespec *value); |
extern unsigned long timeval_to_jiffies(const struct timeval *value); |
extern void jiffies_to_timeval(const unsigned long jiffies, |
struct timeval *value); |
extern clock_t jiffies_to_clock_t(unsigned long x); |
extern unsigned long clock_t_to_jiffies(unsigned long x); |
extern u64 jiffies_64_to_clock_t(u64 x); |
extern u64 nsec_to_clock_t(u64 x); |
extern u64 nsecs_to_jiffies64(u64 n); |
extern unsigned long nsecs_to_jiffies(u64 n); |
#define TIMESTAMP_SIZE 30 |
#endif |
/drivers/include/linux/kernel.h |
---|
13,6 → 13,10 |
#include <linux/compiler.h> |
#include <linux/bitops.h> |
#include <linux/typecheck.h> |
#define __init |
#define USHRT_MAX ((u16)(~0U)) |
#define SHRT_MAX ((s16)(USHRT_MAX>>1)) |
#define SHRT_MIN ((s16)(-SHRT_MAX - 1)) |
31,6 → 35,16 |
#define PTR_ALIGN(p, a) ((typeof(p))ALIGN((unsigned long)(p), (a))) |
#define IS_ALIGNED(x, a) (((x) & ((typeof(x))(a) - 1)) == 0) |
#define roundup(x, y) ((((x) + ((y) - 1)) / (y)) * (y)) |
#define DIV_ROUND_UP(n,d) (((n) + (d) - 1) / (d)) |
#define DIV_ROUND_CLOSEST(x, divisor)( \ |
{ \ |
typeof(divisor) __divisor = divisor; \ |
(((x) + ((__divisor) / 2)) / (__divisor)); \ |
} \ |
) |
/** |
* upper_32_bits - return bits 32-63 of a number |
* @n: the number we're accessing |
220,6 → 234,119 |
typecheck(unsigned long, b) && \ |
((long)(b) - (long)(a) < 0)) |
struct tvec_base; |
struct timer_list { |
struct list_head entry; |
unsigned long expires; |
void (*function)(unsigned long); |
unsigned long data; |
// struct tvec_base *base; |
}; |
struct timespec { |
long tv_sec; /* seconds */ |
long tv_nsec; /* nanoseconds */ |
}; |
#define build_mmio_read(name, size, type, reg, barrier) \ |
static inline type name(const volatile void __iomem *addr) \ |
{ type ret; asm volatile("mov" size " %1,%0":reg (ret) \ |
:"m" (*(volatile type __force *)addr) barrier); return ret; } |
#define build_mmio_write(name, size, type, reg, barrier) \ |
static inline void name(type val, volatile void __iomem *addr) \ |
{ asm volatile("mov" size " %0,%1": :reg (val), \ |
"m" (*(volatile type __force *)addr) barrier); } |
build_mmio_read(readb, "b", unsigned char, "=q", :"memory") |
build_mmio_read(readw, "w", unsigned short, "=r", :"memory") |
build_mmio_read(readl, "l", unsigned int, "=r", :"memory") |
build_mmio_read(__readb, "b", unsigned char, "=q", ) |
build_mmio_read(__readw, "w", unsigned short, "=r", ) |
build_mmio_read(__readl, "l", unsigned int, "=r", ) |
build_mmio_write(writeb, "b", unsigned char, "q", :"memory") |
build_mmio_write(writew, "w", unsigned short, "r", :"memory") |
build_mmio_write(writel, "l", unsigned int, "r", :"memory") |
build_mmio_write(__writeb, "b", unsigned char, "q", ) |
build_mmio_write(__writew, "w", unsigned short, "r", ) |
build_mmio_write(__writel, "l", unsigned int, "r", ) |
#define readb_relaxed(a) __readb(a) |
#define readw_relaxed(a) __readw(a) |
#define readl_relaxed(a) __readl(a) |
#define __raw_readb __readb |
#define __raw_readw __readw |
#define __raw_readl __readl |
#define __raw_writeb __writeb |
#define __raw_writew __writew |
#define __raw_writel __writel |
static inline __u64 readq(const volatile void __iomem *addr) |
{ |
const volatile u32 __iomem *p = addr; |
u32 low, high; |
low = readl(p); |
high = readl(p + 1); |
return low + ((u64)high << 32); |
} |
static inline void writeq(__u64 val, volatile void __iomem *addr) |
{ |
writel(val, addr); |
writel(val >> 32, addr+4); |
} |
#define mmiowb() barrier() |
#define dev_err(dev, format, arg...) \ |
printk("Error %s " format, __func__ , ## arg) |
#define dev_warn(dev, format, arg...) \ |
printk("Warning %s " format, __func__ , ## arg) |
#define dev_info(dev, format, arg...) \ |
printk("Info %s " format , __func__, ## arg) |
#define BUILD_BUG_ON(condition) ((void)sizeof(char[1 - 2*!!(condition)])) |
struct scatterlist { |
unsigned long page_link; |
unsigned int offset; |
unsigned int length; |
dma_addr_t dma_address; |
unsigned int dma_length; |
}; |
struct page |
{ |
unsigned int addr; |
}; |
struct vm_fault { |
unsigned int flags; /* FAULT_FLAG_xxx flags */ |
pgoff_t pgoff; /* Logical page offset based on vma */ |
void __user *virtual_address; /* Faulting virtual address */ |
struct page *page; /* ->fault handlers should return a |
* page here, unless VM_FAULT_NOPAGE |
* is set (which is also implied by |
* VM_FAULT_ERROR). |
*/ |
}; |
#endif |
/drivers/include/linux/lockdep.h |
---|
542,7 → 542,7 |
#endif |
#ifdef CONFIG_PROVE_RCU |
extern void lockdep_rcu_dereference(const char *file, const int line); |
void lockdep_rcu_suspicious(const char *file, const int line, const char *s); |
#endif |
#endif /* __LINUX_LOCKDEP_H */ |
/drivers/include/linux/pci.h |
---|
625,14 → 625,18 |
int enum_pci_devices(void); |
struct pci_device_id* |
find_pci_device(pci_dev_t* pdev, struct pci_device_id *idlist); |
const struct pci_device_id* |
find_pci_device(pci_dev_t* pdev, const struct pci_device_id *idlist); |
#define DMA_BIT_MASK(n) (((n) == 64) ? ~0ULL : ((1ULL<<(n))-1)) |
int pci_set_dma_mask(struct pci_dev *dev, u64 mask); |
struct pci_dev *pci_get_bus_and_slot(unsigned int bus, unsigned int devfn); |
struct pci_dev *pci_get_class(unsigned int class, struct pci_dev *from); |
void __iomem *pci_map_rom(struct pci_dev *pdev, size_t *size); |
#define pci_name(x) "radeon" |
#endif //__PCI__H__ |
/drivers/include/linux/pci_regs.h |
---|
663,6 → 663,26 |
#define PCI_ATS_CTRL_STU(x) ((x) & 0x1f) /* Smallest Translation Unit */ |
#define PCI_ATS_MIN_STU 12 /* shift of minimum STU block */ |
/* Page Request Interface */ |
#define PCI_PRI_CAP 0x13 /* PRI capability ID */ |
#define PCI_PRI_CONTROL_OFF 0x04 /* Offset of control register */ |
#define PCI_PRI_STATUS_OFF 0x06 /* Offset of status register */ |
#define PCI_PRI_ENABLE 0x0001 /* Enable mask */ |
#define PCI_PRI_RESET 0x0002 /* Reset bit mask */ |
#define PCI_PRI_STATUS_RF 0x0001 /* Request Failure */ |
#define PCI_PRI_STATUS_UPRGI 0x0002 /* Unexpected PRG index */ |
#define PCI_PRI_STATUS_STOPPED 0x0100 /* PRI Stopped */ |
#define PCI_PRI_MAX_REQ_OFF 0x08 /* Cap offset for max reqs supported */ |
#define PCI_PRI_ALLOC_REQ_OFF 0x0c /* Cap offset for max reqs allowed */ |
/* PASID capability */ |
#define PCI_PASID_CAP 0x1b /* PASID capability ID */ |
#define PCI_PASID_CAP_OFF 0x04 /* PASID feature register */ |
#define PCI_PASID_CONTROL_OFF 0x06 /* PASID control register */ |
#define PCI_PASID_ENABLE 0x01 /* Enable/Supported bit */ |
#define PCI_PASID_EXEC 0x02 /* Exec permissions Enable/Supported */ |
#define PCI_PASID_PRIV 0x04 /* Priviledge Mode Enable/Support */ |
/* Single Root I/O Virtualization */ |
#define PCI_SRIOV_CAP 0x04 /* SR-IOV Capabilities */ |
#define PCI_SRIOV_CAP_VFM 0x01 /* VF Migration Capable */ |
/drivers/include/linux/poison.h |
---|
40,6 → 40,12 |
#define RED_INACTIVE 0x09F911029D74E35BULL /* when obj is inactive */ |
#define RED_ACTIVE 0xD84156C5635688C0ULL /* when obj is active */ |
#ifdef CONFIG_PHYS_ADDR_T_64BIT |
#define MEMBLOCK_INACTIVE 0x3a84fb0144c9e71bULL |
#else |
#define MEMBLOCK_INACTIVE 0x44c9e71bUL |
#endif |
#define SLUB_RED_INACTIVE 0xbb |
#define SLUB_RED_ACTIVE 0xcc |
/drivers/include/linux/string.h |
---|
114,6 → 114,7 |
#ifndef __HAVE_ARCH_MEMCHR |
extern void * memchr(const void *,int,__kernel_size_t); |
#endif |
void *memchr_inv(const void *s, int c, size_t n); |
extern char *kstrdup(const char *s, gfp_t gfp); |
extern char *kstrndup(const char *s, size_t len, gfp_t gfp); |
/drivers/include/linux/types.h |
---|
246,7 → 246,7 |
typedef unsigned char u8_t; |
typedef unsigned short u16_t; |
typedef unsigned int u32_t; |
typedef unsigned long u32_t; |
typedef unsigned long long u64_t; |
typedef unsigned int addr_t; |
/drivers/include/linux/wait.h |
---|
0,0 → 1,144 |
#ifndef _LINUX_WAIT_H |
#define _LINUX_WAIT_H |
typedef struct __wait_queue wait_queue_t; |
typedef struct __wait_queue_head wait_queue_head_t; |
struct __wait_queue |
{ |
struct list_head task_list; |
evhandle_t evnt; |
}; |
struct __wait_queue_head |
{ |
spinlock_t lock; |
struct list_head task_list; |
}; |
static inline void __add_wait_queue(wait_queue_head_t *head, wait_queue_t *new) |
{ |
list_add(&new->task_list, &head->task_list); |
} |
#define __wait_event(wq, condition) \ |
do { \ |
DEFINE_WAIT(__wait); \ |
\ |
for (;;) { \ |
prepare_to_wait(&wq, &__wait, TASK_UNINTERRUPTIBLE); \ |
if (condition) \ |
break; \ |
schedule(); \ |
} \ |
finish_wait(&wq, &__wait); \ |
} while (0) |
#define wait_event(wq, condition) \ |
do{ \ |
wait_queue_t __wait = { \ |
.task_list = LIST_HEAD_INIT(__wait.task_list), \ |
.evnt = CreateEvent(NULL, MANUAL_DESTROY), \ |
}; \ |
u32 flags; \ |
\ |
spin_lock_irqsave(&wq.lock, flags); \ |
if (list_empty(&__wait.task_list)) \ |
__add_wait_queue(&wq, &__wait); \ |
spin_unlock_irqrestore(&wq.lock, flags); \ |
\ |
for(;;){ \ |
if (condition) \ |
break; \ |
WaitEvent(__wait.evnt); \ |
}; \ |
if (!list_empty_careful(&__wait.task_list)) { \ |
spin_lock_irqsave(&wq.lock, flags); \ |
list_del_init(&__wait.task_list); \ |
spin_unlock_irqrestore(&wq.lock, flags); \ |
}; \ |
DestroyEvent(__wait.evnt); \ |
} while (0) |
static inline |
void wake_up_all(wait_queue_head_t *q) |
{ |
wait_queue_t *curr; |
unsigned long flags; |
spin_lock_irqsave(&q->lock, flags); |
list_for_each_entry(curr, &q->task_list, task_list) |
{ |
kevent_t event; |
event.code = -1; |
RaiseEvent(curr->evnt, 0, &event); |
} |
spin_unlock_irqrestore(&q->lock, flags); |
} |
static inline void |
init_waitqueue_head(wait_queue_head_t *q) |
{ |
spin_lock_init(&q->lock); |
INIT_LIST_HEAD(&q->task_list); |
}; |
/* |
* Workqueue flags and constants. For details, please refer to |
* Documentation/workqueue.txt. |
*/ |
enum { |
WQ_NON_REENTRANT = 1 << 0, /* guarantee non-reentrance */ |
WQ_UNBOUND = 1 << 1, /* not bound to any cpu */ |
WQ_FREEZABLE = 1 << 2, /* freeze during suspend */ |
WQ_MEM_RECLAIM = 1 << 3, /* may be used for memory reclaim */ |
WQ_HIGHPRI = 1 << 4, /* high priority */ |
WQ_CPU_INTENSIVE = 1 << 5, /* cpu instensive workqueue */ |
WQ_DRAINING = 1 << 6, /* internal: workqueue is draining */ |
WQ_RESCUER = 1 << 7, /* internal: workqueue has rescuer */ |
WQ_MAX_ACTIVE = 512, /* I like 512, better ideas? */ |
WQ_MAX_UNBOUND_PER_CPU = 4, /* 4 * #cpus for unbound wq */ |
WQ_DFL_ACTIVE = WQ_MAX_ACTIVE / 2, |
}; |
struct work_struct; |
struct workqueue_struct { |
spinlock_t lock; |
struct list_head worklist; |
}; |
typedef void (*work_func_t)(struct work_struct *work); |
struct work_struct { |
struct list_head entry; |
struct workqueue_struct *data; |
work_func_t func; |
}; |
struct delayed_work { |
struct work_struct work; |
}; |
struct workqueue_struct *alloc_workqueue_key(const char *fmt, |
unsigned int flags, int max_active); |
int queue_delayed_work(struct workqueue_struct *wq, |
struct delayed_work *dwork, unsigned long delay); |
#define INIT_DELAYED_WORK(_work, _func) \ |
do { \ |
INIT_LIST_HEAD(&(_work)->work.entry); \ |
(_work)->work.func = _func; \ |
} while (0) |
#endif |