/*
 * Copyright (C) 2008 The Android Open Source Project
 * All rights reserved.
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 *  * Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 *  * Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in
 *    the documentation and/or other materials provided with the
 *    distribution.
 *
 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
 * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
 * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
 * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
 * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 * SUCH DAMAGE.
 */
#ifndef _SCHED_H_
#define _SCHED_H_

#include <sys/cdefs.h>
#include <sys/time.h>

__BEGIN_DECLS

#define SCHED_NORMAL            0
#define SCHED_OTHER             0
#define SCHED_FIFO              1
#define SCHED_RR                2

struct sched_param {
    int sched_priority;
};

extern int sched_setscheduler(pid_t, int, const struct sched_param *);
extern int sched_getscheduler(pid_t);
extern int sched_yield(void);
extern int sched_get_priority_max(int policy);
extern int sched_get_priority_min(int policy);
extern int sched_setparam(pid_t, const struct sched_param *);
extern int sched_getparam(pid_t, struct sched_param *);
extern int sched_rr_get_interval(pid_t pid, struct timespec *tp);

#define CLONE_VM             0x00000100
#define CLONE_FS             0x00000200
#define CLONE_FILES          0x00000400
#define CLONE_SIGHAND        0x00000800
#define CLONE_PTRACE         0x00002000
#define CLONE_VFORK          0x00004000
#define CLONE_PARENT         0x00008000
#define CLONE_THREAD         0x00010000
#define CLONE_NEWNS          0x00020000
#define CLONE_SYSVSEM        0x00040000
#define CLONE_SETTLS         0x00080000
#define CLONE_PARENT_SETTID  0x00100000
#define CLONE_CHILD_CLEARTID 0x00200000
#define CLONE_DETACHED       0x00400000
#define CLONE_UNTRACED       0x00800000
#define CLONE_CHILD_SETTID   0x01000000
#define CLONE_STOPPED        0x02000000

#ifdef _GNU_SOURCE
extern int clone(int (*fn)(void *), void *child_stack, int flags, void*  arg, ...);
extern int unshare(int);
#endif

/* Support for cpu thread affinity */
#ifdef _GNU_SOURCE

extern int sched_getcpu(void);


/* Our implementation supports up to 32 independent CPUs, which is also
 * the maximum supported by the kernel at the moment. GLibc uses 1024 by
 * default.
 *
 * If you want to use more than that, you should use CPU_ALLOC() / CPU_FREE()
 * and the CPU_XXX_S() macro variants.
 */
#define CPU_SETSIZE   32

#define __CPU_BITTYPE    unsigned long int  /* mandated by the kernel  */
#define __CPU_BITSHIFT   5                  /* should be log2(BITTYPE) */
#define __CPU_BITS       (1 << __CPU_BITSHIFT)
#define __CPU_ELT(x)     ((x) >> __CPU_BITSHIFT)
#define __CPU_MASK(x)    ((__CPU_BITTYPE)1 << ((x) & (__CPU_BITS-1)))

typedef struct {
    __CPU_BITTYPE  __bits[ CPU_SETSIZE / __CPU_BITS ];
} cpu_set_t;

extern int sched_setaffinity(pid_t pid, size_t setsize, const cpu_set_t* set);

extern int sched_getaffinity(pid_t pid, size_t setsize, cpu_set_t* set);

/* Provide optimized implementation for 32-bit cpu_set_t */
#if CPU_SETSIZE == __CPU_BITS

#  define CPU_ZERO(set_)   \
    do{ \
        (set_)->__bits[0] = 0; \
    }while(0)

#  define CPU_SET(cpu_,set_) \
    do {\
        size_t __cpu = (cpu_); \
        if (__cpu < CPU_SETSIZE) \
            (set_)->__bits[0] |= __CPU_MASK(__cpu); \
    }while (0)

#  define CPU_CLR(cpu_,set_) \
    do {\
        size_t __cpu = (cpu_); \
        if (__cpu < CPU_SETSIZE) \
            (set_)->__bits[0] &= ~__CPU_MASK(__cpu); \
    }while (0)

#  define CPU_ISSET(cpu_, set_) \
    (__extension__({\
        size_t  __cpu = (cpu_); \
        (cpu_ < CPU_SETSIZE) \
            ? ((set_)->__bits[0] & __CPU_MASK(__cpu)) != 0 \
            : 0; \
    }))

#  define CPU_EQUAL(set1_, set2_) \
    ((set1_)->__bits[0] == (set2_)->__bits[0])

#  define __CPU_OP(dst_, set1_, set2_, op_) \
    do { \
        (dst_)->__bits[0] = (set1_)->__bits[0] op_ (set2_)->__bits[0]; \
    } while (0)

#  define CPU_COUNT(set_)  __builtin_popcountl((set_)->__bits[0])

#else /* CPU_SETSIZE != __CPU_BITS */

#  define CPU_ZERO(set_)          CPU_ZERO_S(sizeof(cpu_set_t), set_)
#  define CPU_SET(cpu_,set_)      CPU_SET_S(cpu_,sizeof(cpu_set_t),set_)
#  define CPU_CLR(cpu_,set_)      CPU_CLR_S(cpu_,sizeof(cpu_set_t),set_)
#  define CPU_ISSET(cpu_,set_)    CPU_ISSET_S(cpu_,sizeof(cpu_set_t),set_)
#  define CPU_COUNT(set_)         CPU_COUNT_S(sizeof(cpu_set_t),set_)
#  define CPU_EQUAL(set1_,set2_)  CPU_EQUAL_S(sizeof(cpu_set_t),set1_,set2_)

#  define __CPU_OP(dst_,set1_,set2_,op_)  __CPU_OP_S(sizeof(cpu_set_t),dst_,set1_,set2_,op_)

#endif /* CPU_SETSIZE != __CPU_BITS */

#define CPU_AND(set1_,set2_)   __CPU_OP(set1_,set2_,&)
#define CPU_OR(set1_,set2_)    __CPU_OP(set1_,set2_,|)
#define CPU_XOR(set1_,set2_)   __CPU_OP(set1_,set2_,^)

/* Support for dynamically-allocated cpu_set_t */

#define CPU_ALLOC_SIZE(count) \
    __CPU_ELT((count) + (__CPU_BITS-1))*sizeof(__CPU_BITTYPE)

#define CPU_ALLOC(count)   __sched_cpualloc((count));
#define CPU_FREE(set)      __sched_cpufree((set))

extern cpu_set_t* __sched_cpualloc(size_t count);
extern void       __sched_cpufree(cpu_set_t* set);

#define CPU_ZERO_S(setsize_,set_)  \
    do { \
        size_t __nn = 0; \
        size_t __nn_max = (setsize_)/sizeof(__CPU_BITTYPE); \
        for (; __nn < __nn_max; __nn++) \
            (set_)->__bits[__nn] = 0; \
    } while (0)

#define CPU_SET_S(cpu_,setsize_,set_) \
    do { \
        size_t __cpu = (cpu_); \
        if (__cpu < 8*(setsize_)) \
            (set_)->__bits[__CPU_ELT(__cpu)] |= __CPU_MASK(__cpu); \
    } while (0)

#define CPU_CLR_S(cpu_,setsize_,set_) \
    do { \
        size_t __cpu = (cpu_); \
        if (__cpu < 8*(setsize_)) \
            (set_)->__bits[__CPU_ELT(__cpu)] &= ~__CPU_MASK(__cpu); \
    } while (0)

#define CPU_ISSET_S(cpu_, setsize_, set_) \
    (__extension__ ({ \
        size_t __cpu = (cpu_); \
        (__cpu < 8*(setsize_)) \
          ? ((set_)->__bits[__CPU_ELT(__cpu)] & __CPU_MASK(__cpu)) != 0 \
          : 0; \
    }))

#define CPU_EQUAL_S(setsize_, set1_, set2_) \
    (__extension__ ({ \
        __const __CPU_BITTYPE* __src1 = (set1_)->__bits; \
        __const __CPU_BITTYPE* __src2 = (set2_)->__bits; \
        size_t __nn = 0, __nn_max = (setsize_)/sizeof(__CPU_BITTYPE); \
        for (; __nn < __nn_max; __nn++) { \
            if (__src1[__nn] != __src2[__nn]) \
                break; \
        } \
        __nn == __nn_max; \
    }))

#define __CPU_OP_S(setsize_, dstset_, srcset1_, srcset2_, op) \
    do { \
        cpu_set_t* __dst = (dstset); \
        const __CPU_BITTYPE* __src1 = (srcset1)->__bits; \
        const __CPU_BITTYPE* __src2 = (srcset2)->__bits; \
        size_t __nn = 0, __nn_max = (setsize_)/sizeof(__CPU_BITTYPE); \
        for (; __nn < __nn_max; __nn++) \
            (__dst)->__bits[__nn] = __src1[__nn] op __src2[__nn]; \
    } while (0)

#define CPU_COUNT_S(setsize_, set_) \
    __sched_cpucount((setsize_), (set_))

extern int __sched_cpucount(size_t setsize, cpu_set_t* set);

#endif /* _GNU_SOURCE */

__END_DECLS

#endif /* _SCHED_H_ */