X-Git-Url: http://git.ozo.com/?a=blobdiff_plain;f=batman-adv%2Ffiles%2Fcompat-hacks.h;h=8c8a8899995756ebe7620a16aab6fc1b1bea3c6c;hb=1c0419755d9d7c0f723fa3aa7fe166db07172db2;hp=a910c2d9ae2db784656071119e85bf47db9988a2;hpb=fd4a2d7b6d781169ea4581dc211c14921aced9bb;p=lede-routing%2F.git diff --git a/batman-adv/files/compat-hacks.h b/batman-adv/files/compat-hacks.h index a910c2d..8c8a889 100644 --- a/batman-adv/files/compat-hacks.h +++ b/batman-adv/files/compat-hacks.h @@ -1,5 +1,7 @@ /* Please avoid adding hacks here - instead add it to mac80211/backports.git */ +#undef CONFIG_MODULE_STRIPPED + #include /* LINUX_VERSION_CODE */ #include @@ -32,10 +34,35 @@ #endif /* < KERNEL_VERSION(4, 5, 0) */ +#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 0, 0) + +/* wild hack for batadv_getlink_net only */ +#define get_link_net get_xstats_size || 1 ? fallback_net : (struct net*)netdev->rtnl_link_ops->get_xstats_size + +#endif /* < KERNEL_VERSION(4, 0, 0) */ + + +#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 2, 0) + +struct sk_buff *skb_checksum_trimmed(struct sk_buff *skb, + unsigned int transport_len, + __sum16(*skb_chkf)(struct sk_buff *skb)); + +int ip_mc_check_igmp(struct sk_buff *skb, struct sk_buff **skb_trimmed); + +int ipv6_mc_check_mld(struct sk_buff *skb, struct sk_buff **skb_trimmed); + +#endif /* < KERNEL_VERSION(4, 2, 0) */ + #if LINUX_VERSION_CODE < KERNEL_VERSION(4, 3, 0) #define IFF_NO_QUEUE 0; dev->tx_queue_len = 0 +static inline bool hlist_fake(struct hlist_node *h) +{ + return h->pprev == &h->next; +} + #endif /* < KERNEL_VERSION(4, 3, 0) */ #if LINUX_VERSION_CODE < KERNEL_VERSION(4, 6, 0) @@ -84,3 +111,180 @@ static inline void batadv_netif_trans_update(struct net_device *dev) } #endif /* < KERNEL_VERSION(4, 7, 0) */ + + +#include_next + +#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 7, 0) + +#include_next + +static inline bool batadv_nla_need_padding_for_64bit(struct sk_buff *skb); + +static inline int batadv_nla_align_64bit(struct sk_buff *skb, int padattr) +{ + if (batadv_nla_need_padding_for_64bit(skb) && + !nla_reserve(skb, padattr, 0)) + return -EMSGSIZE; + + return 0; +} + +static inline struct nlattr *batadv__nla_reserve_64bit(struct sk_buff *skb, + int attrtype, + int attrlen, int padattr) +{ + if (batadv_nla_need_padding_for_64bit(skb)) + batadv_nla_align_64bit(skb, padattr); + + return __nla_reserve(skb, attrtype, attrlen); +} + +static inline void batadv__nla_put_64bit(struct sk_buff *skb, int attrtype, + int attrlen, const void *data, + int padattr) +{ + struct nlattr *nla; + + nla = batadv__nla_reserve_64bit(skb, attrtype, attrlen, padattr); + memcpy(nla_data(nla), data, attrlen); +} + +static inline bool batadv_nla_need_padding_for_64bit(struct sk_buff *skb) +{ +#ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS + /* The nlattr header is 4 bytes in size, that's why we test + * if the skb->data _is_ aligned. A NOP attribute, plus + * nlattr header for next attribute, will make nla_data() + * 8-byte aligned. + */ + if (IS_ALIGNED((unsigned long)skb_tail_pointer(skb), 8)) + return true; +#endif + return false; +} + +static inline int batadv_nla_total_size_64bit(int payload) +{ + return NLA_ALIGN(nla_attr_size(payload)) +#ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS + + NLA_ALIGN(nla_attr_size(0)) +#endif + ; +} + +static inline int batadv_nla_put_64bit(struct sk_buff *skb, int attrtype, + int attrlen, const void *data, + int padattr) +{ + size_t len; + + if (batadv_nla_need_padding_for_64bit(skb)) + len = batadv_nla_total_size_64bit(attrlen); + else + len = nla_total_size(attrlen); + if (unlikely(skb_tailroom(skb) < len)) + return -EMSGSIZE; + + batadv__nla_put_64bit(skb, attrtype, attrlen, data, padattr); + return 0; +} + +#define nla_put_u64_64bit(_skb, _attrtype, _value, _padattr) \ + batadv_nla_put_u64_64bit(_skb, _attrtype, _value, _padattr) +static inline int batadv_nla_put_u64_64bit(struct sk_buff *skb, int attrtype, + u64 value, int padattr) +{ + return batadv_nla_put_64bit(skb, attrtype, sizeof(u64), &value, + padattr); +} + +#endif /* < KERNEL_VERSION(4, 7, 0) */ + + +#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 10, 0) + +#include_next + +/* hack for netlink.c which marked the family ops as ro */ +#ifdef __ro_after_init +#undef __ro_after_init +#endif +#define __ro_after_init + +#endif /* < KERNEL_VERSION(4, 10, 0) */ + +/* */ + +#include +#include_next + +#include + +#ifdef DECLARE_EWMA +#undef DECLARE_EWMA +#endif /* DECLARE_EWMA */ + +/* + * Exponentially weighted moving average (EWMA) + * + * This implements a fixed-precision EWMA algorithm, with both the + * precision and fall-off coefficient determined at compile-time + * and built into the generated helper funtions. + * + * The first argument to the macro is the name that will be used + * for the struct and helper functions. + * + * The second argument, the precision, expresses how many bits are + * used for the fractional part of the fixed-precision values. + * + * The third argument, the weight reciprocal, determines how the + * new values will be weighed vs. the old state, new values will + * get weight 1/weight_rcp and old values 1-1/weight_rcp. Note + * that this parameter must be a power of two for efficiency. + */ + +#define DECLARE_EWMA(name, _precision, _weight_rcp) \ + struct ewma_##name { \ + unsigned long internal; \ + }; \ + static inline void ewma_##name##_init(struct ewma_##name *e) \ + { \ + BUILD_BUG_ON(!__builtin_constant_p(_precision)); \ + BUILD_BUG_ON(!__builtin_constant_p(_weight_rcp)); \ + /* \ + * Even if you want to feed it just 0/1 you should have \ + * some bits for the non-fractional part... \ + */ \ + BUILD_BUG_ON((_precision) > 30); \ + BUILD_BUG_ON_NOT_POWER_OF_2(_weight_rcp); \ + e->internal = 0; \ + } \ + static inline unsigned long \ + ewma_##name##_read(struct ewma_##name *e) \ + { \ + BUILD_BUG_ON(!__builtin_constant_p(_precision)); \ + BUILD_BUG_ON(!__builtin_constant_p(_weight_rcp)); \ + BUILD_BUG_ON((_precision) > 30); \ + BUILD_BUG_ON_NOT_POWER_OF_2(_weight_rcp); \ + return e->internal >> (_precision); \ + } \ + static inline void ewma_##name##_add(struct ewma_##name *e, \ + unsigned long val) \ + { \ + unsigned long internal = ACCESS_ONCE(e->internal); \ + unsigned long weight_rcp = ilog2(_weight_rcp); \ + unsigned long precision = _precision; \ + \ + BUILD_BUG_ON(!__builtin_constant_p(_precision)); \ + BUILD_BUG_ON(!__builtin_constant_p(_weight_rcp)); \ + BUILD_BUG_ON((_precision) > 30); \ + BUILD_BUG_ON_NOT_POWER_OF_2(_weight_rcp); \ + \ + ACCESS_ONCE(e->internal) = internal ? \ + (((internal << weight_rcp) - internal) + \ + (val << precision)) >> weight_rcp : \ + (val << precision); \ + } + +/* */