replace always-inline attributes
There is a macro __rte_always_inline, forcing functions to be inlined, which is now used where appropriate for consistency. Signed-off-by: Thomas Monjalon <thomas@monjalon.net>
This commit is contained in:
parent
ef5baf3486
commit
33011cb3df
@ -9,7 +9,7 @@
|
||||
|
||||
/* See http://doc.dpdk.org/guides/tools/testeventdev.html for test details */
|
||||
|
||||
static inline __attribute__((always_inline)) void
|
||||
static __rte_always_inline void
|
||||
order_atq_process_stage_0(struct rte_event *const ev)
|
||||
{
|
||||
ev->sub_event_type = 1; /* move to stage 1 (atomic) on the same queue */
|
||||
|
@ -62,7 +62,7 @@ order_nb_event_ports(struct evt_options *opt)
|
||||
return evt_nr_active_lcores(opt->wlcores) + 1 /* producer */;
|
||||
}
|
||||
|
||||
static inline __attribute__((always_inline)) void
|
||||
static __rte_always_inline void
|
||||
order_process_stage_1(struct test_order *const t,
|
||||
struct rte_event *const ev, const uint32_t nb_flows,
|
||||
uint32_t *const expected_flow_seq,
|
||||
@ -87,7 +87,7 @@ order_process_stage_1(struct test_order *const t,
|
||||
rte_atomic64_sub(outstand_pkts, 1);
|
||||
}
|
||||
|
||||
static inline __attribute__((always_inline)) void
|
||||
static __rte_always_inline void
|
||||
order_process_stage_invalid(struct test_order *const t,
|
||||
struct rte_event *const ev)
|
||||
{
|
||||
|
@ -9,7 +9,7 @@
|
||||
|
||||
/* See http://doc.dpdk.org/guides/tools/testeventdev.html for test details */
|
||||
|
||||
static inline __attribute__((always_inline)) void
|
||||
static __rte_always_inline void
|
||||
order_queue_process_stage_0(struct rte_event *const ev)
|
||||
{
|
||||
ev->queue_id = 1; /* q1 atomic queue */
|
||||
|
@ -14,7 +14,7 @@ atq_nb_event_queues(struct evt_options *opt)
|
||||
rte_eth_dev_count_avail() : evt_nr_active_lcores(opt->plcores);
|
||||
}
|
||||
|
||||
static inline __attribute__((always_inline)) void
|
||||
static __rte_always_inline void
|
||||
atq_mark_fwd_latency(struct rte_event *const ev)
|
||||
{
|
||||
if (unlikely(ev->sub_event_type == 0)) {
|
||||
@ -24,7 +24,7 @@ atq_mark_fwd_latency(struct rte_event *const ev)
|
||||
}
|
||||
}
|
||||
|
||||
static inline __attribute__((always_inline)) void
|
||||
static __rte_always_inline void
|
||||
atq_fwd_event(struct rte_event *const ev, uint8_t *const sched_type_list,
|
||||
const uint8_t nb_stages)
|
||||
{
|
||||
|
@ -91,7 +91,7 @@ struct perf_elt {
|
||||
printf("%s(): lcore %d dev_id %d port=%d\n", __func__,\
|
||||
rte_lcore_id(), dev, port)
|
||||
|
||||
static inline __attribute__((always_inline)) int
|
||||
static __rte_always_inline int
|
||||
perf_process_last_stage(struct rte_mempool *const pool,
|
||||
struct rte_event *const ev, struct worker_data *const w,
|
||||
void *bufs[], int const buf_sz, uint8_t count)
|
||||
@ -107,7 +107,7 @@ perf_process_last_stage(struct rte_mempool *const pool,
|
||||
return count;
|
||||
}
|
||||
|
||||
static inline __attribute__((always_inline)) uint8_t
|
||||
static __rte_always_inline uint8_t
|
||||
perf_process_last_stage_latency(struct rte_mempool *const pool,
|
||||
struct rte_event *const ev, struct worker_data *const w,
|
||||
void *bufs[], int const buf_sz, uint8_t count)
|
||||
|
@ -15,7 +15,7 @@ perf_queue_nb_event_queues(struct evt_options *opt)
|
||||
return nb_prod * opt->nb_stages;
|
||||
}
|
||||
|
||||
static inline __attribute__((always_inline)) void
|
||||
static __rte_always_inline void
|
||||
mark_fwd_latency(struct rte_event *const ev,
|
||||
const uint8_t nb_stages)
|
||||
{
|
||||
@ -26,7 +26,7 @@ mark_fwd_latency(struct rte_event *const ev,
|
||||
}
|
||||
}
|
||||
|
||||
static inline __attribute__((always_inline)) void
|
||||
static __rte_always_inline void
|
||||
fwd_event(struct rte_event *const ev, uint8_t *const sched_type_list,
|
||||
const uint8_t nb_stages)
|
||||
{
|
||||
|
@ -2585,7 +2585,7 @@ end_of_tx:
|
||||
return nb_tx;
|
||||
}
|
||||
|
||||
static inline int __attribute__((always_inline))
|
||||
static __rte_always_inline int
|
||||
ice_tx_free_bufs(struct ice_tx_queue *txq)
|
||||
{
|
||||
struct ice_tx_entry *txep;
|
||||
|
@ -25,7 +25,7 @@ extern "C" {
|
||||
#define RTE_XABORT_NESTED (1 << 5)
|
||||
#define RTE_XABORT_CODE(x) (((x) >> 24) & 0xff)
|
||||
|
||||
static __attribute__((__always_inline__)) inline
|
||||
static __rte_always_inline
|
||||
unsigned int rte_xbegin(void)
|
||||
{
|
||||
unsigned int ret = RTE_XBEGIN_STARTED;
|
||||
@ -34,7 +34,7 @@ unsigned int rte_xbegin(void)
|
||||
return ret;
|
||||
}
|
||||
|
||||
static __attribute__((__always_inline__)) inline
|
||||
static __rte_always_inline
|
||||
void rte_xend(void)
|
||||
{
|
||||
asm volatile(".byte 0x0f,0x01,0xd5" ::: "memory");
|
||||
@ -45,7 +45,7 @@ void rte_xend(void)
|
||||
asm volatile(".byte 0xc6,0xf8,%P0" :: "i" (status) : "memory"); \
|
||||
} while (0)
|
||||
|
||||
static __attribute__((__always_inline__)) inline
|
||||
static __rte_always_inline
|
||||
int rte_xtest(void)
|
||||
{
|
||||
unsigned char out;
|
||||
|
@ -52,13 +52,13 @@ set_power_freq(int lcore_id, enum freq_val freq, bool specific_freq)
|
||||
}
|
||||
|
||||
|
||||
static inline void __attribute__((always_inline))
|
||||
static __rte_always_inline void
|
||||
exit_training_state(struct priority_worker *poll_stats)
|
||||
{
|
||||
RTE_SET_USED(poll_stats);
|
||||
}
|
||||
|
||||
static inline void __attribute__((always_inline))
|
||||
static __rte_always_inline void
|
||||
enter_training_state(struct priority_worker *poll_stats)
|
||||
{
|
||||
poll_stats->iter_counter = 0;
|
||||
@ -66,7 +66,7 @@ enter_training_state(struct priority_worker *poll_stats)
|
||||
poll_stats->queue_state = TRAINING;
|
||||
}
|
||||
|
||||
static inline void __attribute__((always_inline))
|
||||
static __rte_always_inline void
|
||||
enter_normal_state(struct priority_worker *poll_stats)
|
||||
{
|
||||
/* Clear the averages arrays and strs */
|
||||
@ -86,7 +86,7 @@ enter_normal_state(struct priority_worker *poll_stats)
|
||||
poll_stats->thresh[HGH].threshold_percent = high_to_med_threshold;
|
||||
}
|
||||
|
||||
static inline void __attribute__((always_inline))
|
||||
static __rte_always_inline void
|
||||
enter_busy_state(struct priority_worker *poll_stats)
|
||||
{
|
||||
memset(poll_stats->edpi_av, 0, sizeof(poll_stats->edpi_av));
|
||||
@ -101,14 +101,14 @@ enter_busy_state(struct priority_worker *poll_stats)
|
||||
set_power_freq(poll_stats->lcore_id, HGH, false);
|
||||
}
|
||||
|
||||
static inline void __attribute__((always_inline))
|
||||
static __rte_always_inline void
|
||||
enter_purge_state(struct priority_worker *poll_stats)
|
||||
{
|
||||
poll_stats->iter_counter = 0;
|
||||
poll_stats->queue_state = LOW_PURGE;
|
||||
}
|
||||
|
||||
static inline void __attribute__((always_inline))
|
||||
static __rte_always_inline void
|
||||
set_state(struct priority_worker *poll_stats,
|
||||
enum queue_state new_state)
|
||||
{
|
||||
@ -131,7 +131,7 @@ set_state(struct priority_worker *poll_stats,
|
||||
}
|
||||
}
|
||||
|
||||
static inline void __attribute__((always_inline))
|
||||
static __rte_always_inline void
|
||||
set_policy(struct priority_worker *poll_stats,
|
||||
struct ep_policy *policy)
|
||||
{
|
||||
@ -204,7 +204,7 @@ update_training_stats(struct priority_worker *poll_stats,
|
||||
}
|
||||
}
|
||||
|
||||
static inline uint32_t __attribute__((always_inline))
|
||||
static __rte_always_inline uint32_t
|
||||
update_stats(struct priority_worker *poll_stats)
|
||||
{
|
||||
uint64_t tot_edpi = 0, tot_ppi = 0;
|
||||
@ -249,7 +249,7 @@ update_stats(struct priority_worker *poll_stats)
|
||||
}
|
||||
|
||||
|
||||
static inline void __attribute__((always_inline))
|
||||
static __rte_always_inline void
|
||||
update_stats_normal(struct priority_worker *poll_stats)
|
||||
{
|
||||
uint32_t percent;
|
||||
|
Loading…
x
Reference in New Issue
Block a user