Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
- index 1359c5fb1f..827d271169 100644
- --- a/lib/vhost/virtio_net.c
- +++ b/lib/vhost/virtio_net.c
- @@ -26,7 +26,7 @@
- #define MAX_BATCH_LEN 256
- -static __rte_always_inline uint16_t
- +static uint16_t
- async_poll_dequeue_completed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts, uint16_t count, int16_t dma_id,
- uint16_t vchan_id, bool legacy_ol_flags);
- @@ -34,13 +34,13 @@ async_poll_dequeue_completed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- /* DMA device copy operation tracking array. */
- struct async_dma_info dma_copy_track[RTE_DMADEV_DEFAULT_MAX];
- -static __rte_always_inline bool
- +static bool
- rxvq_is_mergeable(struct virtio_net *dev)
- {
- return dev->features & (1ULL << VIRTIO_NET_F_MRG_RXBUF);
- }
- -static __rte_always_inline bool
- +static bool
- virtio_net_is_inorder(struct virtio_net *dev)
- {
- return dev->features & (1ULL << VIRTIO_F_IN_ORDER);
- @@ -52,7 +52,7 @@ is_valid_virt_queue_idx(uint32_t idx, int is_tx, uint32_t nr_vring)
- return (is_tx ^ (idx & 1)) == 0 && idx < nr_vring;
- }
- -static inline void
- +static void
- vhost_queue_stats_update(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts, uint16_t count)
- __rte_shared_locks_required(&vq->access_lock)
- @@ -98,7 +98,7 @@ vhost_queue_stats_update(struct virtio_net *dev, struct vhost_virtqueue *vq,
- }
- }
- -static __rte_always_inline int64_t
- +static int64_t
- vhost_async_dma_transfer_one(struct virtio_net *dev, struct vhost_virtqueue *vq,
- int16_t dma_id, uint16_t vchan_id, uint16_t flag_idx,
- struct vhost_iov_iter *pkt)
- @@ -148,7 +148,7 @@ vhost_async_dma_transfer_one(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return nr_segs;
- }
- -static __rte_always_inline uint16_t
- +static uint16_t
- vhost_async_dma_transfer(struct virtio_net *dev, struct vhost_virtqueue *vq,
- int16_t dma_id, uint16_t vchan_id, uint16_t head_idx,
- struct vhost_iov_iter *pkts, uint16_t nr_pkts)
- @@ -180,7 +180,7 @@ vhost_async_dma_transfer(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return pkt_idx;
- }
- -static __rte_always_inline uint16_t
- +static uint16_t
- vhost_async_dma_check_completed(struct virtio_net *dev, int16_t dma_id, uint16_t vchan_id,
- uint16_t max_pkts)
- {
- @@ -231,7 +231,7 @@ vhost_async_dma_check_completed(struct virtio_net *dev, int16_t dma_id, uint16_t
- return nr_copies;
- }
- -static inline void
- +static void
- do_data_copy_enqueue(struct virtio_net *dev, struct vhost_virtqueue *vq)
- __rte_shared_locks_required(&vq->iotlb_lock)
- {
- @@ -249,7 +249,7 @@ do_data_copy_enqueue(struct virtio_net *dev, struct vhost_virtqueue *vq)
- vq->batch_copy_nb_elems = 0;
- }
- -static inline void
- +static void
- do_data_copy_dequeue(struct vhost_virtqueue *vq)
- {
- struct batch_copy_elem *elem = vq->batch_copy_elems;
- @@ -262,7 +262,7 @@ do_data_copy_dequeue(struct vhost_virtqueue *vq)
- vq->batch_copy_nb_elems = 0;
- }
- -static __rte_always_inline void
- +static void
- do_flush_shadow_used_ring_split(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- uint16_t to, uint16_t from, uint16_t size)
- @@ -275,7 +275,7 @@ do_flush_shadow_used_ring_split(struct virtio_net *dev,
- size * sizeof(struct vring_used_elem));
- }
- -static __rte_always_inline void
- +static void
- flush_shadow_used_ring_split(struct virtio_net *dev, struct vhost_virtqueue *vq)
- {
- uint16_t used_idx = vq->last_used_idx & (vq->size - 1);
- @@ -305,7 +305,7 @@ flush_shadow_used_ring_split(struct virtio_net *dev, struct vhost_virtqueue *vq)
- sizeof(vq->used->idx));
- }
- -static __rte_always_inline void
- +static void
- update_shadow_used_ring_split(struct vhost_virtqueue *vq,
- uint16_t desc_idx, uint32_t len)
- {
- @@ -315,7 +315,7 @@ update_shadow_used_ring_split(struct vhost_virtqueue *vq,
- vq->shadow_used_split[i].len = len;
- }
- -static __rte_always_inline void
- +static void
- vhost_flush_enqueue_shadow_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq)
- {
- @@ -379,7 +379,7 @@ vhost_flush_enqueue_shadow_packed(struct virtio_net *dev,
- vhost_log_cache_sync(dev, vq);
- }
- -static __rte_always_inline void
- +static void
- vhost_flush_dequeue_shadow_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq)
- {
- @@ -398,7 +398,7 @@ vhost_flush_dequeue_shadow_packed(struct virtio_net *dev,
- vhost_log_cache_sync(dev, vq);
- }
- -static __rte_always_inline void
- +static void
- vhost_flush_enqueue_batch_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- uint64_t *lens,
- @@ -434,7 +434,7 @@ vhost_flush_enqueue_batch_packed(struct virtio_net *dev,
- vq_inc_last_used_packed(vq, PACKED_BATCH_SIZE);
- }
- -static __rte_always_inline void
- +static void
- vhost_async_shadow_enqueue_packed_batch(struct vhost_virtqueue *vq,
- uint64_t *lens,
- uint16_t *ids)
- @@ -453,7 +453,7 @@ vhost_async_shadow_enqueue_packed_batch(struct vhost_virtqueue *vq,
- }
- }
- -static __rte_always_inline void
- +static void
- vhost_async_shadow_dequeue_packed_batch(struct vhost_virtqueue *vq, uint16_t *ids)
- __rte_shared_locks_required(&vq->access_lock)
- {
- @@ -471,7 +471,7 @@ vhost_async_shadow_dequeue_packed_batch(struct vhost_virtqueue *vq, uint16_t *id
- }
- }
- -static __rte_always_inline void
- +static void
- vhost_shadow_dequeue_batch_packed_inorder(struct vhost_virtqueue *vq,
- uint16_t id)
- {
- @@ -489,7 +489,7 @@ vhost_shadow_dequeue_batch_packed_inorder(struct vhost_virtqueue *vq,
- vq_inc_last_used_packed(vq, PACKED_BATCH_SIZE);
- }
- -static __rte_always_inline void
- +static void
- vhost_shadow_dequeue_batch_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- uint16_t *ids)
- @@ -529,7 +529,7 @@ vhost_shadow_dequeue_batch_packed(struct virtio_net *dev,
- vq_inc_last_used_packed(vq, PACKED_BATCH_SIZE);
- }
- -static __rte_always_inline void
- +static void
- vhost_shadow_dequeue_single_packed(struct vhost_virtqueue *vq,
- uint16_t buf_id,
- uint16_t count)
- @@ -561,7 +561,7 @@ vhost_shadow_dequeue_single_packed(struct vhost_virtqueue *vq,
- vq_inc_last_used_packed(vq, count);
- }
- -static __rte_always_inline void
- +static void
- vhost_shadow_dequeue_single_packed_inorder(struct vhost_virtqueue *vq,
- uint16_t buf_id,
- uint16_t count)
- @@ -589,7 +589,7 @@ vhost_shadow_dequeue_single_packed_inorder(struct vhost_virtqueue *vq,
- vq_inc_last_used_packed(vq, count);
- }
- -static __rte_always_inline void
- +static void
- vhost_shadow_enqueue_packed(struct vhost_virtqueue *vq,
- uint32_t *len,
- uint16_t *id,
- @@ -611,7 +611,7 @@ vhost_shadow_enqueue_packed(struct vhost_virtqueue *vq,
- }
- }
- -static __rte_always_inline void
- +static void
- vhost_async_shadow_enqueue_packed(struct vhost_virtqueue *vq,
- uint32_t *len,
- uint16_t *id,
- @@ -632,7 +632,7 @@ vhost_async_shadow_enqueue_packed(struct vhost_virtqueue *vq,
- }
- }
- -static __rte_always_inline void
- +static void
- vhost_shadow_enqueue_single_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- uint32_t *len,
- @@ -655,7 +655,7 @@ vhost_shadow_enqueue_single_packed(struct virtio_net *dev,
- (var) = (val); \
- } while (0)
- -static __rte_always_inline void
- +static void
- virtio_enqueue_offload(struct rte_mbuf *m_buf, struct virtio_net_hdr *net_hdr)
- {
- uint64_t csum_l4 = m_buf->ol_flags & RTE_MBUF_F_TX_L4_MASK;
- @@ -727,7 +727,7 @@ virtio_enqueue_offload(struct rte_mbuf *m_buf, struct virtio_net_hdr *net_hdr)
- }
- }
- -static __rte_always_inline int
- +static int
- map_one_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct buf_vector *buf_vec, uint16_t *vec_idx,
- uint64_t desc_iova, uint64_t desc_len, uint8_t perm)
- @@ -764,7 +764,7 @@ map_one_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return 0;
- }
- -static __rte_always_inline int
- +static int
- fill_vec_buf_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- uint32_t avail_idx, uint16_t *vec_idx,
- struct buf_vector *buf_vec, uint16_t *desc_chain_head,
- @@ -848,7 +848,7 @@ fill_vec_buf_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- /*
- * Returns -1 on fail, 0 on success
- */
- -static inline int
- +static int
- reserve_avail_buf_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- uint64_t size, struct buf_vector *buf_vec,
- uint16_t *num_buffers, uint16_t avail_head,
- @@ -899,7 +899,7 @@ reserve_avail_buf_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return 0;
- }
- -static __rte_always_inline int
- +static int
- fill_vec_buf_packed_indirect(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct vring_packed_desc *desc, uint16_t *vec_idx,
- @@ -958,7 +958,7 @@ fill_vec_buf_packed_indirect(struct virtio_net *dev,
- return 0;
- }
- -static __rte_always_inline int
- +static int
- fill_vec_buf_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- uint16_t avail_idx, uint16_t *desc_count,
- struct buf_vector *buf_vec, uint16_t *vec_idx,
- @@ -1025,7 +1025,7 @@ fill_vec_buf_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return 0;
- }
- -static __rte_noinline void
- +static void
- copy_vnet_hdr_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct buf_vector *buf_vec,
- struct virtio_net_hdr_mrg_rxbuf *hdr)
- @@ -1056,7 +1056,7 @@ copy_vnet_hdr_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
- }
- }
- -static __rte_always_inline int
- +static int
- async_iter_initialize(struct virtio_net *dev, struct vhost_async *async)
- {
- struct vhost_iov_iter *iter;
- @@ -1073,7 +1073,7 @@ async_iter_initialize(struct virtio_net *dev, struct vhost_async *async)
- return 0;
- }
- -static __rte_always_inline int
- +static int
- async_iter_add_iovec(struct virtio_net *dev, struct vhost_async *async,
- void *src, void *dst, size_t len)
- {
- @@ -1104,13 +1104,13 @@ async_iter_add_iovec(struct virtio_net *dev, struct vhost_async *async,
- return 0;
- }
- -static __rte_always_inline void
- +static void
- async_iter_finalize(struct vhost_async *async)
- {
- async->iter_idx++;
- }
- -static __rte_always_inline void
- +static void
- async_iter_cancel(struct vhost_async *async)
- {
- struct vhost_iov_iter *iter;
- @@ -1121,14 +1121,14 @@ async_iter_cancel(struct vhost_async *async)
- iter->iov = NULL;
- }
- -static __rte_always_inline void
- +static void
- async_iter_reset(struct vhost_async *async)
- {
- async->iter_idx = 0;
- async->iovec_idx = 0;
- }
- -static __rte_always_inline int
- +static int
- async_fill_seg(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf *m, uint32_t mbuf_offset,
- uint64_t buf_iova, uint32_t cpy_len, bool to_desc)
- @@ -1170,7 +1170,7 @@ async_fill_seg(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return 0;
- }
- -static __rte_always_inline void
- +static void
- sync_fill_seg(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf *m, uint32_t mbuf_offset,
- uint64_t buf_addr, uint64_t buf_iova, uint32_t cpy_len, bool to_desc)
- @@ -1208,7 +1208,7 @@ sync_fill_seg(struct virtio_net *dev, struct vhost_virtqueue *vq,
- }
- }
- -static __rte_always_inline int
- +static int
- mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf *m, struct buf_vector *buf_vec,
- uint16_t nr_vec, uint16_t num_buffers, bool is_async)
- @@ -1336,7 +1336,7 @@ mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return -1;
- }
- -static __rte_always_inline int
- +static int
- vhost_enqueue_single_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf *pkt,
- @@ -1401,7 +1401,7 @@ vhost_enqueue_single_packed(struct virtio_net *dev,
- return 0;
- }
- -static __rte_noinline uint32_t
- +static uint32_t
- virtio_dev_rx_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts, uint32_t count)
- __rte_shared_locks_required(&vq->access_lock)
- @@ -1457,7 +1457,7 @@ virtio_dev_rx_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return pkt_idx;
- }
- -static __rte_always_inline int
- +static int
- virtio_dev_rx_sync_batch_check(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts,
- @@ -1509,7 +1509,7 @@ virtio_dev_rx_sync_batch_check(struct virtio_net *dev,
- return 0;
- }
- -static __rte_always_inline int
- +static int
- virtio_dev_rx_async_batch_check(struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts,
- uint64_t *desc_addrs,
- @@ -1561,7 +1561,7 @@ virtio_dev_rx_async_batch_check(struct vhost_virtqueue *vq,
- return 0;
- }
- -static __rte_always_inline void
- +static void
- virtio_dev_rx_batch_packed_copy(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts,
- @@ -1611,7 +1611,7 @@ virtio_dev_rx_batch_packed_copy(struct virtio_net *dev,
- vhost_flush_enqueue_batch_packed(dev, vq, lens, ids);
- }
- -static __rte_always_inline int
- +static int
- virtio_dev_rx_sync_batch_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts)
- @@ -1633,7 +1633,7 @@ virtio_dev_rx_sync_batch_packed(struct virtio_net *dev,
- return 0;
- }
- -static __rte_always_inline int16_t
- +static int16_t
- virtio_dev_rx_single_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf *pkt)
- @@ -1658,7 +1658,7 @@ virtio_dev_rx_single_packed(struct virtio_net *dev,
- return 0;
- }
- -static __rte_noinline uint32_t
- +static uint32_t
- virtio_dev_rx_packed(struct virtio_net *dev,
- struct vhost_virtqueue *__rte_restrict vq,
- struct rte_mbuf **__rte_restrict pkts,
- @@ -1707,7 +1707,7 @@ virtio_dev_vring_translate(struct virtio_net *dev, struct vhost_virtqueue *vq)
- rte_rwlock_write_unlock(&vq->access_lock);
- }
- -static __rte_always_inline uint32_t
- +static uint32_t
- virtio_dev_rx(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts, uint32_t count)
- {
- @@ -1776,7 +1776,7 @@ rte_vhost_enqueue_burst(int vid, uint16_t queue_id,
- return virtio_dev_rx(dev, dev->virtqueue[queue_id], pkts, count);
- }
- -static __rte_always_inline uint16_t
- +static uint16_t
- async_get_first_inflight_pkt_idx(struct vhost_virtqueue *vq)
- __rte_shared_locks_required(&vq->access_lock)
- {
- @@ -1788,7 +1788,7 @@ async_get_first_inflight_pkt_idx(struct vhost_virtqueue *vq)
- return vq->size - async->pkts_inflight_n + async->pkts_idx;
- }
- -static __rte_always_inline void
- +static void
- store_dma_desc_info_split(struct vring_used_elem *s_ring, struct vring_used_elem *d_ring,
- uint16_t ring_size, uint16_t s_idx, uint16_t d_idx, uint16_t count)
- {
- @@ -1804,7 +1804,7 @@ store_dma_desc_info_split(struct vring_used_elem *s_ring, struct vring_used_elem
- }
- }
- -static __rte_noinline uint32_t
- +static uint32_t
- virtio_dev_rx_async_submit_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts, uint32_t count, int16_t dma_id, uint16_t vchan_id)
- __rte_exclusive_locks_required(&vq->access_lock)
- @@ -1909,7 +1909,7 @@ virtio_dev_rx_async_submit_split(struct virtio_net *dev, struct vhost_virtqueue
- }
- -static __rte_always_inline int
- +static int
- vhost_enqueue_async_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf *pkt,
- @@ -1973,7 +1973,7 @@ vhost_enqueue_async_packed(struct virtio_net *dev,
- return 0;
- }
- -static __rte_always_inline int16_t
- +static int16_t
- virtio_dev_rx_async_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf *pkt, uint16_t *nr_descs, uint16_t *nr_buffers)
- __rte_exclusive_locks_required(&vq->access_lock)
- @@ -1994,7 +1994,7 @@ virtio_dev_rx_async_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return 0;
- }
- -static __rte_always_inline void
- +static void
- virtio_dev_rx_async_packed_batch_enqueue(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts,
- @@ -2057,7 +2057,7 @@ virtio_dev_rx_async_packed_batch_enqueue(struct virtio_net *dev,
- vhost_async_shadow_enqueue_packed_batch(vq, lens, ids);
- }
- -static __rte_always_inline int
- +static int
- virtio_dev_rx_async_packed_batch(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts,
- @@ -2076,7 +2076,7 @@ virtio_dev_rx_async_packed_batch(struct virtio_net *dev,
- return 0;
- }
- -static __rte_always_inline void
- +static void
- dma_error_handler_packed(struct vhost_virtqueue *vq, uint16_t slot_idx,
- uint32_t nr_err, uint32_t *pkt_idx)
- __rte_exclusive_locks_required(&vq->access_lock)
- @@ -2107,7 +2107,7 @@ dma_error_handler_packed(struct vhost_virtqueue *vq, uint16_t slot_idx,
- async->buffer_idx_packed = async->buffer_idx_packed + vq->size - buffers_err;
- }
- -static __rte_noinline uint32_t
- +static uint32_t
- virtio_dev_rx_async_submit_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts, uint32_t count, int16_t dma_id, uint16_t vchan_id)
- __rte_exclusive_locks_required(&vq->access_lock)
- @@ -2182,7 +2182,7 @@ virtio_dev_rx_async_submit_packed(struct virtio_net *dev, struct vhost_virtqueue
- return pkt_idx;
- }
- -static __rte_always_inline void
- +static void
- write_back_completed_descs_split(struct vhost_virtqueue *vq, uint16_t n_descs)
- __rte_shared_locks_required(&vq->access_lock)
- {
- @@ -2214,7 +2214,7 @@ write_back_completed_descs_split(struct vhost_virtqueue *vq, uint16_t n_descs)
- } while (nr_left > 0);
- }
- -static __rte_always_inline void
- +static void
- write_back_completed_descs_packed(struct vhost_virtqueue *vq,
- uint16_t n_buffers)
- __rte_shared_locks_required(&vq->access_lock)
- @@ -2279,7 +2279,7 @@ write_back_completed_descs_packed(struct vhost_virtqueue *vq,
- async->last_buffer_idx_packed = from;
- }
- -static __rte_always_inline uint16_t
- +static uint16_t
- vhost_poll_enqueue_completed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts, uint16_t count, int16_t dma_id, uint16_t vchan_id)
- __rte_shared_locks_required(&vq->access_lock)
- @@ -2521,7 +2521,7 @@ rte_vhost_clear_queue(int vid, uint16_t queue_id, struct rte_mbuf **pkts,
- return n_pkts_cpl;
- }
- -static __rte_always_inline uint32_t
- +static uint32_t
- virtio_dev_rx_async_submit(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts, uint32_t count, int16_t dma_id, uint16_t vchan_id)
- {
- @@ -2603,7 +2603,7 @@ rte_vhost_submit_enqueue_burst(int vid, uint16_t queue_id,
- dma_id, vchan_id);
- }
- -static inline bool
- +static bool
- virtio_net_with_host_offload(struct virtio_net *dev)
- {
- if (dev->features &
- @@ -2702,7 +2702,7 @@ parse_headers(struct rte_mbuf *m, uint8_t *l4_proto)
- return -EINVAL;
- }
- -static __rte_always_inline void
- +static void
- vhost_dequeue_offload_legacy(struct virtio_net *dev, struct virtio_net_hdr *hdr,
- struct rte_mbuf *m)
- {
- @@ -2778,7 +2778,7 @@ vhost_dequeue_offload_legacy(struct virtio_net *dev, struct virtio_net_hdr *hdr,
- m->ol_flags = 0;
- }
- -static __rte_always_inline void
- +static void
- vhost_dequeue_offload(struct virtio_net *dev, struct virtio_net_hdr *hdr,
- struct rte_mbuf *m, bool legacy_ol_flags)
- {
- @@ -2866,7 +2866,7 @@ vhost_dequeue_offload(struct virtio_net *dev, struct virtio_net_hdr *hdr,
- }
- }
- -static __rte_noinline void
- +static void
- copy_vnet_hdr_from_desc(struct virtio_net_hdr *hdr,
- struct buf_vector *buf_vec)
- {
- @@ -2887,7 +2887,7 @@ copy_vnet_hdr_from_desc(struct virtio_net_hdr *hdr,
- }
- }
- -static __rte_always_inline int
- +static int
- desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct buf_vector *buf_vec, uint16_t nr_vec,
- struct rte_mbuf *m, struct rte_mempool *mbuf_pool,
- @@ -3077,7 +3077,7 @@ virtio_dev_extbuf_alloc(struct virtio_net *dev, struct rte_mbuf *pkt, uint32_t s
- /*
- * Prepare a host supported pktmbuf.
- */
- -static __rte_always_inline int
- +static int
- virtio_dev_pktmbuf_prep(struct virtio_net *dev, struct rte_mbuf *pkt,
- uint32_t data_len)
- {
- @@ -3095,7 +3095,7 @@ virtio_dev_pktmbuf_prep(struct virtio_net *dev, struct rte_mbuf *pkt,
- return -1;
- }
- -__rte_always_inline
- +
- static uint16_t
- virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mempool *mbuf_pool, struct rte_mbuf **pkts, uint16_t count,
- @@ -3190,7 +3190,7 @@ virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return i;
- }
- -__rte_noinline
- +
- static uint16_t
- virtio_dev_tx_split_legacy(struct virtio_net *dev,
- struct vhost_virtqueue *vq, struct rte_mempool *mbuf_pool,
- @@ -3201,7 +3201,7 @@ virtio_dev_tx_split_legacy(struct virtio_net *dev,
- return virtio_dev_tx_split(dev, vq, mbuf_pool, pkts, count, true);
- }
- -__rte_noinline
- +
- static uint16_t
- virtio_dev_tx_split_compliant(struct virtio_net *dev,
- struct vhost_virtqueue *vq, struct rte_mempool *mbuf_pool,
- @@ -3212,7 +3212,7 @@ virtio_dev_tx_split_compliant(struct virtio_net *dev,
- return virtio_dev_tx_split(dev, vq, mbuf_pool, pkts, count, false);
- }
- -static __rte_always_inline int
- +static int
- vhost_reserve_avail_batch_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts,
- @@ -3284,7 +3284,7 @@ vhost_reserve_avail_batch_packed(struct virtio_net *dev,
- return -1;
- }
- -static __rte_always_inline int
- +static int
- vhost_async_tx_batch_packed_check(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts,
- @@ -3358,7 +3358,7 @@ vhost_async_tx_batch_packed_check(struct virtio_net *dev,
- return -1;
- }
- -static __rte_always_inline int
- +static int
- virtio_dev_tx_batch_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts,
- @@ -3402,7 +3402,7 @@ virtio_dev_tx_batch_packed(struct virtio_net *dev,
- return 0;
- }
- -static __rte_always_inline int
- +static int
- vhost_dequeue_single_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mempool *mbuf_pool,
- @@ -3454,7 +3454,7 @@ vhost_dequeue_single_packed(struct virtio_net *dev,
- return 0;
- }
- -static __rte_always_inline int
- +static int
- virtio_dev_tx_single_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mempool *mbuf_pool,
- @@ -3484,7 +3484,7 @@ virtio_dev_tx_single_packed(struct virtio_net *dev,
- return ret;
- }
- -__rte_always_inline
- +
- static uint16_t
- virtio_dev_tx_packed(struct virtio_net *dev,
- struct vhost_virtqueue *__rte_restrict vq,
- @@ -3534,7 +3534,7 @@ virtio_dev_tx_packed(struct virtio_net *dev,
- return pkt_idx;
- }
- -__rte_noinline
- +
- static uint16_t
- virtio_dev_tx_packed_legacy(struct virtio_net *dev,
- struct vhost_virtqueue *__rte_restrict vq, struct rte_mempool *mbuf_pool,
- @@ -3545,7 +3545,7 @@ virtio_dev_tx_packed_legacy(struct virtio_net *dev,
- return virtio_dev_tx_packed(dev, vq, mbuf_pool, pkts, count, true);
- }
- -__rte_noinline
- +
- static uint16_t
- virtio_dev_tx_packed_compliant(struct virtio_net *dev,
- struct vhost_virtqueue *__rte_restrict vq, struct rte_mempool *mbuf_pool,
- @@ -3666,7 +3666,7 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,
- return count;
- }
- -static __rte_always_inline uint16_t
- +static uint16_t
- async_poll_dequeue_completed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts, uint16_t count, int16_t dma_id,
- uint16_t vchan_id, bool legacy_ol_flags)
- @@ -3714,7 +3714,7 @@ async_poll_dequeue_completed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return nr_cpl_pkts;
- }
- -static __rte_always_inline uint16_t
- +static uint16_t
- virtio_dev_tx_async_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mempool *mbuf_pool, struct rte_mbuf **pkts, uint16_t count,
- int16_t dma_id, uint16_t vchan_id, bool legacy_ol_flags)
- @@ -3863,7 +3863,7 @@ virtio_dev_tx_async_split(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return nr_done_pkts;
- }
- -__rte_noinline
- +
- static uint16_t
- virtio_dev_tx_async_split_legacy(struct virtio_net *dev,
- struct vhost_virtqueue *vq, struct rte_mempool *mbuf_pool,
- @@ -3876,7 +3876,7 @@ virtio_dev_tx_async_split_legacy(struct virtio_net *dev,
- pkts, count, dma_id, vchan_id, true);
- }
- -__rte_noinline
- +
- static uint16_t
- virtio_dev_tx_async_split_compliant(struct virtio_net *dev,
- struct vhost_virtqueue *vq, struct rte_mempool *mbuf_pool,
- @@ -3889,7 +3889,7 @@ virtio_dev_tx_async_split_compliant(struct virtio_net *dev,
- pkts, count, dma_id, vchan_id, false);
- }
- -static __rte_always_inline void
- +static void
- vhost_async_shadow_dequeue_single_packed(struct vhost_virtqueue *vq,
- uint16_t buf_id, uint16_t count)
- __rte_shared_locks_required(&vq->access_lock)
- @@ -3907,7 +3907,7 @@ vhost_async_shadow_dequeue_single_packed(struct vhost_virtqueue *vq,
- }
- -static __rte_always_inline int
- +static int
- virtio_dev_tx_async_single_packed(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mempool *mbuf_pool,
- @@ -3962,7 +3962,7 @@ virtio_dev_tx_async_single_packed(struct virtio_net *dev,
- return err;
- }
- -static __rte_always_inline int
- +static int
- virtio_dev_tx_async_packed_batch(struct virtio_net *dev,
- struct vhost_virtqueue *vq,
- struct rte_mbuf **pkts, uint16_t slot_idx,
- @@ -4021,7 +4021,7 @@ virtio_dev_tx_async_packed_batch(struct virtio_net *dev,
- return 0;
- }
- -static __rte_always_inline uint16_t
- +static uint16_t
- virtio_dev_tx_async_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mempool *mbuf_pool, struct rte_mbuf **pkts,
- uint16_t count, uint16_t dma_id, uint16_t vchan_id, bool legacy_ol_flags)
- @@ -4133,7 +4133,7 @@ virtio_dev_tx_async_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,
- return nr_done_pkts;
- }
- -__rte_noinline
- +
- static uint16_t
- virtio_dev_tx_async_packed_legacy(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mempool *mbuf_pool, struct rte_mbuf **pkts,
- @@ -4145,7 +4145,7 @@ virtio_dev_tx_async_packed_legacy(struct virtio_net *dev, struct vhost_virtqueue
- pkts, count, dma_id, vchan_id, true);
- }
- -__rte_noinline
- +
- static uint16_t
- virtio_dev_tx_async_packed_compliant(struct virtio_net *dev, struct vhost_virtqueue *vq,
- struct rte_mempool *mbuf_pool, struct rte_mbuf **pkts,
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement