kni: move functions to eliminate declarations
Function implementations kept same. Signed-off-by: Ferruh Yigit <ferruh.yigit@intel.com>
This commit is contained in:
parent
d2d7d6fc5f
commit
dec1ffcae7
@ -51,35 +51,6 @@ MODULE_DESCRIPTION("Kernel Module for managing kni devices");
|
||||
extern const struct pci_device_id ixgbe_pci_tbl[];
|
||||
extern const struct pci_device_id igb_pci_tbl[];
|
||||
|
||||
static int kni_open(struct inode *inode, struct file *file);
|
||||
static int kni_release(struct inode *inode, struct file *file);
|
||||
static int kni_ioctl(struct inode *inode, unsigned int ioctl_num,
|
||||
unsigned long ioctl_param);
|
||||
static int kni_compat_ioctl(struct inode *inode, unsigned int ioctl_num,
|
||||
unsigned long ioctl_param);
|
||||
static int kni_dev_remove(struct kni_dev *dev);
|
||||
|
||||
static int __init kni_parse_kthread_mode(void);
|
||||
|
||||
/* KNI processing for single kernel thread mode */
|
||||
static int kni_thread_single(void *unused);
|
||||
/* KNI processing for multiple kernel thread mode */
|
||||
static int kni_thread_multiple(void *param);
|
||||
|
||||
static const struct file_operations kni_fops = {
|
||||
.owner = THIS_MODULE,
|
||||
.open = kni_open,
|
||||
.release = kni_release,
|
||||
.unlocked_ioctl = (void *)kni_ioctl,
|
||||
.compat_ioctl = (void *)kni_compat_ioctl,
|
||||
};
|
||||
|
||||
static struct miscdevice kni_misc = {
|
||||
.minor = MISC_DYNAMIC_MINOR,
|
||||
.name = KNI_DEVICE,
|
||||
.fops = &kni_fops,
|
||||
};
|
||||
|
||||
/* loopback mode */
|
||||
static char *lo_mode;
|
||||
|
||||
@ -156,140 +127,6 @@ static struct pernet_operations kni_net_ops = {
|
||||
#endif
|
||||
};
|
||||
|
||||
static int __init
|
||||
kni_init(void)
|
||||
{
|
||||
int rc;
|
||||
|
||||
pr_debug("######## DPDK kni module loading ########\n");
|
||||
|
||||
if (kni_parse_kthread_mode() < 0) {
|
||||
pr_err("Invalid parameter for kthread_mode\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (multiple_kthread_on == 0)
|
||||
pr_debug("Single kernel thread for all KNI devices\n");
|
||||
else
|
||||
pr_debug("Multiple kernel thread mode enabled\n");
|
||||
|
||||
#ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
|
||||
rc = register_pernet_subsys(&kni_net_ops);
|
||||
#else
|
||||
rc = register_pernet_gen_subsys(&kni_net_id, &kni_net_ops);
|
||||
#endif
|
||||
if (rc)
|
||||
return -EPERM;
|
||||
|
||||
rc = misc_register(&kni_misc);
|
||||
if (rc != 0) {
|
||||
pr_err("Misc registration failed\n");
|
||||
goto out;
|
||||
}
|
||||
|
||||
/* Configure the lo mode according to the input parameter */
|
||||
kni_net_config_lo_mode(lo_mode);
|
||||
|
||||
pr_debug("######## DPDK kni module loaded ########\n");
|
||||
|
||||
return 0;
|
||||
|
||||
out:
|
||||
#ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
|
||||
unregister_pernet_subsys(&kni_net_ops);
|
||||
#else
|
||||
unregister_pernet_gen_subsys(kni_net_id, &kni_net_ops);
|
||||
#endif
|
||||
return rc;
|
||||
}
|
||||
|
||||
static void __exit
|
||||
kni_exit(void)
|
||||
{
|
||||
misc_deregister(&kni_misc);
|
||||
#ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
|
||||
unregister_pernet_subsys(&kni_net_ops);
|
||||
#else
|
||||
unregister_pernet_gen_subsys(kni_net_id, &kni_net_ops);
|
||||
#endif
|
||||
pr_debug("####### DPDK kni module unloaded #######\n");
|
||||
}
|
||||
|
||||
static int __init
|
||||
kni_parse_kthread_mode(void)
|
||||
{
|
||||
if (!kthread_mode)
|
||||
return 0;
|
||||
|
||||
if (strcmp(kthread_mode, "single") == 0)
|
||||
return 0;
|
||||
else if (strcmp(kthread_mode, "multiple") == 0)
|
||||
multiple_kthread_on = 1;
|
||||
else
|
||||
return -1;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
kni_open(struct inode *inode, struct file *file)
|
||||
{
|
||||
struct net *net = current->nsproxy->net_ns;
|
||||
struct kni_net *knet = net_generic(net, kni_net_id);
|
||||
|
||||
/* kni device can be opened by one user only per netns */
|
||||
if (test_and_set_bit(KNI_DEV_IN_USE_BIT_NUM, &knet->device_in_use))
|
||||
return -EBUSY;
|
||||
|
||||
file->private_data = get_net(net);
|
||||
pr_debug("/dev/kni opened\n");
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
kni_release(struct inode *inode, struct file *file)
|
||||
{
|
||||
struct net *net = file->private_data;
|
||||
struct kni_net *knet = net_generic(net, kni_net_id);
|
||||
struct kni_dev *dev, *n;
|
||||
|
||||
/* Stop kernel thread for single mode */
|
||||
if (multiple_kthread_on == 0) {
|
||||
mutex_lock(&knet->kni_kthread_lock);
|
||||
/* Stop kernel thread */
|
||||
if (knet->kni_kthread != NULL) {
|
||||
kthread_stop(knet->kni_kthread);
|
||||
knet->kni_kthread = NULL;
|
||||
}
|
||||
mutex_unlock(&knet->kni_kthread_lock);
|
||||
}
|
||||
|
||||
down_write(&knet->kni_list_lock);
|
||||
list_for_each_entry_safe(dev, n, &knet->kni_list_head, list) {
|
||||
/* Stop kernel thread for multiple mode */
|
||||
if (multiple_kthread_on && dev->pthread != NULL) {
|
||||
kthread_stop(dev->pthread);
|
||||
dev->pthread = NULL;
|
||||
}
|
||||
|
||||
#ifdef RTE_KNI_VHOST
|
||||
kni_vhost_backend_release(dev);
|
||||
#endif
|
||||
kni_dev_remove(dev);
|
||||
list_del(&dev->list);
|
||||
}
|
||||
up_write(&knet->kni_list_lock);
|
||||
|
||||
/* Clear the bit of device in use */
|
||||
clear_bit(KNI_DEV_IN_USE_BIT_NUM, &knet->device_in_use);
|
||||
|
||||
put_net(net);
|
||||
pr_debug("/dev/kni closed\n");
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
kni_thread_single(void *data)
|
||||
{
|
||||
@ -344,6 +181,22 @@ kni_thread_multiple(void *param)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
kni_open(struct inode *inode, struct file *file)
|
||||
{
|
||||
struct net *net = current->nsproxy->net_ns;
|
||||
struct kni_net *knet = net_generic(net, kni_net_id);
|
||||
|
||||
/* kni device can be opened by one user only per netns */
|
||||
if (test_and_set_bit(KNI_DEV_IN_USE_BIT_NUM, &knet->device_in_use))
|
||||
return -EBUSY;
|
||||
|
||||
file->private_data = get_net(net);
|
||||
pr_debug("/dev/kni opened\n");
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
kni_dev_remove(struct kni_dev *dev)
|
||||
{
|
||||
@ -365,6 +218,49 @@ kni_dev_remove(struct kni_dev *dev)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
kni_release(struct inode *inode, struct file *file)
|
||||
{
|
||||
struct net *net = file->private_data;
|
||||
struct kni_net *knet = net_generic(net, kni_net_id);
|
||||
struct kni_dev *dev, *n;
|
||||
|
||||
/* Stop kernel thread for single mode */
|
||||
if (multiple_kthread_on == 0) {
|
||||
mutex_lock(&knet->kni_kthread_lock);
|
||||
/* Stop kernel thread */
|
||||
if (knet->kni_kthread != NULL) {
|
||||
kthread_stop(knet->kni_kthread);
|
||||
knet->kni_kthread = NULL;
|
||||
}
|
||||
mutex_unlock(&knet->kni_kthread_lock);
|
||||
}
|
||||
|
||||
down_write(&knet->kni_list_lock);
|
||||
list_for_each_entry_safe(dev, n, &knet->kni_list_head, list) {
|
||||
/* Stop kernel thread for multiple mode */
|
||||
if (multiple_kthread_on && dev->pthread != NULL) {
|
||||
kthread_stop(dev->pthread);
|
||||
dev->pthread = NULL;
|
||||
}
|
||||
|
||||
#ifdef RTE_KNI_VHOST
|
||||
kni_vhost_backend_release(dev);
|
||||
#endif
|
||||
kni_dev_remove(dev);
|
||||
list_del(&dev->list);
|
||||
}
|
||||
up_write(&knet->kni_list_lock);
|
||||
|
||||
/* Clear the bit of device in use */
|
||||
clear_bit(KNI_DEV_IN_USE_BIT_NUM, &knet->device_in_use);
|
||||
|
||||
put_net(net);
|
||||
pr_debug("/dev/kni closed\n");
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
kni_check_param(struct kni_dev *kni, struct rte_kni_device_info *dev)
|
||||
{
|
||||
@ -685,6 +581,95 @@ kni_compat_ioctl(struct inode *inode,
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
static const struct file_operations kni_fops = {
|
||||
.owner = THIS_MODULE,
|
||||
.open = kni_open,
|
||||
.release = kni_release,
|
||||
.unlocked_ioctl = (void *)kni_ioctl,
|
||||
.compat_ioctl = (void *)kni_compat_ioctl,
|
||||
};
|
||||
|
||||
static struct miscdevice kni_misc = {
|
||||
.minor = MISC_DYNAMIC_MINOR,
|
||||
.name = KNI_DEVICE,
|
||||
.fops = &kni_fops,
|
||||
};
|
||||
|
||||
static int __init
|
||||
kni_parse_kthread_mode(void)
|
||||
{
|
||||
if (!kthread_mode)
|
||||
return 0;
|
||||
|
||||
if (strcmp(kthread_mode, "single") == 0)
|
||||
return 0;
|
||||
else if (strcmp(kthread_mode, "multiple") == 0)
|
||||
multiple_kthread_on = 1;
|
||||
else
|
||||
return -1;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int __init
|
||||
kni_init(void)
|
||||
{
|
||||
int rc;
|
||||
|
||||
pr_debug("######## DPDK kni module loading ########\n");
|
||||
|
||||
if (kni_parse_kthread_mode() < 0) {
|
||||
pr_err("Invalid parameter for kthread_mode\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (multiple_kthread_on == 0)
|
||||
pr_debug("Single kernel thread for all KNI devices\n");
|
||||
else
|
||||
pr_debug("Multiple kernel thread mode enabled\n");
|
||||
|
||||
#ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
|
||||
rc = register_pernet_subsys(&kni_net_ops);
|
||||
#else
|
||||
rc = register_pernet_gen_subsys(&kni_net_id, &kni_net_ops);
|
||||
#endif
|
||||
if (rc)
|
||||
return -EPERM;
|
||||
|
||||
rc = misc_register(&kni_misc);
|
||||
if (rc != 0) {
|
||||
pr_err("Misc registration failed\n");
|
||||
goto out;
|
||||
}
|
||||
|
||||
/* Configure the lo mode according to the input parameter */
|
||||
kni_net_config_lo_mode(lo_mode);
|
||||
|
||||
pr_debug("######## DPDK kni module loaded ########\n");
|
||||
|
||||
return 0;
|
||||
|
||||
out:
|
||||
#ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
|
||||
unregister_pernet_subsys(&kni_net_ops);
|
||||
#else
|
||||
unregister_pernet_gen_subsys(kni_net_id, &kni_net_ops);
|
||||
#endif
|
||||
return rc;
|
||||
}
|
||||
|
||||
static void __exit
|
||||
kni_exit(void)
|
||||
{
|
||||
misc_deregister(&kni_misc);
|
||||
#ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
|
||||
unregister_pernet_subsys(&kni_net_ops);
|
||||
#else
|
||||
unregister_pernet_gen_subsys(kni_net_id, &kni_net_ops);
|
||||
#endif
|
||||
pr_debug("####### DPDK kni module unloaded #######\n");
|
||||
}
|
||||
|
||||
module_init(kni_init);
|
||||
module_exit(kni_exit);
|
||||
|
||||
|
@ -49,12 +49,7 @@
|
||||
/* typedef for rx function */
|
||||
typedef void (*kni_net_rx_t)(struct kni_dev *kni);
|
||||
|
||||
static int kni_net_tx(struct sk_buff *skb, struct net_device *dev);
|
||||
static void kni_net_rx_normal(struct kni_dev *kni);
|
||||
static void kni_net_rx_lo_fifo(struct kni_dev *kni);
|
||||
static void kni_net_rx_lo_fifo_skb(struct kni_dev *kni);
|
||||
static int kni_net_process_request(struct kni_dev *kni,
|
||||
struct rte_kni_request *req);
|
||||
|
||||
/* kni rx function pointer, with default to normal rx */
|
||||
static kni_net_rx_t kni_net_rx_func = kni_net_rx_normal;
|
||||
@ -97,6 +92,55 @@ va2pa(void *va, struct rte_kni_mbuf *m)
|
||||
return pa;
|
||||
}
|
||||
|
||||
/*
|
||||
* It can be called to process the request.
|
||||
*/
|
||||
static int
|
||||
kni_net_process_request(struct kni_dev *kni, struct rte_kni_request *req)
|
||||
{
|
||||
int ret = -1;
|
||||
void *resp_va;
|
||||
unsigned int num;
|
||||
int ret_val;
|
||||
|
||||
if (!kni || !req) {
|
||||
pr_err("No kni instance or request\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
mutex_lock(&kni->sync_lock);
|
||||
|
||||
/* Construct data */
|
||||
memcpy(kni->sync_kva, req, sizeof(struct rte_kni_request));
|
||||
num = kni_fifo_put(kni->req_q, &kni->sync_va, 1);
|
||||
if (num < 1) {
|
||||
pr_err("Cannot send to req_q\n");
|
||||
ret = -EBUSY;
|
||||
goto fail;
|
||||
}
|
||||
|
||||
ret_val = wait_event_interruptible_timeout(kni->wq,
|
||||
kni_fifo_count(kni->resp_q), 3 * HZ);
|
||||
if (signal_pending(current) || ret_val <= 0) {
|
||||
ret = -ETIME;
|
||||
goto fail;
|
||||
}
|
||||
num = kni_fifo_get(kni->resp_q, (void **)&resp_va, 1);
|
||||
if (num != 1 || resp_va != kni->sync_va) {
|
||||
/* This should never happen */
|
||||
pr_err("No data in resp_q\n");
|
||||
ret = -ENODATA;
|
||||
goto fail;
|
||||
}
|
||||
|
||||
memcpy(req, kni->sync_kva, sizeof(struct rte_kni_request));
|
||||
ret = 0;
|
||||
|
||||
fail:
|
||||
mutex_unlock(&kni->sync_lock);
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* Open and close
|
||||
*/
|
||||
@ -151,6 +195,102 @@ kni_net_config(struct net_device *dev, struct ifmap *map)
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Transmit a packet (called by the kernel)
|
||||
*/
|
||||
#ifdef RTE_KNI_VHOST
|
||||
static int
|
||||
kni_net_tx(struct sk_buff *skb, struct net_device *dev)
|
||||
{
|
||||
struct kni_dev *kni = netdev_priv(dev);
|
||||
|
||||
dev_kfree_skb(skb);
|
||||
kni->stats.tx_dropped++;
|
||||
|
||||
return NETDEV_TX_OK;
|
||||
}
|
||||
#else
|
||||
static int
|
||||
kni_net_tx(struct sk_buff *skb, struct net_device *dev)
|
||||
{
|
||||
int len = 0;
|
||||
unsigned int ret;
|
||||
struct kni_dev *kni = netdev_priv(dev);
|
||||
struct rte_kni_mbuf *pkt_kva = NULL;
|
||||
void *pkt_pa = NULL;
|
||||
void *pkt_va = NULL;
|
||||
|
||||
/* save the timestamp */
|
||||
#ifdef HAVE_TRANS_START_HELPER
|
||||
netif_trans_update(dev);
|
||||
#else
|
||||
dev->trans_start = jiffies;
|
||||
#endif
|
||||
|
||||
/* Check if the length of skb is less than mbuf size */
|
||||
if (skb->len > kni->mbuf_size)
|
||||
goto drop;
|
||||
|
||||
/**
|
||||
* Check if it has at least one free entry in tx_q and
|
||||
* one entry in alloc_q.
|
||||
*/
|
||||
if (kni_fifo_free_count(kni->tx_q) == 0 ||
|
||||
kni_fifo_count(kni->alloc_q) == 0) {
|
||||
/**
|
||||
* If no free entry in tx_q or no entry in alloc_q,
|
||||
* drops skb and goes out.
|
||||
*/
|
||||
goto drop;
|
||||
}
|
||||
|
||||
/* dequeue a mbuf from alloc_q */
|
||||
ret = kni_fifo_get(kni->alloc_q, &pkt_pa, 1);
|
||||
if (likely(ret == 1)) {
|
||||
void *data_kva;
|
||||
|
||||
pkt_kva = pa2kva(pkt_pa);
|
||||
data_kva = kva2data_kva(pkt_kva);
|
||||
pkt_va = pa2va(pkt_pa, pkt_kva);
|
||||
|
||||
len = skb->len;
|
||||
memcpy(data_kva, skb->data, len);
|
||||
if (unlikely(len < ETH_ZLEN)) {
|
||||
memset(data_kva + len, 0, ETH_ZLEN - len);
|
||||
len = ETH_ZLEN;
|
||||
}
|
||||
pkt_kva->pkt_len = len;
|
||||
pkt_kva->data_len = len;
|
||||
|
||||
/* enqueue mbuf into tx_q */
|
||||
ret = kni_fifo_put(kni->tx_q, &pkt_va, 1);
|
||||
if (unlikely(ret != 1)) {
|
||||
/* Failing should not happen */
|
||||
pr_err("Fail to enqueue mbuf into tx_q\n");
|
||||
goto drop;
|
||||
}
|
||||
} else {
|
||||
/* Failing should not happen */
|
||||
pr_err("Fail to dequeue mbuf from alloc_q\n");
|
||||
goto drop;
|
||||
}
|
||||
|
||||
/* Free skb and update statistics */
|
||||
dev_kfree_skb(skb);
|
||||
kni->stats.tx_bytes += len;
|
||||
kni->stats.tx_packets++;
|
||||
|
||||
return NETDEV_TX_OK;
|
||||
|
||||
drop:
|
||||
/* Free skb and update statistics */
|
||||
dev_kfree_skb(skb);
|
||||
kni->stats.tx_dropped++;
|
||||
|
||||
return NETDEV_TX_OK;
|
||||
}
|
||||
#endif
|
||||
|
||||
/*
|
||||
* RX: normal working mode
|
||||
*/
|
||||
@ -425,102 +565,6 @@ kni_net_rx(struct kni_dev *kni)
|
||||
(*kni_net_rx_func)(kni);
|
||||
}
|
||||
|
||||
/*
|
||||
* Transmit a packet (called by the kernel)
|
||||
*/
|
||||
#ifdef RTE_KNI_VHOST
|
||||
static int
|
||||
kni_net_tx(struct sk_buff *skb, struct net_device *dev)
|
||||
{
|
||||
struct kni_dev *kni = netdev_priv(dev);
|
||||
|
||||
dev_kfree_skb(skb);
|
||||
kni->stats.tx_dropped++;
|
||||
|
||||
return NETDEV_TX_OK;
|
||||
}
|
||||
#else
|
||||
static int
|
||||
kni_net_tx(struct sk_buff *skb, struct net_device *dev)
|
||||
{
|
||||
int len = 0;
|
||||
unsigned int ret;
|
||||
struct kni_dev *kni = netdev_priv(dev);
|
||||
struct rte_kni_mbuf *pkt_kva = NULL;
|
||||
void *pkt_pa = NULL;
|
||||
void *pkt_va = NULL;
|
||||
|
||||
/* save the timestamp */
|
||||
#ifdef HAVE_TRANS_START_HELPER
|
||||
netif_trans_update(dev);
|
||||
#else
|
||||
dev->trans_start = jiffies;
|
||||
#endif
|
||||
|
||||
/* Check if the length of skb is less than mbuf size */
|
||||
if (skb->len > kni->mbuf_size)
|
||||
goto drop;
|
||||
|
||||
/**
|
||||
* Check if it has at least one free entry in tx_q and
|
||||
* one entry in alloc_q.
|
||||
*/
|
||||
if (kni_fifo_free_count(kni->tx_q) == 0 ||
|
||||
kni_fifo_count(kni->alloc_q) == 0) {
|
||||
/**
|
||||
* If no free entry in tx_q or no entry in alloc_q,
|
||||
* drops skb and goes out.
|
||||
*/
|
||||
goto drop;
|
||||
}
|
||||
|
||||
/* dequeue a mbuf from alloc_q */
|
||||
ret = kni_fifo_get(kni->alloc_q, &pkt_pa, 1);
|
||||
if (likely(ret == 1)) {
|
||||
void *data_kva;
|
||||
|
||||
pkt_kva = pa2kva(pkt_pa);
|
||||
data_kva = kva2data_kva(pkt_kva);
|
||||
pkt_va = pa2va(pkt_pa, pkt_kva);
|
||||
|
||||
len = skb->len;
|
||||
memcpy(data_kva, skb->data, len);
|
||||
if (unlikely(len < ETH_ZLEN)) {
|
||||
memset(data_kva + len, 0, ETH_ZLEN - len);
|
||||
len = ETH_ZLEN;
|
||||
}
|
||||
pkt_kva->pkt_len = len;
|
||||
pkt_kva->data_len = len;
|
||||
|
||||
/* enqueue mbuf into tx_q */
|
||||
ret = kni_fifo_put(kni->tx_q, &pkt_va, 1);
|
||||
if (unlikely(ret != 1)) {
|
||||
/* Failing should not happen */
|
||||
pr_err("Fail to enqueue mbuf into tx_q\n");
|
||||
goto drop;
|
||||
}
|
||||
} else {
|
||||
/* Failing should not happen */
|
||||
pr_err("Fail to dequeue mbuf from alloc_q\n");
|
||||
goto drop;
|
||||
}
|
||||
|
||||
/* Free skb and update statistics */
|
||||
dev_kfree_skb(skb);
|
||||
kni->stats.tx_bytes += len;
|
||||
kni->stats.tx_packets++;
|
||||
|
||||
return NETDEV_TX_OK;
|
||||
|
||||
drop:
|
||||
/* Free skb and update statistics */
|
||||
dev_kfree_skb(skb);
|
||||
kni->stats.tx_dropped++;
|
||||
|
||||
return NETDEV_TX_OK;
|
||||
}
|
||||
#endif
|
||||
|
||||
/*
|
||||
* Deal with a transmit timeout.
|
||||
*/
|
||||
@ -582,55 +626,6 @@ kni_net_poll_resp(struct kni_dev *kni)
|
||||
wake_up_interruptible(&kni->wq);
|
||||
}
|
||||
|
||||
/*
|
||||
* It can be called to process the request.
|
||||
*/
|
||||
static int
|
||||
kni_net_process_request(struct kni_dev *kni, struct rte_kni_request *req)
|
||||
{
|
||||
int ret = -1;
|
||||
void *resp_va;
|
||||
unsigned int num;
|
||||
int ret_val;
|
||||
|
||||
if (!kni || !req) {
|
||||
pr_err("No kni instance or request\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
mutex_lock(&kni->sync_lock);
|
||||
|
||||
/* Construct data */
|
||||
memcpy(kni->sync_kva, req, sizeof(struct rte_kni_request));
|
||||
num = kni_fifo_put(kni->req_q, &kni->sync_va, 1);
|
||||
if (num < 1) {
|
||||
pr_err("Cannot send to req_q\n");
|
||||
ret = -EBUSY;
|
||||
goto fail;
|
||||
}
|
||||
|
||||
ret_val = wait_event_interruptible_timeout(kni->wq,
|
||||
kni_fifo_count(kni->resp_q), 3 * HZ);
|
||||
if (signal_pending(current) || ret_val <= 0) {
|
||||
ret = -ETIME;
|
||||
goto fail;
|
||||
}
|
||||
num = kni_fifo_get(kni->resp_q, (void **)&resp_va, 1);
|
||||
if (num != 1 || resp_va != kni->sync_va) {
|
||||
/* This should never happen */
|
||||
pr_err("No data in resp_q\n");
|
||||
ret = -ENODATA;
|
||||
goto fail;
|
||||
}
|
||||
|
||||
memcpy(req, kni->sync_kva, sizeof(struct rte_kni_request));
|
||||
ret = 0;
|
||||
|
||||
fail:
|
||||
mutex_unlock(&kni->sync_lock);
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* Return statistics to the caller
|
||||
*/
|
||||
|
Loading…
x
Reference in New Issue
Block a user