Quellcodebibliothek Statistik Leitseite products/sources/formale Sprachen/C/Linux/net/ethtool/   (Open Source Betriebssystem Version 6.17.9©)  Datei vom 24.10.2025 mit Größe 11 kB image not shown  

Quelle  rings.c   Sprache: C

 
// SPDX-License-Identifier: GPL-2.0-only

#include <net/netdev_queues.h>

#include "netlink.h"
#include "common.h"

struct rings_req_info {
 struct ethnl_req_info  base;
};

struct rings_reply_data {
 struct ethnl_reply_data  base;
 struct ethtool_ringparam ringparam;
 struct kernel_ethtool_ringparam kernel_ringparam;
 u32    supported_ring_params;
};

#define RINGS_REPDATA(__reply_base) \
 container_of(__reply_base, struct rings_reply_data, base)

const struct nla_policy ethnl_rings_get_policy[] = {
 [ETHTOOL_A_RINGS_HEADER]  =
  NLA_POLICY_NESTED(ethnl_header_policy),
};

static int rings_prepare_data(const struct ethnl_req_info *req_base,
         struct ethnl_reply_data *reply_base,
         const struct genl_info *info)
{
 struct rings_reply_data *data = RINGS_REPDATA(reply_base);
 struct net_device *dev = reply_base->dev;
 int ret;

 if (!dev->ethtool_ops->get_ringparam)
  return -EOPNOTSUPP;

 data->supported_ring_params = dev->ethtool_ops->supported_ring_params;
 ret = ethnl_ops_begin(dev);
 if (ret < 0)
  return ret;

 data->kernel_ringparam.tcp_data_split = dev->cfg->hds_config;
 data->kernel_ringparam.hds_thresh = dev->cfg->hds_thresh;

 dev->ethtool_ops->get_ringparam(dev, &data->ringparam,
     &data->kernel_ringparam, info->extack);
 ethnl_ops_complete(dev);

 return 0;
}

static int rings_reply_size(const struct ethnl_req_info *req_base,
       const struct ethnl_reply_data *reply_base)
{
 return nla_total_size(sizeof(u32)) + /* _RINGS_RX_MAX */
        nla_total_size(sizeof(u32)) + /* _RINGS_RX_MINI_MAX */
        nla_total_size(sizeof(u32)) + /* _RINGS_RX_JUMBO_MAX */
        nla_total_size(sizeof(u32)) + /* _RINGS_TX_MAX */
        nla_total_size(sizeof(u32)) + /* _RINGS_RX */
        nla_total_size(sizeof(u32)) + /* _RINGS_RX_MINI */
        nla_total_size(sizeof(u32)) + /* _RINGS_RX_JUMBO */
        nla_total_size(sizeof(u32)) + /* _RINGS_TX */
        nla_total_size(sizeof(u32)) + /* _RINGS_RX_BUF_LEN */
        nla_total_size(sizeof(u8))  + /* _RINGS_TCP_DATA_SPLIT */
        nla_total_size(sizeof(u32)  + /* _RINGS_CQE_SIZE */
        nla_total_size(sizeof(u8))  + /* _RINGS_TX_PUSH */
        nla_total_size(sizeof(u8))) + /* _RINGS_RX_PUSH */
        nla_total_size(sizeof(u32)) + /* _RINGS_TX_PUSH_BUF_LEN */
        nla_total_size(sizeof(u32)) + /* _RINGS_TX_PUSH_BUF_LEN_MAX */
        nla_total_size(sizeof(u32)) + /* _RINGS_HDS_THRESH */
        nla_total_size(sizeof(u32)); /* _RINGS_HDS_THRESH_MAX*/
}

static int rings_fill_reply(struct sk_buff *skb,
       const struct ethnl_req_info *req_base,
       const struct ethnl_reply_data *reply_base)
{
 const struct rings_reply_data *data = RINGS_REPDATA(reply_base);
 const struct kernel_ethtool_ringparam *kr = &data->kernel_ringparam;
 const struct ethtool_ringparam *ringparam = &data->ringparam;
 u32 supported_ring_params = data->supported_ring_params;

 WARN_ON(kr->tcp_data_split > ETHTOOL_TCP_DATA_SPLIT_ENABLED);

 if ((ringparam->rx_max_pending &&
      (nla_put_u32(skb, ETHTOOL_A_RINGS_RX_MAX,
     ringparam->rx_max_pending) ||
       nla_put_u32(skb, ETHTOOL_A_RINGS_RX,
     ringparam->rx_pending))) ||
     (ringparam->rx_mini_max_pending &&
      (nla_put_u32(skb, ETHTOOL_A_RINGS_RX_MINI_MAX,
     ringparam->rx_mini_max_pending) ||
       nla_put_u32(skb, ETHTOOL_A_RINGS_RX_MINI,
     ringparam->rx_mini_pending))) ||
     (ringparam->rx_jumbo_max_pending &&
      (nla_put_u32(skb, ETHTOOL_A_RINGS_RX_JUMBO_MAX,
     ringparam->rx_jumbo_max_pending) ||
       nla_put_u32(skb, ETHTOOL_A_RINGS_RX_JUMBO,
     ringparam->rx_jumbo_pending))) ||
     (ringparam->tx_max_pending &&
      (nla_put_u32(skb, ETHTOOL_A_RINGS_TX_MAX,
     ringparam->tx_max_pending) ||
       nla_put_u32(skb, ETHTOOL_A_RINGS_TX,
     ringparam->tx_pending)))  ||
     (kr->rx_buf_len &&
      (nla_put_u32(skb, ETHTOOL_A_RINGS_RX_BUF_LEN, kr->rx_buf_len))) ||
     (kr->tcp_data_split &&
      (nla_put_u8(skb, ETHTOOL_A_RINGS_TCP_DATA_SPLIT,
    kr->tcp_data_split))) ||
     (kr->cqe_size &&
      (nla_put_u32(skb, ETHTOOL_A_RINGS_CQE_SIZE, kr->cqe_size))) ||
     nla_put_u8(skb, ETHTOOL_A_RINGS_TX_PUSH, !!kr->tx_push) ||
     nla_put_u8(skb, ETHTOOL_A_RINGS_RX_PUSH, !!kr->rx_push) ||
     ((supported_ring_params & ETHTOOL_RING_USE_TX_PUSH_BUF_LEN) &&
      (nla_put_u32(skb, ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN_MAX,
     kr->tx_push_buf_max_len) ||
       nla_put_u32(skb, ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN,
     kr->tx_push_buf_len))) ||
     ((supported_ring_params & ETHTOOL_RING_USE_HDS_THRS) &&
      (nla_put_u32(skb, ETHTOOL_A_RINGS_HDS_THRESH,
     kr->hds_thresh) ||
       nla_put_u32(skb, ETHTOOL_A_RINGS_HDS_THRESH_MAX,
     kr->hds_thresh_max))))
  return -EMSGSIZE;

 return 0;
}

/* RINGS_SET */

const struct nla_policy ethnl_rings_set_policy[] = {
 [ETHTOOL_A_RINGS_HEADER]  =
  NLA_POLICY_NESTED(ethnl_header_policy),
 [ETHTOOL_A_RINGS_RX]   = { .type = NLA_U32 },
 [ETHTOOL_A_RINGS_RX_MINI]  = { .type = NLA_U32 },
 [ETHTOOL_A_RINGS_RX_JUMBO]  = { .type = NLA_U32 },
 [ETHTOOL_A_RINGS_TX]   = { .type = NLA_U32 },
 [ETHTOOL_A_RINGS_RX_BUF_LEN]            = NLA_POLICY_MIN(NLA_U32, 1),
 [ETHTOOL_A_RINGS_TCP_DATA_SPLIT] =
  NLA_POLICY_MAX(NLA_U8, ETHTOOL_TCP_DATA_SPLIT_ENABLED),
 [ETHTOOL_A_RINGS_CQE_SIZE]  = NLA_POLICY_MIN(NLA_U32, 1),
 [ETHTOOL_A_RINGS_TX_PUSH]  = NLA_POLICY_MAX(NLA_U8, 1),
 [ETHTOOL_A_RINGS_RX_PUSH]  = NLA_POLICY_MAX(NLA_U8, 1),
 [ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN] = { .type = NLA_U32 },
 [ETHTOOL_A_RINGS_HDS_THRESH]  = { .type = NLA_U32 },
};

static int
ethnl_set_rings_validate(struct ethnl_req_info *req_info,
    struct genl_info *info)
{
 const struct ethtool_ops *ops = req_info->dev->ethtool_ops;
 struct nlattr **tb = info->attrs;

 if (tb[ETHTOOL_A_RINGS_RX_BUF_LEN] &&
     !(ops->supported_ring_params & ETHTOOL_RING_USE_RX_BUF_LEN)) {
  NL_SET_ERR_MSG_ATTR(info->extack,
        tb[ETHTOOL_A_RINGS_RX_BUF_LEN],
        "setting rx buf len not supported");
  return -EOPNOTSUPP;
 }

 if (tb[ETHTOOL_A_RINGS_TCP_DATA_SPLIT] &&
     !(ops->supported_ring_params & ETHTOOL_RING_USE_TCP_DATA_SPLIT)) {
  NL_SET_ERR_MSG_ATTR(info->extack,
        tb[ETHTOOL_A_RINGS_TCP_DATA_SPLIT],
        "setting TCP data split is not supported");
  return -EOPNOTSUPP;
 }

 if (tb[ETHTOOL_A_RINGS_HDS_THRESH] &&
     !(ops->supported_ring_params & ETHTOOL_RING_USE_HDS_THRS)) {
  NL_SET_ERR_MSG_ATTR(info->extack,
        tb[ETHTOOL_A_RINGS_HDS_THRESH],
        "setting hds-thresh is not supported");
  return -EOPNOTSUPP;
 }

 if (tb[ETHTOOL_A_RINGS_CQE_SIZE] &&
     !(ops->supported_ring_params & ETHTOOL_RING_USE_CQE_SIZE)) {
  NL_SET_ERR_MSG_ATTR(info->extack,
        tb[ETHTOOL_A_RINGS_CQE_SIZE],
        "setting cqe size not supported");
  return -EOPNOTSUPP;
 }

 if (tb[ETHTOOL_A_RINGS_TX_PUSH] &&
     !(ops->supported_ring_params & ETHTOOL_RING_USE_TX_PUSH)) {
  NL_SET_ERR_MSG_ATTR(info->extack,
        tb[ETHTOOL_A_RINGS_TX_PUSH],
        "setting tx push not supported");
  return -EOPNOTSUPP;
 }

 if (tb[ETHTOOL_A_RINGS_RX_PUSH] &&
     !(ops->supported_ring_params & ETHTOOL_RING_USE_RX_PUSH)) {
  NL_SET_ERR_MSG_ATTR(info->extack,
        tb[ETHTOOL_A_RINGS_RX_PUSH],
        "setting rx push not supported");
  return -EOPNOTSUPP;
 }

 if (tb[ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN] &&
     !(ops->supported_ring_params & ETHTOOL_RING_USE_TX_PUSH_BUF_LEN)) {
  NL_SET_ERR_MSG_ATTR(info->extack,
        tb[ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN],
        "setting tx push buf len is not supported");
  return -EOPNOTSUPP;
 }

 return ops->get_ringparam && ops->set_ringparam ? 1 : -EOPNOTSUPP;
}

static int
ethnl_set_rings(struct ethnl_req_info *req_info, struct genl_info *info)
{
 struct kernel_ethtool_ringparam kernel_ringparam;
 struct net_device *dev = req_info->dev;
 struct ethtool_ringparam ringparam;
 struct nlattr **tb = info->attrs;
 const struct nlattr *err_attr;
 bool mod = false;
 int ret;

 ethtool_ringparam_get_cfg(dev, &ringparam, &kernel_ringparam,
      info->extack);

 ethnl_update_u32(&ringparam.rx_pending, tb[ETHTOOL_A_RINGS_RX], &mod);
 ethnl_update_u32(&ringparam.rx_mini_pending,
    tb[ETHTOOL_A_RINGS_RX_MINI], &mod);
 ethnl_update_u32(&ringparam.rx_jumbo_pending,
    tb[ETHTOOL_A_RINGS_RX_JUMBO], &mod);
 ethnl_update_u32(&ringparam.tx_pending, tb[ETHTOOL_A_RINGS_TX], &mod);
 ethnl_update_u32(&kernel_ringparam.rx_buf_len,
    tb[ETHTOOL_A_RINGS_RX_BUF_LEN], &mod);
 ethnl_update_u8(&kernel_ringparam.tcp_data_split,
   tb[ETHTOOL_A_RINGS_TCP_DATA_SPLIT], &mod);
 ethnl_update_u32(&kernel_ringparam.cqe_size,
    tb[ETHTOOL_A_RINGS_CQE_SIZE], &mod);
 ethnl_update_u8(&kernel_ringparam.tx_push,
   tb[ETHTOOL_A_RINGS_TX_PUSH], &mod);
 ethnl_update_u8(&kernel_ringparam.rx_push,
   tb[ETHTOOL_A_RINGS_RX_PUSH], &mod);
 ethnl_update_u32(&kernel_ringparam.tx_push_buf_len,
    tb[ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN], &mod);
 ethnl_update_u32(&kernel_ringparam.hds_thresh,
    tb[ETHTOOL_A_RINGS_HDS_THRESH], &mod);
 if (!mod)
  return 0;

 if (kernel_ringparam.tcp_data_split == ETHTOOL_TCP_DATA_SPLIT_ENABLED &&
     dev_xdp_sb_prog_count(dev)) {
  NL_SET_ERR_MSG_ATTR(info->extack,
        tb[ETHTOOL_A_RINGS_TCP_DATA_SPLIT],
        "tcp-data-split can not be enabled with single buffer XDP");
  return -EINVAL;
 }

 if (dev_get_min_mp_channel_count(dev)) {
  if (kernel_ringparam.tcp_data_split !=
      ETHTOOL_TCP_DATA_SPLIT_ENABLED) {
   NL_SET_ERR_MSG(info->extack,
           "can't disable tcp-data-split while device has memory provider enabled");
   return -EINVAL;
  } else if (kernel_ringparam.hds_thresh) {
   NL_SET_ERR_MSG(info->extack,
           "can't set non-zero hds_thresh while device is memory provider enabled");
   return -EINVAL;
  }
 }

 /* ensure new ring parameters are within limits */
 if (ringparam.rx_pending > ringparam.rx_max_pending)
  err_attr = tb[ETHTOOL_A_RINGS_RX];
 else if (ringparam.rx_mini_pending > ringparam.rx_mini_max_pending)
  err_attr = tb[ETHTOOL_A_RINGS_RX_MINI];
 else if (ringparam.rx_jumbo_pending > ringparam.rx_jumbo_max_pending)
  err_attr = tb[ETHTOOL_A_RINGS_RX_JUMBO];
 else if (ringparam.tx_pending > ringparam.tx_max_pending)
  err_attr = tb[ETHTOOL_A_RINGS_TX];
 else if (kernel_ringparam.hds_thresh > kernel_ringparam.hds_thresh_max)
  err_attr = tb[ETHTOOL_A_RINGS_HDS_THRESH];
 else
  err_attr = NULL;
 if (err_attr) {
  NL_SET_ERR_MSG_ATTR(info->extack, err_attr,
        "requested ring size exceeds maximum");
  return -EINVAL;
 }

 if (kernel_ringparam.tx_push_buf_len > kernel_ringparam.tx_push_buf_max_len) {
  NL_SET_ERR_MSG_ATTR_FMT(info->extack, tb[ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN],
     "Requested TX push buffer exceeds the maximum of %u",
     kernel_ringparam.tx_push_buf_max_len);

  return -EINVAL;
 }

 dev->cfg_pending->hds_config = kernel_ringparam.tcp_data_split;
 dev->cfg_pending->hds_thresh = kernel_ringparam.hds_thresh;

 ret = dev->ethtool_ops->set_ringparam(dev, &ringparam,
           &kernel_ringparam, info->extack);
 return ret < 0 ? ret : 1;
}

const struct ethnl_request_ops ethnl_rings_request_ops = {
 .request_cmd  = ETHTOOL_MSG_RINGS_GET,
 .reply_cmd  = ETHTOOL_MSG_RINGS_GET_REPLY,
 .hdr_attr  = ETHTOOL_A_RINGS_HEADER,
 .req_info_size  = sizeof(struct rings_req_info),
 .reply_data_size = sizeof(struct rings_reply_data),

 .prepare_data  = rings_prepare_data,
 .reply_size  = rings_reply_size,
 .fill_reply  = rings_fill_reply,

 .set_validate  = ethnl_set_rings_validate,
 .set   = ethnl_set_rings,
 .set_ntf_cmd  = ETHTOOL_MSG_RINGS_NTF,
};

Messung V0.5
C=95 H=96 G=95

¤ Dauer der Verarbeitung: 0.4 Sekunden  ¤

*© Formatika GbR, Deutschland






Wurzel

Suchen

Beweissystem der NASA

Beweissystem Isabelle

NIST Cobol Testsuite

Cephes Mathematical Library

Wiener Entwicklungsmethode

Haftungshinweis

Die Informationen auf dieser Webseite wurden nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit, noch Qualität der bereit gestellten Informationen zugesichert.

Bemerkung:

Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.