mirror of https://github.com/torvalds/linux.git
669 lines
20 KiB
C
669 lines
20 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/* Copyright (C) 2024 Intel Corporation */
|
|
|
|
#include "idpf.h"
|
|
#include "idpf_ptp.h"
|
|
#include "idpf_virtchnl.h"
|
|
|
|
/**
|
|
* idpf_ptp_get_caps - Send virtchnl get ptp capabilities message
|
|
* @adapter: Driver specific private structure
|
|
*
|
|
* Send virtchnl get PTP capabilities message.
|
|
*
|
|
* Return: 0 on success, -errno on failure.
|
|
*/
|
|
int idpf_ptp_get_caps(struct idpf_adapter *adapter)
|
|
{
|
|
struct virtchnl2_ptp_get_caps *recv_ptp_caps_msg __free(kfree) = NULL;
|
|
struct virtchnl2_ptp_get_caps send_ptp_caps_msg = {
|
|
.caps = cpu_to_le32(VIRTCHNL2_CAP_PTP_GET_DEVICE_CLK_TIME |
|
|
VIRTCHNL2_CAP_PTP_GET_DEVICE_CLK_TIME_MB |
|
|
VIRTCHNL2_CAP_PTP_GET_CROSS_TIME |
|
|
VIRTCHNL2_CAP_PTP_SET_DEVICE_CLK_TIME_MB |
|
|
VIRTCHNL2_CAP_PTP_ADJ_DEVICE_CLK_MB |
|
|
VIRTCHNL2_CAP_PTP_TX_TSTAMPS_MB)
|
|
};
|
|
struct idpf_vc_xn_params xn_params = {
|
|
.vc_op = VIRTCHNL2_OP_PTP_GET_CAPS,
|
|
.send_buf.iov_base = &send_ptp_caps_msg,
|
|
.send_buf.iov_len = sizeof(send_ptp_caps_msg),
|
|
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
|
|
};
|
|
struct virtchnl2_ptp_cross_time_reg_offsets cross_tstamp_offsets;
|
|
struct virtchnl2_ptp_clk_adj_reg_offsets clk_adj_offsets;
|
|
struct virtchnl2_ptp_clk_reg_offsets clock_offsets;
|
|
struct idpf_ptp_secondary_mbx *scnd_mbx;
|
|
struct idpf_ptp *ptp = adapter->ptp;
|
|
enum idpf_ptp_access access_type;
|
|
u32 temp_offset;
|
|
int reply_sz;
|
|
|
|
recv_ptp_caps_msg = kzalloc(sizeof(struct virtchnl2_ptp_get_caps),
|
|
GFP_KERNEL);
|
|
if (!recv_ptp_caps_msg)
|
|
return -ENOMEM;
|
|
|
|
xn_params.recv_buf.iov_base = recv_ptp_caps_msg;
|
|
xn_params.recv_buf.iov_len = sizeof(*recv_ptp_caps_msg);
|
|
|
|
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
|
|
if (reply_sz < 0)
|
|
return reply_sz;
|
|
else if (reply_sz != sizeof(*recv_ptp_caps_msg))
|
|
return -EIO;
|
|
|
|
ptp->caps = le32_to_cpu(recv_ptp_caps_msg->caps);
|
|
ptp->base_incval = le64_to_cpu(recv_ptp_caps_msg->base_incval);
|
|
ptp->max_adj = le32_to_cpu(recv_ptp_caps_msg->max_adj);
|
|
|
|
scnd_mbx = &ptp->secondary_mbx;
|
|
scnd_mbx->peer_mbx_q_id = le16_to_cpu(recv_ptp_caps_msg->peer_mbx_q_id);
|
|
|
|
/* if the ptp_mb_q_id holds invalid value (0xffff), the secondary
|
|
* mailbox is not supported.
|
|
*/
|
|
scnd_mbx->valid = scnd_mbx->peer_mbx_q_id != 0xffff;
|
|
if (scnd_mbx->valid)
|
|
scnd_mbx->peer_id = recv_ptp_caps_msg->peer_id;
|
|
|
|
/* Determine the access type for the PTP features */
|
|
idpf_ptp_get_features_access(adapter);
|
|
|
|
access_type = ptp->get_dev_clk_time_access;
|
|
if (access_type != IDPF_PTP_DIRECT)
|
|
goto cross_tstamp;
|
|
|
|
clock_offsets = recv_ptp_caps_msg->clk_offsets;
|
|
|
|
temp_offset = le32_to_cpu(clock_offsets.dev_clk_ns_l);
|
|
ptp->dev_clk_regs.dev_clk_ns_l = idpf_get_reg_addr(adapter,
|
|
temp_offset);
|
|
temp_offset = le32_to_cpu(clock_offsets.dev_clk_ns_h);
|
|
ptp->dev_clk_regs.dev_clk_ns_h = idpf_get_reg_addr(adapter,
|
|
temp_offset);
|
|
temp_offset = le32_to_cpu(clock_offsets.phy_clk_ns_l);
|
|
ptp->dev_clk_regs.phy_clk_ns_l = idpf_get_reg_addr(adapter,
|
|
temp_offset);
|
|
temp_offset = le32_to_cpu(clock_offsets.phy_clk_ns_h);
|
|
ptp->dev_clk_regs.phy_clk_ns_h = idpf_get_reg_addr(adapter,
|
|
temp_offset);
|
|
temp_offset = le32_to_cpu(clock_offsets.cmd_sync_trigger);
|
|
ptp->dev_clk_regs.cmd_sync = idpf_get_reg_addr(adapter, temp_offset);
|
|
|
|
cross_tstamp:
|
|
access_type = ptp->get_cross_tstamp_access;
|
|
if (access_type != IDPF_PTP_DIRECT)
|
|
goto discipline_clock;
|
|
|
|
cross_tstamp_offsets = recv_ptp_caps_msg->cross_time_offsets;
|
|
|
|
temp_offset = le32_to_cpu(cross_tstamp_offsets.sys_time_ns_l);
|
|
ptp->dev_clk_regs.sys_time_ns_l = idpf_get_reg_addr(adapter,
|
|
temp_offset);
|
|
temp_offset = le32_to_cpu(cross_tstamp_offsets.sys_time_ns_h);
|
|
ptp->dev_clk_regs.sys_time_ns_h = idpf_get_reg_addr(adapter,
|
|
temp_offset);
|
|
temp_offset = le32_to_cpu(cross_tstamp_offsets.cmd_sync_trigger);
|
|
ptp->dev_clk_regs.cmd_sync = idpf_get_reg_addr(adapter, temp_offset);
|
|
|
|
discipline_clock:
|
|
access_type = ptp->adj_dev_clk_time_access;
|
|
if (access_type != IDPF_PTP_DIRECT)
|
|
return 0;
|
|
|
|
clk_adj_offsets = recv_ptp_caps_msg->clk_adj_offsets;
|
|
|
|
/* Device clock offsets */
|
|
temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_cmd_type);
|
|
ptp->dev_clk_regs.cmd = idpf_get_reg_addr(adapter, temp_offset);
|
|
temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_incval_l);
|
|
ptp->dev_clk_regs.incval_l = idpf_get_reg_addr(adapter, temp_offset);
|
|
temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_incval_h);
|
|
ptp->dev_clk_regs.incval_h = idpf_get_reg_addr(adapter, temp_offset);
|
|
temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_shadj_l);
|
|
ptp->dev_clk_regs.shadj_l = idpf_get_reg_addr(adapter, temp_offset);
|
|
temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_shadj_h);
|
|
ptp->dev_clk_regs.shadj_h = idpf_get_reg_addr(adapter, temp_offset);
|
|
|
|
/* PHY clock offsets */
|
|
temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_cmd_type);
|
|
ptp->dev_clk_regs.phy_cmd = idpf_get_reg_addr(adapter, temp_offset);
|
|
temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_incval_l);
|
|
ptp->dev_clk_regs.phy_incval_l = idpf_get_reg_addr(adapter,
|
|
temp_offset);
|
|
temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_incval_h);
|
|
ptp->dev_clk_regs.phy_incval_h = idpf_get_reg_addr(adapter,
|
|
temp_offset);
|
|
temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_shadj_l);
|
|
ptp->dev_clk_regs.phy_shadj_l = idpf_get_reg_addr(adapter, temp_offset);
|
|
temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_shadj_h);
|
|
ptp->dev_clk_regs.phy_shadj_h = idpf_get_reg_addr(adapter, temp_offset);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_ptp_get_dev_clk_time - Send virtchnl get device clk time message
|
|
* @adapter: Driver specific private structure
|
|
* @dev_clk_time: Pointer to the device clock structure where the value is set
|
|
*
|
|
* Send virtchnl get time message to get the time of the clock.
|
|
*
|
|
* Return: 0 on success, -errno otherwise.
|
|
*/
|
|
int idpf_ptp_get_dev_clk_time(struct idpf_adapter *adapter,
|
|
struct idpf_ptp_dev_timers *dev_clk_time)
|
|
{
|
|
struct virtchnl2_ptp_get_dev_clk_time get_dev_clk_time_msg;
|
|
struct idpf_vc_xn_params xn_params = {
|
|
.vc_op = VIRTCHNL2_OP_PTP_GET_DEV_CLK_TIME,
|
|
.send_buf.iov_base = &get_dev_clk_time_msg,
|
|
.send_buf.iov_len = sizeof(get_dev_clk_time_msg),
|
|
.recv_buf.iov_base = &get_dev_clk_time_msg,
|
|
.recv_buf.iov_len = sizeof(get_dev_clk_time_msg),
|
|
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
|
|
};
|
|
int reply_sz;
|
|
u64 dev_time;
|
|
|
|
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
|
|
if (reply_sz < 0)
|
|
return reply_sz;
|
|
if (reply_sz != sizeof(get_dev_clk_time_msg))
|
|
return -EIO;
|
|
|
|
dev_time = le64_to_cpu(get_dev_clk_time_msg.dev_time_ns);
|
|
dev_clk_time->dev_clk_time_ns = dev_time;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_ptp_get_cross_time - Send virtchnl get cross time message
|
|
* @adapter: Driver specific private structure
|
|
* @cross_time: Pointer to the device clock structure where the value is set
|
|
*
|
|
* Send virtchnl get cross time message to get the time of the clock and the
|
|
* system time.
|
|
*
|
|
* Return: 0 on success, -errno otherwise.
|
|
*/
|
|
int idpf_ptp_get_cross_time(struct idpf_adapter *adapter,
|
|
struct idpf_ptp_dev_timers *cross_time)
|
|
{
|
|
struct virtchnl2_ptp_get_cross_time cross_time_msg;
|
|
struct idpf_vc_xn_params xn_params = {
|
|
.vc_op = VIRTCHNL2_OP_PTP_GET_CROSS_TIME,
|
|
.send_buf.iov_base = &cross_time_msg,
|
|
.send_buf.iov_len = sizeof(cross_time_msg),
|
|
.recv_buf.iov_base = &cross_time_msg,
|
|
.recv_buf.iov_len = sizeof(cross_time_msg),
|
|
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
|
|
};
|
|
int reply_sz;
|
|
|
|
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
|
|
if (reply_sz < 0)
|
|
return reply_sz;
|
|
if (reply_sz != sizeof(cross_time_msg))
|
|
return -EIO;
|
|
|
|
cross_time->dev_clk_time_ns = le64_to_cpu(cross_time_msg.dev_time_ns);
|
|
cross_time->sys_time_ns = le64_to_cpu(cross_time_msg.sys_time_ns);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_ptp_set_dev_clk_time - Send virtchnl set device time message
|
|
* @adapter: Driver specific private structure
|
|
* @time: New time value
|
|
*
|
|
* Send virtchnl set time message to set the time of the clock.
|
|
*
|
|
* Return: 0 on success, -errno otherwise.
|
|
*/
|
|
int idpf_ptp_set_dev_clk_time(struct idpf_adapter *adapter, u64 time)
|
|
{
|
|
struct virtchnl2_ptp_set_dev_clk_time set_dev_clk_time_msg = {
|
|
.dev_time_ns = cpu_to_le64(time),
|
|
};
|
|
struct idpf_vc_xn_params xn_params = {
|
|
.vc_op = VIRTCHNL2_OP_PTP_SET_DEV_CLK_TIME,
|
|
.send_buf.iov_base = &set_dev_clk_time_msg,
|
|
.send_buf.iov_len = sizeof(set_dev_clk_time_msg),
|
|
.recv_buf.iov_base = &set_dev_clk_time_msg,
|
|
.recv_buf.iov_len = sizeof(set_dev_clk_time_msg),
|
|
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
|
|
};
|
|
int reply_sz;
|
|
|
|
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
|
|
if (reply_sz < 0)
|
|
return reply_sz;
|
|
if (reply_sz != sizeof(set_dev_clk_time_msg))
|
|
return -EIO;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_ptp_adj_dev_clk_time - Send virtchnl adj device clock time message
|
|
* @adapter: Driver specific private structure
|
|
* @delta: Offset in nanoseconds to adjust the time by
|
|
*
|
|
* Send virtchnl adj time message to adjust the clock by the indicated delta.
|
|
*
|
|
* Return: 0 on success, -errno otherwise.
|
|
*/
|
|
int idpf_ptp_adj_dev_clk_time(struct idpf_adapter *adapter, s64 delta)
|
|
{
|
|
struct virtchnl2_ptp_adj_dev_clk_time adj_dev_clk_time_msg = {
|
|
.delta = cpu_to_le64(delta),
|
|
};
|
|
struct idpf_vc_xn_params xn_params = {
|
|
.vc_op = VIRTCHNL2_OP_PTP_ADJ_DEV_CLK_TIME,
|
|
.send_buf.iov_base = &adj_dev_clk_time_msg,
|
|
.send_buf.iov_len = sizeof(adj_dev_clk_time_msg),
|
|
.recv_buf.iov_base = &adj_dev_clk_time_msg,
|
|
.recv_buf.iov_len = sizeof(adj_dev_clk_time_msg),
|
|
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
|
|
};
|
|
int reply_sz;
|
|
|
|
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
|
|
if (reply_sz < 0)
|
|
return reply_sz;
|
|
if (reply_sz != sizeof(adj_dev_clk_time_msg))
|
|
return -EIO;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_ptp_adj_dev_clk_fine - Send virtchnl adj time message
|
|
* @adapter: Driver specific private structure
|
|
* @incval: Source timer increment value per clock cycle
|
|
*
|
|
* Send virtchnl adj fine message to adjust the frequency of the clock by
|
|
* incval.
|
|
*
|
|
* Return: 0 on success, -errno otherwise.
|
|
*/
|
|
int idpf_ptp_adj_dev_clk_fine(struct idpf_adapter *adapter, u64 incval)
|
|
{
|
|
struct virtchnl2_ptp_adj_dev_clk_fine adj_dev_clk_fine_msg = {
|
|
.incval = cpu_to_le64(incval),
|
|
};
|
|
struct idpf_vc_xn_params xn_params = {
|
|
.vc_op = VIRTCHNL2_OP_PTP_ADJ_DEV_CLK_FINE,
|
|
.send_buf.iov_base = &adj_dev_clk_fine_msg,
|
|
.send_buf.iov_len = sizeof(adj_dev_clk_fine_msg),
|
|
.recv_buf.iov_base = &adj_dev_clk_fine_msg,
|
|
.recv_buf.iov_len = sizeof(adj_dev_clk_fine_msg),
|
|
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
|
|
};
|
|
int reply_sz;
|
|
|
|
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
|
|
if (reply_sz < 0)
|
|
return reply_sz;
|
|
if (reply_sz != sizeof(adj_dev_clk_fine_msg))
|
|
return -EIO;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_ptp_get_vport_tstamps_caps - Send virtchnl to get tstamps caps for vport
|
|
* @vport: Virtual port structure
|
|
*
|
|
* Send virtchnl get vport tstamps caps message to receive the set of tstamp
|
|
* capabilities per vport.
|
|
*
|
|
* Return: 0 on success, -errno otherwise.
|
|
*/
|
|
int idpf_ptp_get_vport_tstamps_caps(struct idpf_vport *vport)
|
|
{
|
|
struct virtchnl2_ptp_get_vport_tx_tstamp_caps send_tx_tstamp_caps;
|
|
struct virtchnl2_ptp_get_vport_tx_tstamp_caps *rcv_tx_tstamp_caps;
|
|
struct virtchnl2_ptp_tx_tstamp_latch_caps tx_tstamp_latch_caps;
|
|
struct idpf_ptp_vport_tx_tstamp_caps *tstamp_caps;
|
|
struct idpf_ptp_tx_tstamp *ptp_tx_tstamp, *tmp;
|
|
struct idpf_vc_xn_params xn_params = {
|
|
.vc_op = VIRTCHNL2_OP_PTP_GET_VPORT_TX_TSTAMP_CAPS,
|
|
.send_buf.iov_base = &send_tx_tstamp_caps,
|
|
.send_buf.iov_len = sizeof(send_tx_tstamp_caps),
|
|
.recv_buf.iov_len = IDPF_CTLQ_MAX_BUF_LEN,
|
|
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
|
|
};
|
|
enum idpf_ptp_access tstamp_access, get_dev_clk_access;
|
|
struct idpf_ptp *ptp = vport->adapter->ptp;
|
|
struct list_head *head;
|
|
int err = 0, reply_sz;
|
|
u16 num_latches;
|
|
u32 size;
|
|
|
|
if (!ptp)
|
|
return -EOPNOTSUPP;
|
|
|
|
tstamp_access = ptp->tx_tstamp_access;
|
|
get_dev_clk_access = ptp->get_dev_clk_time_access;
|
|
if (tstamp_access == IDPF_PTP_NONE ||
|
|
get_dev_clk_access == IDPF_PTP_NONE)
|
|
return -EOPNOTSUPP;
|
|
|
|
rcv_tx_tstamp_caps = kzalloc(IDPF_CTLQ_MAX_BUF_LEN, GFP_KERNEL);
|
|
if (!rcv_tx_tstamp_caps)
|
|
return -ENOMEM;
|
|
|
|
send_tx_tstamp_caps.vport_id = cpu_to_le32(vport->vport_id);
|
|
xn_params.recv_buf.iov_base = rcv_tx_tstamp_caps;
|
|
|
|
reply_sz = idpf_vc_xn_exec(vport->adapter, &xn_params);
|
|
if (reply_sz < 0) {
|
|
err = reply_sz;
|
|
goto get_tstamp_caps_out;
|
|
}
|
|
|
|
num_latches = le16_to_cpu(rcv_tx_tstamp_caps->num_latches);
|
|
size = struct_size(rcv_tx_tstamp_caps, tstamp_latches, num_latches);
|
|
if (reply_sz != size) {
|
|
err = -EIO;
|
|
goto get_tstamp_caps_out;
|
|
}
|
|
|
|
size = struct_size(tstamp_caps, tx_tstamp_status, num_latches);
|
|
tstamp_caps = kzalloc(size, GFP_KERNEL);
|
|
if (!tstamp_caps) {
|
|
err = -ENOMEM;
|
|
goto get_tstamp_caps_out;
|
|
}
|
|
|
|
tstamp_caps->access = true;
|
|
tstamp_caps->num_entries = num_latches;
|
|
|
|
INIT_LIST_HEAD(&tstamp_caps->latches_in_use);
|
|
INIT_LIST_HEAD(&tstamp_caps->latches_free);
|
|
|
|
spin_lock_init(&tstamp_caps->latches_lock);
|
|
spin_lock_init(&tstamp_caps->status_lock);
|
|
|
|
tstamp_caps->tstamp_ns_lo_bit = rcv_tx_tstamp_caps->tstamp_ns_lo_bit;
|
|
|
|
for (u16 i = 0; i < tstamp_caps->num_entries; i++) {
|
|
__le32 offset_l, offset_h;
|
|
|
|
ptp_tx_tstamp = kzalloc(sizeof(*ptp_tx_tstamp), GFP_KERNEL);
|
|
if (!ptp_tx_tstamp) {
|
|
err = -ENOMEM;
|
|
goto err_free_ptp_tx_stamp_list;
|
|
}
|
|
|
|
tx_tstamp_latch_caps = rcv_tx_tstamp_caps->tstamp_latches[i];
|
|
|
|
if (tstamp_access != IDPF_PTP_DIRECT)
|
|
goto skip_offsets;
|
|
|
|
offset_l = tx_tstamp_latch_caps.tx_latch_reg_offset_l;
|
|
offset_h = tx_tstamp_latch_caps.tx_latch_reg_offset_h;
|
|
ptp_tx_tstamp->tx_latch_reg_offset_l = le32_to_cpu(offset_l);
|
|
ptp_tx_tstamp->tx_latch_reg_offset_h = le32_to_cpu(offset_h);
|
|
|
|
skip_offsets:
|
|
ptp_tx_tstamp->idx = tx_tstamp_latch_caps.index;
|
|
|
|
list_add(&ptp_tx_tstamp->list_member,
|
|
&tstamp_caps->latches_free);
|
|
|
|
tstamp_caps->tx_tstamp_status[i].state = IDPF_PTP_FREE;
|
|
}
|
|
|
|
vport->tx_tstamp_caps = tstamp_caps;
|
|
kfree(rcv_tx_tstamp_caps);
|
|
|
|
return 0;
|
|
|
|
err_free_ptp_tx_stamp_list:
|
|
head = &tstamp_caps->latches_free;
|
|
list_for_each_entry_safe(ptp_tx_tstamp, tmp, head, list_member) {
|
|
list_del(&ptp_tx_tstamp->list_member);
|
|
kfree(ptp_tx_tstamp);
|
|
}
|
|
|
|
kfree(tstamp_caps);
|
|
get_tstamp_caps_out:
|
|
kfree(rcv_tx_tstamp_caps);
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* idpf_ptp_update_tstamp_tracker - Update the Tx timestamp tracker based on
|
|
* the skb compatibility.
|
|
* @caps: Tx timestamp capabilities that monitor the latch status
|
|
* @skb: skb for which the tstamp value is returned through virtchnl message
|
|
* @current_state: Current state of the Tx timestamp latch
|
|
* @expected_state: Expected state of the Tx timestamp latch
|
|
*
|
|
* Find a proper skb tracker for which the Tx timestamp is received and change
|
|
* the state to expected value.
|
|
*
|
|
* Return: true if the tracker has been found and updated, false otherwise.
|
|
*/
|
|
static bool
|
|
idpf_ptp_update_tstamp_tracker(struct idpf_ptp_vport_tx_tstamp_caps *caps,
|
|
struct sk_buff *skb,
|
|
enum idpf_ptp_tx_tstamp_state current_state,
|
|
enum idpf_ptp_tx_tstamp_state expected_state)
|
|
{
|
|
bool updated = false;
|
|
|
|
spin_lock(&caps->status_lock);
|
|
for (u16 i = 0; i < caps->num_entries; i++) {
|
|
struct idpf_ptp_tx_tstamp_status *status;
|
|
|
|
status = &caps->tx_tstamp_status[i];
|
|
|
|
if (skb == status->skb && status->state == current_state) {
|
|
status->state = expected_state;
|
|
updated = true;
|
|
break;
|
|
}
|
|
}
|
|
spin_unlock(&caps->status_lock);
|
|
|
|
return updated;
|
|
}
|
|
|
|
/**
|
|
* idpf_ptp_get_tstamp_value - Get the Tx timestamp value and provide it
|
|
* back to the skb.
|
|
* @vport: Virtual port structure
|
|
* @tstamp_latch: Tx timestamp latch structure fulfilled by the Control Plane
|
|
* @ptp_tx_tstamp: Tx timestamp latch to add to the free list
|
|
*
|
|
* Read the value of the Tx timestamp for a given latch received from the
|
|
* Control Plane, extend it to 64 bit and provide back to the skb.
|
|
*
|
|
* Return: 0 on success, -errno otherwise.
|
|
*/
|
|
static int
|
|
idpf_ptp_get_tstamp_value(struct idpf_vport *vport,
|
|
struct virtchnl2_ptp_tx_tstamp_latch *tstamp_latch,
|
|
struct idpf_ptp_tx_tstamp *ptp_tx_tstamp)
|
|
{
|
|
struct idpf_ptp_vport_tx_tstamp_caps *tx_tstamp_caps;
|
|
struct skb_shared_hwtstamps shhwtstamps;
|
|
bool state_upd = false;
|
|
u8 tstamp_ns_lo_bit;
|
|
u64 tstamp;
|
|
|
|
tx_tstamp_caps = vport->tx_tstamp_caps;
|
|
tstamp_ns_lo_bit = tx_tstamp_caps->tstamp_ns_lo_bit;
|
|
|
|
ptp_tx_tstamp->tstamp = le64_to_cpu(tstamp_latch->tstamp);
|
|
ptp_tx_tstamp->tstamp >>= tstamp_ns_lo_bit;
|
|
|
|
state_upd = idpf_ptp_update_tstamp_tracker(tx_tstamp_caps,
|
|
ptp_tx_tstamp->skb,
|
|
IDPF_PTP_READ_VALUE,
|
|
IDPF_PTP_FREE);
|
|
if (!state_upd)
|
|
return -EINVAL;
|
|
|
|
tstamp = idpf_ptp_extend_ts(vport, ptp_tx_tstamp->tstamp);
|
|
shhwtstamps.hwtstamp = ns_to_ktime(tstamp);
|
|
skb_tstamp_tx(ptp_tx_tstamp->skb, &shhwtstamps);
|
|
consume_skb(ptp_tx_tstamp->skb);
|
|
|
|
list_add(&ptp_tx_tstamp->list_member,
|
|
&tx_tstamp_caps->latches_free);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_ptp_get_tx_tstamp_async_handler - Async callback for getting Tx tstamps
|
|
* @adapter: Driver specific private structure
|
|
* @xn: transaction for message
|
|
* @ctlq_msg: received message
|
|
*
|
|
* Read the tstamps Tx tstamp values from a received message and put them
|
|
* directly to the skb. The number of timestamps to read is specified by
|
|
* the virtchnl message.
|
|
*
|
|
* Return: 0 on success, -errno otherwise.
|
|
*/
|
|
static int
|
|
idpf_ptp_get_tx_tstamp_async_handler(struct idpf_adapter *adapter,
|
|
struct idpf_vc_xn *xn,
|
|
const struct idpf_ctlq_msg *ctlq_msg)
|
|
{
|
|
struct virtchnl2_ptp_get_vport_tx_tstamp_latches *recv_tx_tstamp_msg;
|
|
struct idpf_ptp_vport_tx_tstamp_caps *tx_tstamp_caps;
|
|
struct virtchnl2_ptp_tx_tstamp_latch tstamp_latch;
|
|
struct idpf_ptp_tx_tstamp *tx_tstamp, *tmp;
|
|
struct idpf_vport *tstamp_vport = NULL;
|
|
struct list_head *head;
|
|
u16 num_latches;
|
|
u32 vport_id;
|
|
int err = 0;
|
|
|
|
recv_tx_tstamp_msg = ctlq_msg->ctx.indirect.payload->va;
|
|
vport_id = le32_to_cpu(recv_tx_tstamp_msg->vport_id);
|
|
|
|
idpf_for_each_vport(adapter, vport) {
|
|
if (!vport)
|
|
continue;
|
|
|
|
if (vport->vport_id == vport_id) {
|
|
tstamp_vport = vport;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!tstamp_vport || !tstamp_vport->tx_tstamp_caps)
|
|
return -EINVAL;
|
|
|
|
tx_tstamp_caps = tstamp_vport->tx_tstamp_caps;
|
|
num_latches = le16_to_cpu(recv_tx_tstamp_msg->num_latches);
|
|
|
|
spin_lock_bh(&tx_tstamp_caps->latches_lock);
|
|
head = &tx_tstamp_caps->latches_in_use;
|
|
|
|
for (u16 i = 0; i < num_latches; i++) {
|
|
tstamp_latch = recv_tx_tstamp_msg->tstamp_latches[i];
|
|
|
|
if (!tstamp_latch.valid)
|
|
continue;
|
|
|
|
if (list_empty(head)) {
|
|
err = -ENOBUFS;
|
|
goto unlock;
|
|
}
|
|
|
|
list_for_each_entry_safe(tx_tstamp, tmp, head, list_member) {
|
|
if (tstamp_latch.index == tx_tstamp->idx) {
|
|
list_del(&tx_tstamp->list_member);
|
|
err = idpf_ptp_get_tstamp_value(tstamp_vport,
|
|
&tstamp_latch,
|
|
tx_tstamp);
|
|
if (err)
|
|
goto unlock;
|
|
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
unlock:
|
|
spin_unlock_bh(&tx_tstamp_caps->latches_lock);
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* idpf_ptp_get_tx_tstamp - Send virtchnl get Tx timestamp latches message
|
|
* @vport: Virtual port structure
|
|
*
|
|
* Send virtchnl get Tx tstamp message to read the value of the HW timestamp.
|
|
* The message contains a list of indexes set in the Tx descriptors.
|
|
*
|
|
* Return: 0 on success, -errno otherwise.
|
|
*/
|
|
int idpf_ptp_get_tx_tstamp(struct idpf_vport *vport)
|
|
{
|
|
struct virtchnl2_ptp_get_vport_tx_tstamp_latches *send_tx_tstamp_msg;
|
|
struct idpf_ptp_vport_tx_tstamp_caps *tx_tstamp_caps;
|
|
struct idpf_vc_xn_params xn_params = {
|
|
.vc_op = VIRTCHNL2_OP_PTP_GET_VPORT_TX_TSTAMP,
|
|
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
|
|
.async = true,
|
|
.async_handler = idpf_ptp_get_tx_tstamp_async_handler,
|
|
};
|
|
struct idpf_ptp_tx_tstamp *ptp_tx_tstamp;
|
|
int reply_sz, size, msg_size;
|
|
struct list_head *head;
|
|
bool state_upd;
|
|
u16 id = 0;
|
|
|
|
tx_tstamp_caps = vport->tx_tstamp_caps;
|
|
head = &tx_tstamp_caps->latches_in_use;
|
|
|
|
size = struct_size(send_tx_tstamp_msg, tstamp_latches,
|
|
tx_tstamp_caps->num_entries);
|
|
send_tx_tstamp_msg = kzalloc(size, GFP_KERNEL);
|
|
if (!send_tx_tstamp_msg)
|
|
return -ENOMEM;
|
|
|
|
spin_lock_bh(&tx_tstamp_caps->latches_lock);
|
|
list_for_each_entry(ptp_tx_tstamp, head, list_member) {
|
|
u8 idx;
|
|
|
|
state_upd = idpf_ptp_update_tstamp_tracker(tx_tstamp_caps,
|
|
ptp_tx_tstamp->skb,
|
|
IDPF_PTP_REQUEST,
|
|
IDPF_PTP_READ_VALUE);
|
|
if (!state_upd)
|
|
continue;
|
|
|
|
idx = ptp_tx_tstamp->idx;
|
|
send_tx_tstamp_msg->tstamp_latches[id].index = idx;
|
|
id++;
|
|
}
|
|
spin_unlock_bh(&tx_tstamp_caps->latches_lock);
|
|
|
|
msg_size = struct_size(send_tx_tstamp_msg, tstamp_latches, id);
|
|
send_tx_tstamp_msg->vport_id = cpu_to_le32(vport->vport_id);
|
|
send_tx_tstamp_msg->num_latches = cpu_to_le16(id);
|
|
xn_params.send_buf.iov_base = send_tx_tstamp_msg;
|
|
xn_params.send_buf.iov_len = msg_size;
|
|
|
|
reply_sz = idpf_vc_xn_exec(vport->adapter, &xn_params);
|
|
kfree(send_tx_tstamp_msg);
|
|
|
|
return min(reply_sz, 0);
|
|
}
|