// SPDX-License-Identifier: GPL-2.0-only /* Copyright (C) 2024 Intel Corporation */ #include "idpf.h" #include "idpf_ptp.h" #include "idpf_virtchnl.h" /** * idpf_ptp_get_caps - Send virtchnl get ptp capabilities message * @adapter: Driver specific private structure * * Send virtchnl get PTP capabilities message. * * Return: 0 on success, -errno on failure. */ int idpf_ptp_get_caps(struct idpf_adapter *adapter) { struct virtchnl2_ptp_get_caps *recv_ptp_caps_msg __free(kfree) = NULL; struct virtchnl2_ptp_get_caps send_ptp_caps_msg = { .caps = cpu_to_le32(VIRTCHNL2_CAP_PTP_GET_DEVICE_CLK_TIME | VIRTCHNL2_CAP_PTP_GET_DEVICE_CLK_TIME_MB | VIRTCHNL2_CAP_PTP_GET_CROSS_TIME | VIRTCHNL2_CAP_PTP_SET_DEVICE_CLK_TIME_MB | VIRTCHNL2_CAP_PTP_ADJ_DEVICE_CLK_MB | VIRTCHNL2_CAP_PTP_TX_TSTAMPS_MB) }; struct idpf_vc_xn_params xn_params = { .vc_op = VIRTCHNL2_OP_PTP_GET_CAPS, .send_buf.iov_base = &send_ptp_caps_msg, .send_buf.iov_len = sizeof(send_ptp_caps_msg), .timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC, }; struct virtchnl2_ptp_clk_adj_reg_offsets clk_adj_offsets; struct virtchnl2_ptp_clk_reg_offsets clock_offsets; struct idpf_ptp_secondary_mbx *scnd_mbx; struct idpf_ptp *ptp = adapter->ptp; enum idpf_ptp_access access_type; u32 temp_offset; int reply_sz; recv_ptp_caps_msg = kzalloc(sizeof(struct virtchnl2_ptp_get_caps), GFP_KERNEL); if (!recv_ptp_caps_msg) return -ENOMEM; xn_params.recv_buf.iov_base = recv_ptp_caps_msg; xn_params.recv_buf.iov_len = sizeof(*recv_ptp_caps_msg); reply_sz = idpf_vc_xn_exec(adapter, &xn_params); if (reply_sz < 0) return reply_sz; else if (reply_sz != sizeof(*recv_ptp_caps_msg)) return -EIO; ptp->caps = le32_to_cpu(recv_ptp_caps_msg->caps); ptp->base_incval = le64_to_cpu(recv_ptp_caps_msg->base_incval); ptp->max_adj = le32_to_cpu(recv_ptp_caps_msg->max_adj); scnd_mbx = &ptp->secondary_mbx; scnd_mbx->peer_mbx_q_id = le16_to_cpu(recv_ptp_caps_msg->peer_mbx_q_id); /* if the ptp_mb_q_id holds invalid value (0xffff), the secondary * mailbox is not supported. */ scnd_mbx->valid = scnd_mbx->peer_mbx_q_id != 0xffff; if (scnd_mbx->valid) scnd_mbx->peer_id = recv_ptp_caps_msg->peer_id; /* Determine the access type for the PTP features */ idpf_ptp_get_features_access(adapter); access_type = ptp->get_dev_clk_time_access; if (access_type != IDPF_PTP_DIRECT) goto discipline_clock; clock_offsets = recv_ptp_caps_msg->clk_offsets; temp_offset = le32_to_cpu(clock_offsets.dev_clk_ns_l); ptp->dev_clk_regs.dev_clk_ns_l = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clock_offsets.dev_clk_ns_h); ptp->dev_clk_regs.dev_clk_ns_h = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clock_offsets.phy_clk_ns_l); ptp->dev_clk_regs.phy_clk_ns_l = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clock_offsets.phy_clk_ns_h); ptp->dev_clk_regs.phy_clk_ns_h = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clock_offsets.cmd_sync_trigger); ptp->dev_clk_regs.cmd_sync = idpf_get_reg_addr(adapter, temp_offset); discipline_clock: access_type = ptp->adj_dev_clk_time_access; if (access_type != IDPF_PTP_DIRECT) return 0; clk_adj_offsets = recv_ptp_caps_msg->clk_adj_offsets; /* Device clock offsets */ temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_cmd_type); ptp->dev_clk_regs.cmd = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_incval_l); ptp->dev_clk_regs.incval_l = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_incval_h); ptp->dev_clk_regs.incval_h = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_shadj_l); ptp->dev_clk_regs.shadj_l = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_shadj_h); ptp->dev_clk_regs.shadj_h = idpf_get_reg_addr(adapter, temp_offset); /* PHY clock offsets */ temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_cmd_type); ptp->dev_clk_regs.phy_cmd = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_incval_l); ptp->dev_clk_regs.phy_incval_l = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_incval_h); ptp->dev_clk_regs.phy_incval_h = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_shadj_l); ptp->dev_clk_regs.phy_shadj_l = idpf_get_reg_addr(adapter, temp_offset); temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_shadj_h); ptp->dev_clk_regs.phy_shadj_h = idpf_get_reg_addr(adapter, temp_offset); return 0; } /** * idpf_ptp_get_dev_clk_time - Send virtchnl get device clk time message * @adapter: Driver specific private structure * @dev_clk_time: Pointer to the device clock structure where the value is set * * Send virtchnl get time message to get the time of the clock. * * Return: 0 on success, -errno otherwise. */ int idpf_ptp_get_dev_clk_time(struct idpf_adapter *adapter, struct idpf_ptp_dev_timers *dev_clk_time) { struct virtchnl2_ptp_get_dev_clk_time get_dev_clk_time_msg; struct idpf_vc_xn_params xn_params = { .vc_op = VIRTCHNL2_OP_PTP_GET_DEV_CLK_TIME, .send_buf.iov_base = &get_dev_clk_time_msg, .send_buf.iov_len = sizeof(get_dev_clk_time_msg), .recv_buf.iov_base = &get_dev_clk_time_msg, .recv_buf.iov_len = sizeof(get_dev_clk_time_msg), .timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC, }; int reply_sz; u64 dev_time; reply_sz = idpf_vc_xn_exec(adapter, &xn_params); if (reply_sz < 0) return reply_sz; if (reply_sz != sizeof(get_dev_clk_time_msg)) return -EIO; dev_time = le64_to_cpu(get_dev_clk_time_msg.dev_time_ns); dev_clk_time->dev_clk_time_ns = dev_time; return 0; } /** * idpf_ptp_set_dev_clk_time - Send virtchnl set device time message * @adapter: Driver specific private structure * @time: New time value * * Send virtchnl set time message to set the time of the clock. * * Return: 0 on success, -errno otherwise. */ int idpf_ptp_set_dev_clk_time(struct idpf_adapter *adapter, u64 time) { struct virtchnl2_ptp_set_dev_clk_time set_dev_clk_time_msg = { .dev_time_ns = cpu_to_le64(time), }; struct idpf_vc_xn_params xn_params = { .vc_op = VIRTCHNL2_OP_PTP_SET_DEV_CLK_TIME, .send_buf.iov_base = &set_dev_clk_time_msg, .send_buf.iov_len = sizeof(set_dev_clk_time_msg), .recv_buf.iov_base = &set_dev_clk_time_msg, .recv_buf.iov_len = sizeof(set_dev_clk_time_msg), .timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC, }; int reply_sz; reply_sz = idpf_vc_xn_exec(adapter, &xn_params); if (reply_sz < 0) return reply_sz; if (reply_sz != sizeof(set_dev_clk_time_msg)) return -EIO; return 0; } /** * idpf_ptp_adj_dev_clk_time - Send virtchnl adj device clock time message * @adapter: Driver specific private structure * @delta: Offset in nanoseconds to adjust the time by * * Send virtchnl adj time message to adjust the clock by the indicated delta. * * Return: 0 on success, -errno otherwise. */ int idpf_ptp_adj_dev_clk_time(struct idpf_adapter *adapter, s64 delta) { struct virtchnl2_ptp_adj_dev_clk_time adj_dev_clk_time_msg = { .delta = cpu_to_le64(delta), }; struct idpf_vc_xn_params xn_params = { .vc_op = VIRTCHNL2_OP_PTP_ADJ_DEV_CLK_TIME, .send_buf.iov_base = &adj_dev_clk_time_msg, .send_buf.iov_len = sizeof(adj_dev_clk_time_msg), .recv_buf.iov_base = &adj_dev_clk_time_msg, .recv_buf.iov_len = sizeof(adj_dev_clk_time_msg), .timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC, }; int reply_sz; reply_sz = idpf_vc_xn_exec(adapter, &xn_params); if (reply_sz < 0) return reply_sz; if (reply_sz != sizeof(adj_dev_clk_time_msg)) return -EIO; return 0; } /** * idpf_ptp_adj_dev_clk_fine - Send virtchnl adj time message * @adapter: Driver specific private structure * @incval: Source timer increment value per clock cycle * * Send virtchnl adj fine message to adjust the frequency of the clock by * incval. * * Return: 0 on success, -errno otherwise. */ int idpf_ptp_adj_dev_clk_fine(struct idpf_adapter *adapter, u64 incval) { struct virtchnl2_ptp_adj_dev_clk_fine adj_dev_clk_fine_msg = { .incval = cpu_to_le64(incval), }; struct idpf_vc_xn_params xn_params = { .vc_op = VIRTCHNL2_OP_PTP_ADJ_DEV_CLK_FINE, .send_buf.iov_base = &adj_dev_clk_fine_msg, .send_buf.iov_len = sizeof(adj_dev_clk_fine_msg), .recv_buf.iov_base = &adj_dev_clk_fine_msg, .recv_buf.iov_len = sizeof(adj_dev_clk_fine_msg), .timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC, }; int reply_sz; reply_sz = idpf_vc_xn_exec(adapter, &xn_params); if (reply_sz < 0) return reply_sz; if (reply_sz != sizeof(adj_dev_clk_fine_msg)) return -EIO; return 0; } /** * idpf_ptp_get_vport_tstamps_caps - Send virtchnl to get tstamps caps for vport * @vport: Virtual port structure * * Send virtchnl get vport tstamps caps message to receive the set of tstamp * capabilities per vport. * * Return: 0 on success, -errno otherwise. */ int idpf_ptp_get_vport_tstamps_caps(struct idpf_vport *vport) { struct virtchnl2_ptp_get_vport_tx_tstamp_caps send_tx_tstamp_caps; struct virtchnl2_ptp_get_vport_tx_tstamp_caps *rcv_tx_tstamp_caps; struct virtchnl2_ptp_tx_tstamp_latch_caps tx_tstamp_latch_caps; struct idpf_ptp_vport_tx_tstamp_caps *tstamp_caps; struct idpf_ptp_tx_tstamp *ptp_tx_tstamp, *tmp; struct idpf_vc_xn_params xn_params = { .vc_op = VIRTCHNL2_OP_PTP_GET_VPORT_TX_TSTAMP_CAPS, .send_buf.iov_base = &send_tx_tstamp_caps, .send_buf.iov_len = sizeof(send_tx_tstamp_caps), .recv_buf.iov_len = IDPF_CTLQ_MAX_BUF_LEN, .timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC, }; enum idpf_ptp_access tstamp_access, get_dev_clk_access; struct idpf_ptp *ptp = vport->adapter->ptp; struct list_head *head; int err = 0, reply_sz; u16 num_latches; u32 size; if (!ptp) return -EOPNOTSUPP; tstamp_access = ptp->tx_tstamp_access; get_dev_clk_access = ptp->get_dev_clk_time_access; if (tstamp_access == IDPF_PTP_NONE || get_dev_clk_access == IDPF_PTP_NONE) return -EOPNOTSUPP; rcv_tx_tstamp_caps = kzalloc(IDPF_CTLQ_MAX_BUF_LEN, GFP_KERNEL); if (!rcv_tx_tstamp_caps) return -ENOMEM; send_tx_tstamp_caps.vport_id = cpu_to_le32(vport->vport_id); xn_params.recv_buf.iov_base = rcv_tx_tstamp_caps; reply_sz = idpf_vc_xn_exec(vport->adapter, &xn_params); if (reply_sz < 0) { err = reply_sz; goto get_tstamp_caps_out; } num_latches = le16_to_cpu(rcv_tx_tstamp_caps->num_latches); size = struct_size(rcv_tx_tstamp_caps, tstamp_latches, num_latches); if (reply_sz != size) { err = -EIO; goto get_tstamp_caps_out; } size = struct_size(tstamp_caps, tx_tstamp_status, num_latches); tstamp_caps = kzalloc(size, GFP_KERNEL); if (!tstamp_caps) { err = -ENOMEM; goto get_tstamp_caps_out; } tstamp_caps->access = true; tstamp_caps->num_entries = num_latches; INIT_LIST_HEAD(&tstamp_caps->latches_in_use); INIT_LIST_HEAD(&tstamp_caps->latches_free); spin_lock_init(&tstamp_caps->latches_lock); spin_lock_init(&tstamp_caps->status_lock); tstamp_caps->tstamp_ns_lo_bit = rcv_tx_tstamp_caps->tstamp_ns_lo_bit; for (u16 i = 0; i < tstamp_caps->num_entries; i++) { __le32 offset_l, offset_h; ptp_tx_tstamp = kzalloc(sizeof(*ptp_tx_tstamp), GFP_KERNEL); if (!ptp_tx_tstamp) { err = -ENOMEM; goto err_free_ptp_tx_stamp_list; } tx_tstamp_latch_caps = rcv_tx_tstamp_caps->tstamp_latches[i]; if (tstamp_access != IDPF_PTP_DIRECT) goto skip_offsets; offset_l = tx_tstamp_latch_caps.tx_latch_reg_offset_l; offset_h = tx_tstamp_latch_caps.tx_latch_reg_offset_h; ptp_tx_tstamp->tx_latch_reg_offset_l = le32_to_cpu(offset_l); ptp_tx_tstamp->tx_latch_reg_offset_h = le32_to_cpu(offset_h); skip_offsets: ptp_tx_tstamp->idx = tx_tstamp_latch_caps.index; list_add(&ptp_tx_tstamp->list_member, &tstamp_caps->latches_free); tstamp_caps->tx_tstamp_status[i].state = IDPF_PTP_FREE; } vport->tx_tstamp_caps = tstamp_caps; kfree(rcv_tx_tstamp_caps); return 0; err_free_ptp_tx_stamp_list: head = &tstamp_caps->latches_free; list_for_each_entry_safe(ptp_tx_tstamp, tmp, head, list_member) { list_del(&ptp_tx_tstamp->list_member); kfree(ptp_tx_tstamp); } kfree(tstamp_caps); get_tstamp_caps_out: kfree(rcv_tx_tstamp_caps); return err; } /** * idpf_ptp_update_tstamp_tracker - Update the Tx timestamp tracker based on * the skb compatibility. * @caps: Tx timestamp capabilities that monitor the latch status * @skb: skb for which the tstamp value is returned through virtchnl message * @current_state: Current state of the Tx timestamp latch * @expected_state: Expected state of the Tx timestamp latch * * Find a proper skb tracker for which the Tx timestamp is received and change * the state to expected value. * * Return: true if the tracker has been found and updated, false otherwise. */ static bool idpf_ptp_update_tstamp_tracker(struct idpf_ptp_vport_tx_tstamp_caps *caps, struct sk_buff *skb, enum idpf_ptp_tx_tstamp_state current_state, enum idpf_ptp_tx_tstamp_state expected_state) { bool updated = false; spin_lock(&caps->status_lock); for (u16 i = 0; i < caps->num_entries; i++) { struct idpf_ptp_tx_tstamp_status *status; status = &caps->tx_tstamp_status[i]; if (skb == status->skb && status->state == current_state) { status->state = expected_state; updated = true; break; } } spin_unlock(&caps->status_lock); return updated; } /** * idpf_ptp_get_tstamp_value - Get the Tx timestamp value and provide it * back to the skb. * @vport: Virtual port structure * @tstamp_latch: Tx timestamp latch structure fulfilled by the Control Plane * @ptp_tx_tstamp: Tx timestamp latch to add to the free list * * Read the value of the Tx timestamp for a given latch received from the * Control Plane, extend it to 64 bit and provide back to the skb. * * Return: 0 on success, -errno otherwise. */ static int idpf_ptp_get_tstamp_value(struct idpf_vport *vport, struct virtchnl2_ptp_tx_tstamp_latch *tstamp_latch, struct idpf_ptp_tx_tstamp *ptp_tx_tstamp) { struct idpf_ptp_vport_tx_tstamp_caps *tx_tstamp_caps; struct skb_shared_hwtstamps shhwtstamps; bool state_upd = false; u8 tstamp_ns_lo_bit; u64 tstamp; tx_tstamp_caps = vport->tx_tstamp_caps; tstamp_ns_lo_bit = tx_tstamp_caps->tstamp_ns_lo_bit; ptp_tx_tstamp->tstamp = le64_to_cpu(tstamp_latch->tstamp); ptp_tx_tstamp->tstamp >>= tstamp_ns_lo_bit; state_upd = idpf_ptp_update_tstamp_tracker(tx_tstamp_caps, ptp_tx_tstamp->skb, IDPF_PTP_READ_VALUE, IDPF_PTP_FREE); if (!state_upd) return -EINVAL; tstamp = idpf_ptp_extend_ts(vport, ptp_tx_tstamp->tstamp); shhwtstamps.hwtstamp = ns_to_ktime(tstamp); skb_tstamp_tx(ptp_tx_tstamp->skb, &shhwtstamps); consume_skb(ptp_tx_tstamp->skb); list_add(&ptp_tx_tstamp->list_member, &tx_tstamp_caps->latches_free); return 0; } /** * idpf_ptp_get_tx_tstamp_async_handler - Async callback for getting Tx tstamps * @adapter: Driver specific private structure * @xn: transaction for message * @ctlq_msg: received message * * Read the tstamps Tx tstamp values from a received message and put them * directly to the skb. The number of timestamps to read is specified by * the virtchnl message. * * Return: 0 on success, -errno otherwise. */ static int idpf_ptp_get_tx_tstamp_async_handler(struct idpf_adapter *adapter, struct idpf_vc_xn *xn, const struct idpf_ctlq_msg *ctlq_msg) { struct virtchnl2_ptp_get_vport_tx_tstamp_latches *recv_tx_tstamp_msg; struct idpf_ptp_vport_tx_tstamp_caps *tx_tstamp_caps; struct virtchnl2_ptp_tx_tstamp_latch tstamp_latch; struct idpf_ptp_tx_tstamp *tx_tstamp, *tmp; struct idpf_vport *tstamp_vport = NULL; struct list_head *head; u16 num_latches; u32 vport_id; int err = 0; recv_tx_tstamp_msg = ctlq_msg->ctx.indirect.payload->va; vport_id = le32_to_cpu(recv_tx_tstamp_msg->vport_id); idpf_for_each_vport(adapter, vport) { if (!vport) continue; if (vport->vport_id == vport_id) { tstamp_vport = vport; break; } } if (!tstamp_vport || !tstamp_vport->tx_tstamp_caps) return -EINVAL; tx_tstamp_caps = tstamp_vport->tx_tstamp_caps; num_latches = le16_to_cpu(recv_tx_tstamp_msg->num_latches); spin_lock_bh(&tx_tstamp_caps->latches_lock); head = &tx_tstamp_caps->latches_in_use; for (u16 i = 0; i < num_latches; i++) { tstamp_latch = recv_tx_tstamp_msg->tstamp_latches[i]; if (!tstamp_latch.valid) continue; if (list_empty(head)) { err = -ENOBUFS; goto unlock; } list_for_each_entry_safe(tx_tstamp, tmp, head, list_member) { if (tstamp_latch.index == tx_tstamp->idx) { list_del(&tx_tstamp->list_member); err = idpf_ptp_get_tstamp_value(tstamp_vport, &tstamp_latch, tx_tstamp); if (err) goto unlock; break; } } } unlock: spin_unlock_bh(&tx_tstamp_caps->latches_lock); return err; } /** * idpf_ptp_get_tx_tstamp - Send virtchnl get Tx timestamp latches message * @vport: Virtual port structure * * Send virtchnl get Tx tstamp message to read the value of the HW timestamp. * The message contains a list of indexes set in the Tx descriptors. * * Return: 0 on success, -errno otherwise. */ int idpf_ptp_get_tx_tstamp(struct idpf_vport *vport) { struct virtchnl2_ptp_get_vport_tx_tstamp_latches *send_tx_tstamp_msg; struct idpf_ptp_vport_tx_tstamp_caps *tx_tstamp_caps; struct idpf_vc_xn_params xn_params = { .vc_op = VIRTCHNL2_OP_PTP_GET_VPORT_TX_TSTAMP, .timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC, .async = true, .async_handler = idpf_ptp_get_tx_tstamp_async_handler, }; struct idpf_ptp_tx_tstamp *ptp_tx_tstamp; int reply_sz, size, msg_size; struct list_head *head; bool state_upd; u16 id = 0; tx_tstamp_caps = vport->tx_tstamp_caps; head = &tx_tstamp_caps->latches_in_use; size = struct_size(send_tx_tstamp_msg, tstamp_latches, tx_tstamp_caps->num_entries); send_tx_tstamp_msg = kzalloc(size, GFP_KERNEL); if (!send_tx_tstamp_msg) return -ENOMEM; spin_lock_bh(&tx_tstamp_caps->latches_lock); list_for_each_entry(ptp_tx_tstamp, head, list_member) { u8 idx; state_upd = idpf_ptp_update_tstamp_tracker(tx_tstamp_caps, ptp_tx_tstamp->skb, IDPF_PTP_REQUEST, IDPF_PTP_READ_VALUE); if (!state_upd) continue; idx = ptp_tx_tstamp->idx; send_tx_tstamp_msg->tstamp_latches[id].index = idx; id++; } spin_unlock_bh(&tx_tstamp_caps->latches_lock); msg_size = struct_size(send_tx_tstamp_msg, tstamp_latches, id); send_tx_tstamp_msg->vport_id = cpu_to_le32(vport->vport_id); send_tx_tstamp_msg->num_latches = cpu_to_le16(id); xn_params.send_buf.iov_base = send_tx_tstamp_msg; xn_params.send_buf.iov_len = msg_size; reply_sz = idpf_vc_xn_exec(vport->adapter, &xn_params); kfree(send_tx_tstamp_msg); return min(reply_sz, 0); }