/* * Copyright (c) 2015, Mellanox Technologies. All rights reserved. * * This software is available to you under a choice of one of two * licenses. You may choose to be licensed under the terms of the GNU * General Public License (GPL) Version 2, available from the file * COPYING in the main directory of this source tree, or the * OpenIB.org BSD license below: * * Redistribution and use in source and binary forms, with or * without modification, are permitted provided that the following * conditions are met: * * - Redistributions of source code must retain the above * copyright notice, this list of conditions and the following * disclaimer. * * - Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials * provided with the distribution. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include #include #include #include #include "lib/eq.h" #include "en.h" #include "clock.h" enum { MLX5_CYCLES_SHIFT = 23 }; enum { MLX5_PIN_MODE_IN = 0x0, MLX5_PIN_MODE_OUT = 0x1, }; enum { MLX5_OUT_PATTERN_PULSE = 0x0, MLX5_OUT_PATTERN_PERIODIC = 0x1, }; enum { MLX5_EVENT_MODE_DISABLE = 0x0, MLX5_EVENT_MODE_REPETETIVE = 0x1, MLX5_EVENT_MODE_ONCE_TILL_ARM = 0x2, }; enum { MLX5_MTPPS_FS_ENABLE = BIT(0x0), MLX5_MTPPS_FS_PATTERN = BIT(0x2), MLX5_MTPPS_FS_PIN_MODE = BIT(0x3), MLX5_MTPPS_FS_TIME_STAMP = BIT(0x4), MLX5_MTPPS_FS_OUT_PULSE_DURATION = BIT(0x5), MLX5_MTPPS_FS_ENH_OUT_PER_ADJ = BIT(0x7), }; static bool mlx5_real_time_mode(struct mlx5_core_dev *mdev) { return (mlx5_is_real_time_rq(mdev) || mlx5_is_real_time_sq(mdev)); } static bool mlx5_modify_mtutc_allowed(struct mlx5_core_dev *mdev) { return MLX5_CAP_MCAM_FEATURE(mdev, ptpcyc2realtime_modify); } static int mlx5_set_mtutc(struct mlx5_core_dev *dev, u32 *mtutc, u32 size) { u32 out[MLX5_ST_SZ_DW(mtutc_reg)] = {}; if (!MLX5_CAP_MCAM_REG(dev, mtutc)) return -EOPNOTSUPP; return mlx5_core_access_reg(dev, mtutc, size, out, sizeof(out), MLX5_REG_MTUTC, 0, 1); } static u64 mlx5_read_time(struct mlx5_core_dev *dev, struct ptp_system_timestamp *sts, bool real_time) { u32 timer_h, timer_h1, timer_l; timer_h = ioread32be(real_time ? &dev->iseg->real_time_h : &dev->iseg->internal_timer_h); ptp_read_system_prets(sts); timer_l = ioread32be(real_time ? &dev->iseg->real_time_l : &dev->iseg->internal_timer_l); ptp_read_system_postts(sts); timer_h1 = ioread32be(real_time ? &dev->iseg->real_time_h : &dev->iseg->internal_timer_h); if (timer_h != timer_h1) { /* wrap around */ ptp_read_system_prets(sts); timer_l = ioread32be(real_time ? &dev->iseg->real_time_l : &dev->iseg->internal_timer_l); ptp_read_system_postts(sts); } return real_time ? REAL_TIME_TO_NS(timer_h1, timer_l) : (u64)timer_l | (u64)timer_h1 << 32; } static u64 read_internal_timer(const struct cyclecounter *cc) { struct mlx5_timer *timer = container_of(cc, struct mlx5_timer, cycles); struct mlx5_clock *clock = container_of(timer, struct mlx5_clock, timer); struct mlx5_core_dev *mdev = container_of(clock, struct mlx5_core_dev, clock); return mlx5_read_time(mdev, NULL, false) & cc->mask; } static void mlx5_update_clock_info_page(struct mlx5_core_dev *mdev) { struct mlx5_ib_clock_info *clock_info = mdev->clock_info; struct mlx5_clock *clock = &mdev->clock; struct mlx5_timer *timer; u32 sign; if (!clock_info) return; sign = smp_load_acquire(&clock_info->sign); smp_store_mb(clock_info->sign, sign | MLX5_IB_CLOCK_INFO_KERNEL_UPDATING); timer = &clock->timer; clock_info->cycles = timer->tc.cycle_last; clock_info->mult = timer->cycles.mult; clock_info->nsec = timer->tc.nsec; clock_info->frac = timer->tc.frac; smp_store_release(&clock_info->sign, sign + MLX5_IB_CLOCK_INFO_KERNEL_UPDATING * 2); } static void mlx5_pps_out(struct work_struct *work) { struct mlx5_pps *pps_info = container_of(work, struct mlx5_pps, out_work); struct mlx5_clock *clock = container_of(pps_info, struct mlx5_clock, pps_info); struct mlx5_core_dev *mdev = container_of(clock, struct mlx5_core_dev, clock); u32 in[MLX5_ST_SZ_DW(mtpps_reg)] = {0}; unsigned long flags; int i; for (i = 0; i < clock->ptp_info.n_pins; i++) { u64 tstart; write_seqlock_irqsave(&clock->lock, flags); tstart = clock->pps_info.start[i]; clock->pps_info.start[i] = 0; write_sequnlock_irqrestore(&clock->lock, flags); if (!tstart) continue; MLX5_SET(mtpps_reg, in, pin, i); MLX5_SET64(mtpps_reg, in, time_stamp, tstart); MLX5_SET(mtpps_reg, in, field_select, MLX5_MTPPS_FS_TIME_STAMP); mlx5_set_mtpps(mdev, in, sizeof(in)); } } static void mlx5_timestamp_overflow(struct work_struct *work) { struct delayed_work *dwork = to_delayed_work(work); struct mlx5_core_dev *mdev; struct mlx5_timer *timer; struct mlx5_clock *clock; unsigned long flags; timer = container_of(dwork, struct mlx5_timer, overflow_work); clock = container_of(timer, struct mlx5_clock, timer); mdev = container_of(clock, struct mlx5_core_dev, clock); if (mdev->state == MLX5_DEVICE_STATE_INTERNAL_ERROR) goto out; write_seqlock_irqsave(&clock->lock, flags); timecounter_read(&timer->tc); mlx5_update_clock_info_page(mdev); write_sequnlock_irqrestore(&clock->lock, flags); out: schedule_delayed_work(&timer->overflow_work, timer->overflow_period); } static int mlx5_ptp_settime_real_time(struct mlx5_core_dev *mdev, const struct timespec64 *ts) { u32 in[MLX5_ST_SZ_DW(mtutc_reg)] = {}; if (!mlx5_modify_mtutc_allowed(mdev)) return 0; if (ts->tv_sec < 0 || ts->tv_sec > U32_MAX || ts->tv_nsec < 0 || ts->tv_nsec > NSEC_PER_SEC) return -EINVAL; MLX5_SET(mtutc_reg, in, operation, MLX5_MTUTC_OPERATION_SET_TIME_IMMEDIATE); MLX5_SET(mtutc_reg, in, utc_sec, ts->tv_sec); MLX5_SET(mtutc_reg, in, utc_nsec, ts->tv_nsec); return mlx5_set_mtutc(mdev, in, sizeof(in)); } static int mlx5_ptp_settime(struct ptp_clock_info *ptp, const struct timespec64 *ts) { struct mlx5_clock *clock = container_of(ptp, struct mlx5_clock, ptp_info); struct mlx5_timer *timer = &clock->timer; struct mlx5_core_dev *mdev; unsigned long flags; int err; mdev = container_of(clock, struct mlx5_core_dev, clock); err = mlx5_ptp_settime_real_time(mdev, ts); if (err) return err; write_seqlock_irqsave(&clock->lock, flags); timecounter_init(&timer->tc, &timer->cycles, timespec64_to_ns(ts)); mlx5_update_clock_info_page(mdev); write_sequnlock_irqrestore(&clock->lock, flags); return 0; } static struct timespec64 mlx5_ptp_gettimex_real_time(struct mlx5_core_dev *mdev, struct ptp_system_timestamp *sts) { struct timespec64 ts; u64 time; time = mlx5_read_time(mdev, sts, true); ts = ns_to_timespec64(time); return ts; } static int mlx5_ptp_gettimex(struct ptp_clock_info *ptp, struct timespec64 *ts, struct ptp_system_timestamp *sts) { struct mlx5_clock *clock = container_of(ptp, struct mlx5_clock, ptp_info); struct mlx5_timer *timer = &clock->timer; struct mlx5_core_dev *mdev; unsigned long flags; u64 cycles, ns; mdev = container_of(clock, struct mlx5_core_dev, clock); if (mlx5_real_time_mode(mdev)) { *ts = mlx5_ptp_gettimex_real_time(mdev, sts); goto out; } write_seqlock_irqsave(&clock->lock, flags); cycles = mlx5_read_time(mdev, sts, false); ns = timecounter_cyc2time(&timer->tc, cycles); write_sequnlock_irqrestore(&clock->lock, flags); *ts = ns_to_timespec64(ns); out: return 0; } static int mlx5_ptp_adjtime_real_time(struct mlx5_core_dev *mdev, s64 delta) { u32 in[MLX5_ST_SZ_DW(mtutc_reg)] = {}; if (!mlx5_modify_mtutc_allowed(mdev)) return 0; /* HW time adjustment range is s16. If out of range, settime instead */ if (delta < S16_MIN || delta > S16_MAX) { struct timespec64 ts; s64 ns; ts = mlx5_ptp_gettimex_real_time(mdev, NULL); ns = timespec64_to_ns(&ts) + delta; ts = ns_to_timespec64(ns); return mlx5_ptp_settime_real_time(mdev, &ts); } MLX5_SET(mtutc_reg, in, operation, MLX5_MTUTC_OPERATION_ADJUST_TIME); MLX5_SET(mtutc_reg, in, time_adjustment, delta); return mlx5_set_mtutc(mdev, in, sizeof(in)); } static int mlx5_ptp_adjtime(struct ptp_clock_info *ptp, s64 delta) { struct mlx5_clock *clock = container_of(ptp, struct mlx5_clock, ptp_info); struct mlx5_timer *timer = &clock->timer; struct mlx5_core_dev *mdev; unsigned long flags; int err; mdev = container_of(clock, struct mlx5_core_dev, clock); err = mlx5_ptp_adjtime_real_time(mdev, delta); if (err) return err; write_seqlock_irqsave(&clock->lock, flags); timecounter_adjtime(&timer->tc, delta); mlx5_update_clock_info_page(mdev); write_sequnlock_irqrestore(&clock->lock, flags); return 0; } static int mlx5_ptp_adjfreq_real_time(struct mlx5_core_dev *mdev, s32 freq) { u32 in[MLX5_ST_SZ_DW(mtutc_reg)] = {}; if (!mlx5_modify_mtutc_allowed(mdev)) return 0; MLX5_SET(mtutc_reg, in, operation, MLX5_MTUTC_OPERATION_ADJUST_FREQ_UTC); MLX5_SET(mtutc_reg, in, freq_adjustment, freq); return mlx5_set_mtutc(mdev, in, sizeof(in)); } static int mlx5_ptp_adjfreq(struct ptp_clock_info *ptp, s32 delta) { struct mlx5_clock *clock = container_of(ptp, struct mlx5_clock, ptp_info); struct mlx5_timer *timer = &clock->timer; struct mlx5_core_dev *mdev; unsigned long flags; int neg_adj = 0; u32 diff; u64 adj; int err; mdev = container_of(clock, struct mlx5_core_dev, clock); err = mlx5_ptp_adjfreq_real_time(mdev, delta); if (err) return err; if (delta < 0) { neg_adj = 1; delta = -delta; } adj = timer->nominal_c_mult; adj *= delta; diff = div_u64(adj, 1000000000ULL); write_seqlock_irqsave(&clock->lock, flags); timecounter_read(&timer->tc); timer->cycles.mult = neg_adj ? timer->nominal_c_mult - diff : timer->nominal_c_mult + diff; mlx5_update_clock_info_page(mdev); write_sequnlock_irqrestore(&clock->lock, flags); return 0; } static int mlx5_extts_configure(struct ptp_clock_info *ptp, struct ptp_clock_request *rq, int on) { struct mlx5_clock *clock = container_of(ptp, struct mlx5_clock, ptp_info); struct mlx5_core_dev *mdev = container_of(clock, struct mlx5_core_dev, clock); u32 in[MLX5_ST_SZ_DW(mtpps_reg)] = {0}; u32 field_select = 0; u8 pin_mode = 0; u8 pattern = 0; int pin = -1; int err = 0; if (!MLX5_PPS_CAP(mdev)) return -EOPNOTSUPP; /* Reject requests with unsupported flags */ if (rq->extts.flags & ~(PTP_ENABLE_FEATURE | PTP_RISING_EDGE | PTP_FALLING_EDGE | PTP_STRICT_FLAGS)) return -EOPNOTSUPP; /* Reject requests to enable time stamping on both edges. */ if ((rq->extts.flags & PTP_STRICT_FLAGS) && (rq->extts.flags & PTP_ENABLE_FEATURE) && (rq->extts.flags & PTP_EXTTS_EDGES) == PTP_EXTTS_EDGES) return -EOPNOTSUPP; if (rq->extts.index >= clock->ptp_info.n_pins) return -EINVAL; pin = ptp_find_pin(clock->ptp, PTP_PF_EXTTS, rq->extts.index); if (pin < 0) return -EBUSY; if (on) { pin_mode = MLX5_PIN_MODE_IN; pattern = !!(rq->extts.flags & PTP_FALLING_EDGE); field_select = MLX5_MTPPS_FS_PIN_MODE | MLX5_MTPPS_FS_PATTERN | MLX5_MTPPS_FS_ENABLE; } else { field_select = MLX5_MTPPS_FS_ENABLE; } MLX5_SET(mtpps_reg, in, pin, pin); MLX5_SET(mtpps_reg, in, pin_mode, pin_mode); MLX5_SET(mtpps_reg, in, pattern, pattern); MLX5_SET(mtpps_reg, in, enable, on); MLX5_SET(mtpps_reg, in, field_select, field_select); err = mlx5_set_mtpps(mdev, in, sizeof(in)); if (err) return err; return mlx5_set_mtppse(mdev, pin, 0, MLX5_EVENT_MODE_REPETETIVE & on); } static u64 find_target_cycles(struct mlx5_core_dev *mdev, s64 target_ns) { struct mlx5_clock *clock = &mdev->clock; u64 cycles_now, cycles_delta; u64 nsec_now, nsec_delta; struct mlx5_timer *timer; unsigned long flags; timer = &clock->timer; cycles_now = mlx5_read_time(mdev, NULL, false); write_seqlock_irqsave(&clock->lock, flags); nsec_now = timecounter_cyc2time(&timer->tc, cycles_now); nsec_delta = target_ns - nsec_now; cycles_delta = div64_u64(nsec_delta << timer->cycles.shift, timer->cycles.mult); write_sequnlock_irqrestore(&clock->lock, flags); return cycles_now + cycles_delta; } static u64 perout_conf_internal_timer(struct mlx5_core_dev *mdev, s64 sec) { struct timespec64 ts = {}; s64 target_ns; ts.tv_sec = sec; target_ns = timespec64_to_ns(&ts); return find_target_cycles(mdev, target_ns); } static u64 perout_conf_real_time(s64 sec) { return (u64)sec << 32; } static int mlx5_perout_configure(struct ptp_clock_info *ptp, struct ptp_clock_request *rq, int on) { struct mlx5_clock *clock = container_of(ptp, struct mlx5_clock, ptp_info); struct mlx5_core_dev *mdev = container_of(clock, struct mlx5_core_dev, clock); bool rt_mode = mlx5_real_time_mode(mdev); u32 in[MLX5_ST_SZ_DW(mtpps_reg)] = {0}; struct timespec64 ts; u32 field_select = 0; u64 time_stamp = 0; u8 pin_mode = 0; u8 pattern = 0; int pin = -1; int err = 0; s64 ns; if (!MLX5_PPS_CAP(mdev)) return -EOPNOTSUPP; /* Reject requests with unsupported flags */ if (rq->perout.flags) return -EOPNOTSUPP; if (rq->perout.index >= clock->ptp_info.n_pins) return -EINVAL; field_select = MLX5_MTPPS_FS_ENABLE; pin = ptp_find_pin(clock->ptp, PTP_PF_PEROUT, rq->perout.index); if (pin < 0) return -EBUSY; if (on) { bool rt_mode = mlx5_real_time_mode(mdev); s64 sec = rq->perout.start.sec; if (rq->perout.start.nsec) return -EINVAL; pin_mode = MLX5_PIN_MODE_OUT; pattern = MLX5_OUT_PATTERN_PERIODIC; ts.tv_sec = rq->perout.period.sec; ts.tv_nsec = rq->perout.period.nsec; ns = timespec64_to_ns(&ts); if ((ns >> 1) != 500000000LL) return -EINVAL; if (rt_mode && sec > U32_MAX) return -EINVAL; time_stamp = rt_mode ? perout_conf_real_time(sec) : perout_conf_internal_timer(mdev, sec); field_select |= MLX5_MTPPS_FS_PIN_MODE | MLX5_MTPPS_FS_PATTERN | MLX5_MTPPS_FS_TIME_STAMP; } MLX5_SET(mtpps_reg, in, pin, pin); MLX5_SET(mtpps_reg, in, pin_mode, pin_mode); MLX5_SET(mtpps_reg, in, pattern, pattern); MLX5_SET(mtpps_reg, in, enable, on); MLX5_SET64(mtpps_reg, in, time_stamp, time_stamp); MLX5_SET(mtpps_reg, in, field_select, field_select); err = mlx5_set_mtpps(mdev, in, sizeof(in)); if (err) return err; if (rt_mode) return 0; return mlx5_set_mtppse(mdev, pin, 0, MLX5_EVENT_MODE_REPETETIVE & on); } static int mlx5_pps_configure(struct ptp_clock_info *ptp, struct ptp_clock_request *rq, int on) { struct mlx5_clock *clock = container_of(ptp, struct mlx5_clock, ptp_info); clock->pps_info.enabled = !!on; return 0; } static int mlx5_ptp_enable(struct ptp_clock_info *ptp, struct ptp_clock_request *rq, int on) { switch (rq->type) { case PTP_CLK_REQ_EXTTS: return mlx5_extts_configure(ptp, rq, on); case PTP_CLK_REQ_PEROUT: return mlx5_perout_configure(ptp, rq, on); case PTP_CLK_REQ_PPS: return mlx5_pps_configure(ptp, rq, on); default: return -EOPNOTSUPP; } return 0; } enum { MLX5_MTPPS_REG_CAP_PIN_X_MODE_SUPPORT_PPS_IN = BIT(0), MLX5_MTPPS_REG_CAP_PIN_X_MODE_SUPPORT_PPS_OUT = BIT(1), }; static int mlx5_ptp_verify(struct ptp_clock_info *ptp, unsigned int pin, enum ptp_pin_function func, unsigned int chan) { struct mlx5_clock *clock = container_of(ptp, struct mlx5_clock, ptp_info); switch (func) { case PTP_PF_NONE: return 0; case PTP_PF_EXTTS: return !(clock->pps_info.pin_caps[pin] & MLX5_MTPPS_REG_CAP_PIN_X_MODE_SUPPORT_PPS_IN); case PTP_PF_PEROUT: return !(clock->pps_info.pin_caps[pin] & MLX5_MTPPS_REG_CAP_PIN_X_MODE_SUPPORT_PPS_OUT); default: return -EOPNOTSUPP; } } static const struct ptp_clock_info mlx5_ptp_clock_info = { .owner = THIS_MODULE, .name = "mlx5_ptp", .max_adj = 50000000, .n_alarm = 0, .n_ext_ts = 0, .n_per_out = 0, .n_pins = 0, .pps = 0, .adjfreq = mlx5_ptp_adjfreq, .adjtime = mlx5_ptp_adjtime, .gettimex64 = mlx5_ptp_gettimex, .settime64 = mlx5_ptp_settime, .enable = NULL, .verify = NULL, }; static int mlx5_query_mtpps_pin_mode(struct mlx5_core_dev *mdev, u8 pin, u32 *mtpps, u32 mtpps_size) { u32 in[MLX5_ST_SZ_DW(mtpps_reg)] = {}; MLX5_SET(mtpps_reg, in, pin, pin); return mlx5_core_access_reg(mdev, in, sizeof(in), mtpps, mtpps_size, MLX5_REG_MTPPS, 0, 0); } static int mlx5_get_pps_pin_mode(struct mlx5_clock *clock, u8 pin) { struct mlx5_core_dev *mdev = container_of(clock, struct mlx5_core_dev, clock); u32 out[MLX5_ST_SZ_DW(mtpps_reg)] = {}; u8 mode; int err; err = mlx5_query_mtpps_pin_mode(mdev, pin, out, sizeof(out)); if (err || !MLX5_GET(mtpps_reg, out, enable)) return PTP_PF_NONE; mode = MLX5_GET(mtpps_reg, out, pin_mode); if (mode == MLX5_PIN_MODE_IN) return PTP_PF_EXTTS; else if (mode == MLX5_PIN_MODE_OUT) return PTP_PF_PEROUT; return PTP_PF_NONE; } static void mlx5_init_pin_config(struct mlx5_clock *clock) { int i; if (!clock->ptp_info.n_pins) return; clock->ptp_info.pin_config = kcalloc(clock->ptp_info.n_pins, sizeof(*clock->ptp_info.pin_config), GFP_KERNEL); if (!clock->ptp_info.pin_config) return; clock->ptp_info.enable = mlx5_ptp_enable; clock->ptp_info.verify = mlx5_ptp_verify; clock->ptp_info.pps = 1; for (i = 0; i < clock->ptp_info.n_pins; i++) { snprintf(clock->ptp_info.pin_config[i].name, sizeof(clock->ptp_info.pin_config[i].name), "mlx5_pps%d", i); clock->ptp_info.pin_config[i].index = i; clock->ptp_info.pin_config[i].func = mlx5_get_pps_pin_mode(clock, i); clock->ptp_info.pin_config[i].chan = 0; } } static void mlx5_get_pps_caps(struct mlx5_core_dev *mdev) { struct mlx5_clock *clock = &mdev->clock; u32 out[MLX5_ST_SZ_DW(mtpps_reg)] = {0}; mlx5_query_mtpps(mdev, out, sizeof(out)); clock->ptp_info.n_pins = MLX5_GET(mtpps_reg, out, cap_number_of_pps_pins); clock->ptp_info.n_ext_ts = MLX5_GET(mtpps_reg, out, cap_max_num_of_pps_in_pins); clock->ptp_info.n_per_out = MLX5_GET(mtpps_reg, out, cap_max_num_of_pps_out_pins); clock->pps_info.pin_caps[0] = MLX5_GET(mtpps_reg, out, cap_pin_0_mode); clock->pps_info.pin_caps[1] = MLX5_GET(mtpps_reg, out, cap_pin_1_mode); clock->pps_info.pin_caps[2] = MLX5_GET(mtpps_reg, out, cap_pin_2_mode); clock->pps_info.pin_caps[3] = MLX5_GET(mtpps_reg, out, cap_pin_3_mode); clock->pps_info.pin_caps[4] = MLX5_GET(mtpps_reg, out, cap_pin_4_mode); clock->pps_info.pin_caps[5] = MLX5_GET(mtpps_reg, out, cap_pin_5_mode); clock->pps_info.pin_caps[6] = MLX5_GET(mtpps_reg, out, cap_pin_6_mode); clock->pps_info.pin_caps[7] = MLX5_GET(mtpps_reg, out, cap_pin_7_mode); } static void ts_next_sec(struct timespec64 *ts) { ts->tv_sec += 1; ts->tv_nsec = 0; } static u64 perout_conf_next_event_timer(struct mlx5_core_dev *mdev, struct mlx5_clock *clock) { struct timespec64 ts; s64 target_ns; mlx5_ptp_gettimex(&clock->ptp_info, &ts, NULL); ts_next_sec(&ts); target_ns = timespec64_to_ns(&ts); return find_target_cycles(mdev, target_ns); } static int mlx5_pps_event(struct notifier_block *nb, unsigned long type, void *data) { struct mlx5_clock *clock = mlx5_nb_cof(nb, struct mlx5_clock, pps_nb); struct ptp_clock_event ptp_event; struct mlx5_eqe *eqe = data; int pin = eqe->data.pps.pin; struct mlx5_core_dev *mdev; unsigned long flags; u64 ns; mdev = container_of(clock, struct mlx5_core_dev, clock); switch (clock->ptp_info.pin_config[pin].func) { case PTP_PF_EXTTS: ptp_event.index = pin; ptp_event.timestamp = mlx5_real_time_mode(mdev) ? mlx5_real_time_cyc2time(clock, be64_to_cpu(eqe->data.pps.time_stamp)) : mlx5_timecounter_cyc2time(clock, be64_to_cpu(eqe->data.pps.time_stamp)); if (clock->pps_info.enabled) { ptp_event.type = PTP_CLOCK_PPSUSR; ptp_event.pps_times.ts_real = ns_to_timespec64(ptp_event.timestamp); } else { ptp_event.type = PTP_CLOCK_EXTTS; } /* TODOL clock->ptp can be NULL if ptp_clock_register fails */ ptp_clock_event(clock->ptp, &ptp_event); break; case PTP_PF_PEROUT: ns = perout_conf_next_event_timer(mdev, clock); write_seqlock_irqsave(&clock->lock, flags); clock->pps_info.start[pin] = ns; write_sequnlock_irqrestore(&clock->lock, flags); schedule_work(&clock->pps_info.out_work); break; default: mlx5_core_err(mdev, " Unhandled clock PPS event, func %d\n", clock->ptp_info.pin_config[pin].func); } return NOTIFY_OK; } static void mlx5_timecounter_init(struct mlx5_core_dev *mdev) { struct mlx5_clock *clock = &mdev->clock; struct mlx5_timer *timer = &clock->timer; u32 dev_freq; dev_freq = MLX5_CAP_GEN(mdev, device_frequency_khz); timer->cycles.read = read_internal_timer; timer->cycles.shift = MLX5_CYCLES_SHIFT; timer->cycles.mult = clocksource_khz2mult(dev_freq, timer->cycles.shift); timer->nominal_c_mult = timer->cycles.mult; timer->cycles.mask = CLOCKSOURCE_MASK(41); timecounter_init(&timer->tc, &timer->cycles, ktime_to_ns(ktime_get_real())); } static void mlx5_init_overflow_period(struct mlx5_clock *clock) { struct mlx5_core_dev *mdev = container_of(clock, struct mlx5_core_dev, clock); struct mlx5_ib_clock_info *clock_info = mdev->clock_info; struct mlx5_timer *timer = &clock->timer; u64 overflow_cycles; u64 frac = 0; u64 ns; /* Calculate period in seconds to call the overflow watchdog - to make * sure counter is checked at least twice every wrap around. * The period is calculated as the minimum between max HW cycles count * (The clock source mask) and max amount of cycles that can be * multiplied by clock multiplier where the result doesn't exceed * 64bits. */ overflow_cycles = div64_u64(~0ULL >> 1, timer->cycles.mult); overflow_cycles = min(overflow_cycles, div_u64(timer->cycles.mask, 3)); ns = cyclecounter_cyc2ns(&timer->cycles, overflow_cycles, frac, &frac); do_div(ns, NSEC_PER_SEC / HZ); timer->overflow_period = ns; INIT_DELAYED_WORK(&timer->overflow_work, mlx5_timestamp_overflow); if (timer->overflow_period) schedule_delayed_work(&timer->overflow_work, 0); else mlx5_core_warn(mdev, "invalid overflow period, overflow_work is not scheduled\n"); if (clock_info) clock_info->overflow_period = timer->overflow_period; } static void mlx5_init_clock_info(struct mlx5_core_dev *mdev) { struct mlx5_clock *clock = &mdev->clock; struct mlx5_ib_clock_info *info; struct mlx5_timer *timer; mdev->clock_info = (struct mlx5_ib_clock_info *)get_zeroed_page(GFP_KERNEL); if (!mdev->clock_info) { mlx5_core_warn(mdev, "Failed to allocate IB clock info page\n"); return; } info = mdev->clock_info; timer = &clock->timer; info->nsec = timer->tc.nsec; info->cycles = timer->tc.cycle_last; info->mask = timer->cycles.mask; info->mult = timer->nominal_c_mult; info->shift = timer->cycles.shift; info->frac = timer->tc.frac; } static void mlx5_init_timer_clock(struct mlx5_core_dev *mdev) { struct mlx5_clock *clock = &mdev->clock; mlx5_timecounter_init(mdev); mlx5_init_clock_info(mdev); mlx5_init_overflow_period(clock); clock->ptp_info = mlx5_ptp_clock_info; if (mlx5_real_time_mode(mdev)) { struct timespec64 ts; ktime_get_real_ts64(&ts); mlx5_ptp_settime(&clock->ptp_info, &ts); } } static void mlx5_init_pps(struct mlx5_core_dev *mdev) { struct mlx5_clock *clock = &mdev->clock; if (!MLX5_PPS_CAP(mdev)) return; mlx5_get_pps_caps(mdev); mlx5_init_pin_config(clock); } void mlx5_init_clock(struct mlx5_core_dev *mdev) { struct mlx5_clock *clock = &mdev->clock; if (!MLX5_CAP_GEN(mdev, device_frequency_khz)) { mlx5_core_warn(mdev, "invalid device_frequency_khz, aborting HW clock init\n"); return; } seqlock_init(&clock->lock); mlx5_init_timer_clock(mdev); INIT_WORK(&clock->pps_info.out_work, mlx5_pps_out); /* Configure the PHC */ clock->ptp_info = mlx5_ptp_clock_info; /* Initialize 1PPS data structures */ mlx5_init_pps(mdev); clock->ptp = ptp_clock_register(&clock->ptp_info, &mdev->pdev->dev); if (IS_ERR(clock->ptp)) { mlx5_core_warn(mdev, "ptp_clock_register failed %ld\n", PTR_ERR(clock->ptp)); clock->ptp = NULL; } MLX5_NB_INIT(&clock->pps_nb, mlx5_pps_event, PPS_EVENT); mlx5_eq_notifier_register(mdev, &clock->pps_nb); } void mlx5_cleanup_clock(struct mlx5_core_dev *mdev) { struct mlx5_clock *clock = &mdev->clock; if (!MLX5_CAP_GEN(mdev, device_frequency_khz)) return; mlx5_eq_notifier_unregister(mdev, &clock->pps_nb); if (clock->ptp) { ptp_clock_unregister(clock->ptp); clock->ptp = NULL; } cancel_work_sync(&clock->pps_info.out_work); cancel_delayed_work_sync(&clock->timer.overflow_work); if (mdev->clock_info) { free_page((unsigned long)mdev->clock_info); mdev->clock_info = NULL; } kfree(clock->ptp_info.pin_config); }