1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * PTP 1588 clock support
5 * Copyright (C) 2010 OMICRON electronics GmbH
8 #include <linux/device.h>
10 #include <linux/init.h>
11 #include <linux/kernel.h>
12 #include <linux/module.h>
13 #include <linux/posix-clock.h>
14 #include <linux/pps_kernel.h>
15 #include <linux/slab.h>
16 #include <linux/syscalls.h>
17 #include <linux/uaccess.h>
18 #include <linux/debugfs.h>
19 #include <uapi/linux/sched/types.h>
21 #include "ptp_private.h"
23 #define PTP_MAX_ALARMS 4
24 #define PTP_PPS_DEFAULTS (PPS_CAPTUREASSERT | PPS_OFFSETASSERT)
25 #define PTP_PPS_EVENT PPS_CAPTUREASSERT
26 #define PTP_PPS_MODE (PTP_PPS_DEFAULTS | PPS_CANWAIT | PPS_TSFMT_TSPEC)
28 struct class *ptp_class;
32 static dev_t ptp_devt;
34 static DEFINE_IDA(ptp_clocks_map);
36 /* time stamp event queue operations */
38 static inline int queue_free(struct timestamp_event_queue *q)
40 return PTP_MAX_TIMESTAMPS - queue_cnt(q) - 1;
43 static void enqueue_external_timestamp(struct timestamp_event_queue *queue,
44 struct ptp_clock_event *src)
46 struct ptp_extts_event *dst;
51 seconds = div_u64_rem(src->timestamp, 1000000000, &remainder);
53 spin_lock_irqsave(&queue->lock, flags);
55 dst = &queue->buf[queue->tail];
56 dst->index = src->index;
58 dst->t.nsec = remainder;
60 /* Both WRITE_ONCE() are paired with READ_ONCE() in queue_cnt() */
61 if (!queue_free(queue))
62 WRITE_ONCE(queue->head, (queue->head + 1) % PTP_MAX_TIMESTAMPS);
64 WRITE_ONCE(queue->tail, (queue->tail + 1) % PTP_MAX_TIMESTAMPS);
66 spin_unlock_irqrestore(&queue->lock, flags);
69 /* posix clock implementation */
71 static int ptp_clock_getres(struct posix_clock *pc, struct timespec64 *tp)
78 static int ptp_clock_settime(struct posix_clock *pc, const struct timespec64 *tp)
80 struct ptp_clock *ptp = container_of(pc, struct ptp_clock, clock);
82 if (ptp_clock_freerun(ptp)) {
83 pr_err("ptp: physical clock is free running\n");
87 return ptp->info->settime64(ptp->info, tp);
90 static int ptp_clock_gettime(struct posix_clock *pc, struct timespec64 *tp)
92 struct ptp_clock *ptp = container_of(pc, struct ptp_clock, clock);
95 if (ptp->info->gettimex64)
96 err = ptp->info->gettimex64(ptp->info, tp, NULL);
98 err = ptp->info->gettime64(ptp->info, tp);
102 static int ptp_clock_adjtime(struct posix_clock *pc, struct __kernel_timex *tx)
104 struct ptp_clock *ptp = container_of(pc, struct ptp_clock, clock);
105 struct ptp_clock_info *ops;
106 int err = -EOPNOTSUPP;
108 if (ptp_clock_freerun(ptp)) {
109 pr_err("ptp: physical clock is free running\n");
115 if (tx->modes & ADJ_SETOFFSET) {
116 struct timespec64 ts;
120 ts.tv_sec = tx->time.tv_sec;
121 ts.tv_nsec = tx->time.tv_usec;
123 if (!(tx->modes & ADJ_NANO))
126 if ((unsigned long) ts.tv_nsec >= NSEC_PER_SEC)
129 kt = timespec64_to_ktime(ts);
130 delta = ktime_to_ns(kt);
131 err = ops->adjtime(ops, delta);
132 } else if (tx->modes & ADJ_FREQUENCY) {
133 long ppb = scaled_ppm_to_ppb(tx->freq);
134 if (ppb > ops->max_adj || ppb < -ops->max_adj)
136 err = ops->adjfine(ops, tx->freq);
137 ptp->dialed_frequency = tx->freq;
138 } else if (tx->modes & ADJ_OFFSET) {
140 s32 max_phase_adj = ops->getmaxphase(ops);
141 s32 offset = tx->offset;
143 if (!(tx->modes & ADJ_NANO))
144 offset *= NSEC_PER_USEC;
146 if (offset > max_phase_adj || offset < -max_phase_adj)
149 err = ops->adjphase(ops, offset);
151 } else if (tx->modes == 0) {
152 tx->freq = ptp->dialed_frequency;
159 static struct posix_clock_operations ptp_clock_ops = {
160 .owner = THIS_MODULE,
161 .clock_adjtime = ptp_clock_adjtime,
162 .clock_gettime = ptp_clock_gettime,
163 .clock_getres = ptp_clock_getres,
164 .clock_settime = ptp_clock_settime,
167 .release = ptp_release,
172 static void ptp_clock_release(struct device *dev)
174 struct ptp_clock *ptp = container_of(dev, struct ptp_clock, dev);
175 struct timestamp_event_queue *tsevq;
178 ptp_cleanup_pin_groups(ptp);
179 kfree(ptp->vclock_index);
180 mutex_destroy(&ptp->pincfg_mux);
181 mutex_destroy(&ptp->n_vclocks_mux);
182 /* Delete first entry */
183 spin_lock_irqsave(&ptp->tsevqs_lock, flags);
184 tsevq = list_first_entry(&ptp->tsevqs, struct timestamp_event_queue,
186 list_del(&tsevq->qlist);
187 spin_unlock_irqrestore(&ptp->tsevqs_lock, flags);
188 bitmap_free(tsevq->mask);
190 debugfs_remove(ptp->debugfs_root);
191 ida_free(&ptp_clocks_map, ptp->index);
195 static int ptp_getcycles64(struct ptp_clock_info *info, struct timespec64 *ts)
197 if (info->getcyclesx64)
198 return info->getcyclesx64(info, ts, NULL);
200 return info->gettime64(info, ts);
203 static void ptp_aux_kworker(struct kthread_work *work)
205 struct ptp_clock *ptp = container_of(work, struct ptp_clock,
207 struct ptp_clock_info *info = ptp->info;
210 delay = info->do_aux_work(info);
213 kthread_queue_delayed_work(ptp->kworker, &ptp->aux_work, delay);
216 /* public interface */
218 struct ptp_clock *ptp_clock_register(struct ptp_clock_info *info,
219 struct device *parent)
221 struct ptp_clock *ptp;
222 struct timestamp_event_queue *queue = NULL;
223 int err = 0, index, major = MAJOR(ptp_devt);
224 char debugfsname[16];
227 if (info->n_alarm > PTP_MAX_ALARMS)
228 return ERR_PTR(-EINVAL);
230 /* Initialize a clock structure. */
232 ptp = kzalloc(sizeof(struct ptp_clock), GFP_KERNEL);
236 index = ida_alloc_max(&ptp_clocks_map, MINORMASK, GFP_KERNEL);
242 ptp->clock.ops = ptp_clock_ops;
244 ptp->devid = MKDEV(major, index);
246 INIT_LIST_HEAD(&ptp->tsevqs);
247 queue = kzalloc(sizeof(*queue), GFP_KERNEL);
249 goto no_memory_queue;
250 list_add_tail(&queue->qlist, &ptp->tsevqs);
251 spin_lock_init(&ptp->tsevqs_lock);
252 queue->mask = bitmap_alloc(PTP_MAX_CHANNELS, GFP_KERNEL);
254 goto no_memory_bitmap;
255 bitmap_set(queue->mask, 0, PTP_MAX_CHANNELS);
256 spin_lock_init(&queue->lock);
257 mutex_init(&ptp->pincfg_mux);
258 mutex_init(&ptp->n_vclocks_mux);
259 init_waitqueue_head(&ptp->tsev_wq);
261 if (ptp->info->getcycles64 || ptp->info->getcyclesx64) {
262 ptp->has_cycles = true;
263 if (!ptp->info->getcycles64 && ptp->info->getcyclesx64)
264 ptp->info->getcycles64 = ptp_getcycles64;
266 /* Free running cycle counter not supported, use time. */
267 ptp->info->getcycles64 = ptp_getcycles64;
269 if (ptp->info->gettimex64)
270 ptp->info->getcyclesx64 = ptp->info->gettimex64;
272 if (ptp->info->getcrosststamp)
273 ptp->info->getcrosscycles = ptp->info->getcrosststamp;
276 if (ptp->info->do_aux_work) {
277 kthread_init_delayed_work(&ptp->aux_work, ptp_aux_kworker);
278 ptp->kworker = kthread_create_worker(0, "ptp%d", ptp->index);
279 if (IS_ERR(ptp->kworker)) {
280 err = PTR_ERR(ptp->kworker);
281 pr_err("failed to create ptp aux_worker %d\n", err);
286 /* PTP virtual clock is being registered under physical clock */
287 if (parent && parent->class && parent->class->name &&
288 strcmp(parent->class->name, "ptp") == 0)
289 ptp->is_virtual_clock = true;
291 if (!ptp->is_virtual_clock) {
292 ptp->max_vclocks = PTP_DEFAULT_MAX_VCLOCKS;
294 size = sizeof(int) * ptp->max_vclocks;
295 ptp->vclock_index = kzalloc(size, GFP_KERNEL);
296 if (!ptp->vclock_index) {
298 goto no_mem_for_vclocks;
302 err = ptp_populate_pin_groups(ptp);
306 /* Register a new PPS source. */
308 struct pps_source_info pps;
309 memset(&pps, 0, sizeof(pps));
310 snprintf(pps.name, PPS_MAX_NAME_LEN, "ptp%d", index);
311 pps.mode = PTP_PPS_MODE;
312 pps.owner = info->owner;
313 ptp->pps_source = pps_register_source(&pps, PTP_PPS_DEFAULTS);
314 if (IS_ERR(ptp->pps_source)) {
315 err = PTR_ERR(ptp->pps_source);
316 pr_err("failed to register pps source\n");
319 ptp->pps_source->lookup_cookie = ptp;
322 /* Initialize a new device of our class in our clock structure. */
323 device_initialize(&ptp->dev);
324 ptp->dev.devt = ptp->devid;
325 ptp->dev.class = ptp_class;
326 ptp->dev.parent = parent;
327 ptp->dev.groups = ptp->pin_attr_groups;
328 ptp->dev.release = ptp_clock_release;
329 dev_set_drvdata(&ptp->dev, ptp);
330 dev_set_name(&ptp->dev, "ptp%d", ptp->index);
332 /* Create a posix clock and link it to the device. */
333 err = posix_clock_register(&ptp->clock, &ptp->dev);
336 pps_unregister_source(ptp->pps_source);
339 kthread_destroy_worker(ptp->kworker);
341 put_device(&ptp->dev);
343 pr_err("failed to create posix clock\n");
347 /* Debugfs initialization */
348 snprintf(debugfsname, sizeof(debugfsname), "ptp%d", ptp->index);
349 ptp->debugfs_root = debugfs_create_dir(debugfsname, NULL);
354 ptp_cleanup_pin_groups(ptp);
356 kfree(ptp->vclock_index);
359 kthread_destroy_worker(ptp->kworker);
361 mutex_destroy(&ptp->pincfg_mux);
362 mutex_destroy(&ptp->n_vclocks_mux);
363 bitmap_free(queue->mask);
365 list_del(&queue->qlist);
368 ida_free(&ptp_clocks_map, index);
374 EXPORT_SYMBOL(ptp_clock_register);
376 static int unregister_vclock(struct device *dev, void *data)
378 struct ptp_clock *ptp = dev_get_drvdata(dev);
380 ptp_vclock_unregister(info_to_vclock(ptp->info));
384 int ptp_clock_unregister(struct ptp_clock *ptp)
386 if (ptp_vclock_in_use(ptp)) {
387 device_for_each_child(&ptp->dev, NULL, unregister_vclock);
391 wake_up_interruptible(&ptp->tsev_wq);
394 kthread_cancel_delayed_work_sync(&ptp->aux_work);
395 kthread_destroy_worker(ptp->kworker);
398 /* Release the clock's resources. */
400 pps_unregister_source(ptp->pps_source);
402 posix_clock_unregister(&ptp->clock);
406 EXPORT_SYMBOL(ptp_clock_unregister);
408 void ptp_clock_event(struct ptp_clock *ptp, struct ptp_clock_event *event)
410 struct timestamp_event_queue *tsevq;
411 struct pps_event_time evt;
414 switch (event->type) {
416 case PTP_CLOCK_ALARM:
419 case PTP_CLOCK_EXTTS:
420 /* Enqueue timestamp on selected queues */
421 spin_lock_irqsave(&ptp->tsevqs_lock, flags);
422 list_for_each_entry(tsevq, &ptp->tsevqs, qlist) {
423 if (test_bit((unsigned int)event->index, tsevq->mask))
424 enqueue_external_timestamp(tsevq, event);
426 spin_unlock_irqrestore(&ptp->tsevqs_lock, flags);
427 wake_up_interruptible(&ptp->tsev_wq);
432 pps_event(ptp->pps_source, &evt, PTP_PPS_EVENT, NULL);
435 case PTP_CLOCK_PPSUSR:
436 pps_event(ptp->pps_source, &event->pps_times,
437 PTP_PPS_EVENT, NULL);
441 EXPORT_SYMBOL(ptp_clock_event);
443 int ptp_clock_index(struct ptp_clock *ptp)
447 EXPORT_SYMBOL(ptp_clock_index);
449 int ptp_find_pin(struct ptp_clock *ptp,
450 enum ptp_pin_function func, unsigned int chan)
452 struct ptp_pin_desc *pin = NULL;
455 for (i = 0; i < ptp->info->n_pins; i++) {
456 if (ptp->info->pin_config[i].func == func &&
457 ptp->info->pin_config[i].chan == chan) {
458 pin = &ptp->info->pin_config[i];
465 EXPORT_SYMBOL(ptp_find_pin);
467 int ptp_find_pin_unlocked(struct ptp_clock *ptp,
468 enum ptp_pin_function func, unsigned int chan)
472 mutex_lock(&ptp->pincfg_mux);
474 result = ptp_find_pin(ptp, func, chan);
476 mutex_unlock(&ptp->pincfg_mux);
480 EXPORT_SYMBOL(ptp_find_pin_unlocked);
482 int ptp_schedule_worker(struct ptp_clock *ptp, unsigned long delay)
484 return kthread_mod_delayed_work(ptp->kworker, &ptp->aux_work, delay);
486 EXPORT_SYMBOL(ptp_schedule_worker);
488 void ptp_cancel_worker_sync(struct ptp_clock *ptp)
490 kthread_cancel_delayed_work_sync(&ptp->aux_work);
492 EXPORT_SYMBOL(ptp_cancel_worker_sync);
494 /* module operations */
496 static void __exit ptp_exit(void)
498 class_destroy(ptp_class);
499 unregister_chrdev_region(ptp_devt, MINORMASK + 1);
500 ida_destroy(&ptp_clocks_map);
503 static int __init ptp_init(void)
507 ptp_class = class_create("ptp");
508 if (IS_ERR(ptp_class)) {
509 pr_err("ptp: failed to allocate class\n");
510 return PTR_ERR(ptp_class);
513 err = alloc_chrdev_region(&ptp_devt, 0, MINORMASK + 1, "ptp");
515 pr_err("ptp: failed to allocate device region\n");
519 ptp_class->dev_groups = ptp_groups;
520 pr_info("PTP clock support registered\n");
524 class_destroy(ptp_class);
528 subsys_initcall(ptp_init);
529 module_exit(ptp_exit);
531 MODULE_AUTHOR("Richard Cochran <richardcochran@gmail.com>");
532 MODULE_DESCRIPTION("PTP clocks support");
533 MODULE_LICENSE("GPL");