Orange Pi5 kernel

Deprecated Linux kernel 5.10.110 for OrangePi 5/5B/5+ boards

3 Commits   0 Branches   0 Tags
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   1) // SPDX-License-Identifier: GPL-2.0-or-later
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   3)  *  Copyright (C) 2008 Nokia Corporation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   4)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   5)  *  Based on lirc_serial.c
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   6)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   7) #include <linux/clk.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   8) #include <linux/module.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   9) #include <linux/platform_device.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  10) #include <linux/wait.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  11) #include <linux/pwm.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  12) #include <linux/of.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  13) #include <linux/hrtimer.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  14) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  15) #include <media/rc-core.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  16) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  17) #define WBUF_LEN 256
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  18) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  19) struct ir_rx51 {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  20) 	struct rc_dev *rcdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  21) 	struct pwm_device *pwm;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  22) 	struct hrtimer timer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  23) 	struct device	     *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  24) 	wait_queue_head_t     wqueue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  25) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  26) 	unsigned int	freq;		/* carrier frequency */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  27) 	unsigned int	duty_cycle;	/* carrier duty cycle */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  28) 	int		wbuf[WBUF_LEN];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  29) 	int		wbuf_index;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  30) 	unsigned long	device_is_open;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  31) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  32) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  33) static inline void ir_rx51_on(struct ir_rx51 *ir_rx51)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  34) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  35) 	pwm_enable(ir_rx51->pwm);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  36) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  37) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  38) static inline void ir_rx51_off(struct ir_rx51 *ir_rx51)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  39) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  40) 	pwm_disable(ir_rx51->pwm);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  41) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  42) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  43) static int init_timing_params(struct ir_rx51 *ir_rx51)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  44) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  45) 	struct pwm_device *pwm = ir_rx51->pwm;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  46) 	int duty, period = DIV_ROUND_CLOSEST(NSEC_PER_SEC, ir_rx51->freq);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  47) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  48) 	duty = DIV_ROUND_CLOSEST(ir_rx51->duty_cycle * period, 100);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  49) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  50) 	pwm_config(pwm, duty, period);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  51) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  52) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  53) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  54) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  55) static enum hrtimer_restart ir_rx51_timer_cb(struct hrtimer *timer)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  56) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  57) 	struct ir_rx51 *ir_rx51 = container_of(timer, struct ir_rx51, timer);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  58) 	ktime_t now;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  59) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  60) 	if (ir_rx51->wbuf_index < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  61) 		dev_err_ratelimited(ir_rx51->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  62) 				    "BUG wbuf_index has value of %i\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  63) 				    ir_rx51->wbuf_index);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  64) 		goto end;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  65) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  66) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  67) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  68) 	 * If we happen to hit an odd latency spike, loop through the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  69) 	 * pulses until we catch up.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  70) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  71) 	do {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  72) 		u64 ns;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  73) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  74) 		if (ir_rx51->wbuf_index >= WBUF_LEN)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  75) 			goto end;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  76) 		if (ir_rx51->wbuf[ir_rx51->wbuf_index] == -1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  77) 			goto end;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  78) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  79) 		if (ir_rx51->wbuf_index % 2)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  80) 			ir_rx51_off(ir_rx51);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  81) 		else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  82) 			ir_rx51_on(ir_rx51);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  83) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  84) 		ns = US_TO_NS(ir_rx51->wbuf[ir_rx51->wbuf_index]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  85) 		hrtimer_add_expires_ns(timer, ns);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  86) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  87) 		ir_rx51->wbuf_index++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  88) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  89) 		now = timer->base->get_time();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  90) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  91) 	} while (hrtimer_get_expires_tv64(timer) < now);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  92) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  93) 	return HRTIMER_RESTART;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  94) end:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  95) 	/* Stop TX here */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  96) 	ir_rx51_off(ir_rx51);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  97) 	ir_rx51->wbuf_index = -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  98) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  99) 	wake_up_interruptible(&ir_rx51->wqueue);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) 	return HRTIMER_NORESTART;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104) static int ir_rx51_tx(struct rc_dev *dev, unsigned int *buffer,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) 		      unsigned int count)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) 	struct ir_rx51 *ir_rx51 = dev->priv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) 	if (count > WBUF_LEN)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) 		return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112) 	memcpy(ir_rx51->wbuf, buffer, count * sizeof(unsigned int));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) 	/* Wait any pending transfers to finish */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115) 	wait_event_interruptible(ir_rx51->wqueue, ir_rx51->wbuf_index < 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) 	init_timing_params(ir_rx51);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) 	if (count < WBUF_LEN)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) 		ir_rx51->wbuf[count] = -1; /* Insert termination mark */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) 	 * REVISIT: Adjust latency requirements so the device doesn't go in too
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) 	 * deep sleep states with pm_qos_add_request().
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) 	ir_rx51_on(ir_rx51);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127) 	ir_rx51->wbuf_index = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128) 	hrtimer_start(&ir_rx51->timer,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) 		      ns_to_ktime(US_TO_NS(ir_rx51->wbuf[0])),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) 		      HRTIMER_MODE_REL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) 	 * Don't return back to the userspace until the transfer has
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) 	 * finished
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) 	wait_event_interruptible(ir_rx51->wqueue, ir_rx51->wbuf_index < 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) 	/* REVISIT: Remove pm_qos constraint, we can sleep again */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) 	return count;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) static int ir_rx51_open(struct rc_dev *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) 	struct ir_rx51 *ir_rx51 = dev->priv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146) 	if (test_and_set_bit(1, &ir_rx51->device_is_open))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147) 		return -EBUSY;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) 	ir_rx51->pwm = pwm_get(ir_rx51->dev, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150) 	if (IS_ERR(ir_rx51->pwm)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) 		int res = PTR_ERR(ir_rx51->pwm);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153) 		dev_err(ir_rx51->dev, "pwm_get failed: %d\n", res);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) 		return res;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) static void ir_rx51_release(struct rc_dev *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162) 	struct ir_rx51 *ir_rx51 = dev->priv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164) 	hrtimer_cancel(&ir_rx51->timer);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165) 	ir_rx51_off(ir_rx51);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) 	pwm_put(ir_rx51->pwm);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168) 	clear_bit(1, &ir_rx51->device_is_open);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) static struct ir_rx51 ir_rx51 = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) 	.duty_cycle	= 50,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) 	.wbuf_index	= -1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) static int ir_rx51_set_duty_cycle(struct rc_dev *dev, u32 duty)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) 	struct ir_rx51 *ir_rx51 = dev->priv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180) 	ir_rx51->duty_cycle = duty;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) static int ir_rx51_set_tx_carrier(struct rc_dev *dev, u32 carrier)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) 	struct ir_rx51 *ir_rx51 = dev->priv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) 	if (carrier > 500000 || carrier < 20000)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190) 		return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192) 	ir_rx51->freq = carrier;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197) #ifdef CONFIG_PM
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) static int ir_rx51_suspend(struct platform_device *dev, pm_message_t state)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202) 	 * In case the device is still open, do not suspend. Normally
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203) 	 * this should not be a problem as lircd only keeps the device
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204) 	 * open only for short periods of time. We also don't want to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205) 	 * get involved with race conditions that might happen if we
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206) 	 * were in a middle of a transmit. Thus, we defer any suspend
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207) 	 * actions until transmit has completed.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209) 	if (test_and_set_bit(1, &ir_rx51.device_is_open))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210) 		return -EAGAIN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212) 	clear_bit(1, &ir_rx51.device_is_open);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) static int ir_rx51_resume(struct platform_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224) #define ir_rx51_suspend	NULL
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225) #define ir_rx51_resume	NULL
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227) #endif /* CONFIG_PM */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) static int ir_rx51_probe(struct platform_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) 	struct pwm_device *pwm;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232) 	struct rc_dev *rcdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 233) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 234) 	pwm = pwm_get(&dev->dev, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 235) 	if (IS_ERR(pwm)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 236) 		int err = PTR_ERR(pwm);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 237) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 238) 		if (err != -EPROBE_DEFER)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 239) 			dev_err(&dev->dev, "pwm_get failed: %d\n", err);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 240) 		return err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 241) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 242) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 243) 	/* Use default, in case userspace does not set the carrier */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 244) 	ir_rx51.freq = DIV_ROUND_CLOSEST_ULL(pwm_get_period(pwm), NSEC_PER_SEC);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 245) 	pwm_put(pwm);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 246) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 247) 	hrtimer_init(&ir_rx51.timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 248) 	ir_rx51.timer.function = ir_rx51_timer_cb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 249) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 250) 	ir_rx51.dev = &dev->dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 251) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 252) 	rcdev = devm_rc_allocate_device(&dev->dev, RC_DRIVER_IR_RAW_TX);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 253) 	if (!rcdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 254) 		return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 255) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 256) 	rcdev->priv = &ir_rx51;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 257) 	rcdev->open = ir_rx51_open;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 258) 	rcdev->close = ir_rx51_release;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 259) 	rcdev->tx_ir = ir_rx51_tx;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 260) 	rcdev->s_tx_duty_cycle = ir_rx51_set_duty_cycle;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 261) 	rcdev->s_tx_carrier = ir_rx51_set_tx_carrier;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 262) 	rcdev->driver_name = KBUILD_MODNAME;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 263) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 264) 	ir_rx51.rcdev = rcdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 265) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 266) 	return devm_rc_register_device(&dev->dev, ir_rx51.rcdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 267) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 268) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 269) static int ir_rx51_remove(struct platform_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 270) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 271) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 272) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 273) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 274) static const struct of_device_id ir_rx51_match[] = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 275) 	{
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 276) 		.compatible = "nokia,n900-ir",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 277) 	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 278) 	{},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 279) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 280) MODULE_DEVICE_TABLE(of, ir_rx51_match);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 281) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 282) static struct platform_driver ir_rx51_platform_driver = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 283) 	.probe		= ir_rx51_probe,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 284) 	.remove		= ir_rx51_remove,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 285) 	.suspend	= ir_rx51_suspend,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 286) 	.resume		= ir_rx51_resume,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 287) 	.driver		= {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 288) 		.name	= KBUILD_MODNAME,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 289) 		.of_match_table = of_match_ptr(ir_rx51_match),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 290) 	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 291) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 292) module_platform_driver(ir_rx51_platform_driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 293) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 294) MODULE_DESCRIPTION("IR TX driver for Nokia RX51");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 295) MODULE_AUTHOR("Nokia Corporation");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 296) MODULE_LICENSE("GPL");