^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) // SPDX-License-Identifier: GPL-2.0-only
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) * Implementation of the Xen vTPM device frontend
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) * Author: Daniel De Graaf <dgdegra@tycho.nsa.gov>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) #include <linux/errno.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) #include <linux/err.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) #include <linux/interrupt.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10) #include <linux/freezer.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) #include <xen/xen.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12) #include <xen/events.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) #include <xen/interface/io/tpmif.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14) #include <xen/grant_table.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15) #include <xen/xenbus.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) #include <xen/page.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17) #include "tpm.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) #include <xen/platform_pci.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20) struct tpm_private {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21) struct tpm_chip *chip;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22) struct xenbus_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) struct vtpm_shared_page *shr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26) unsigned int evtchn;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27) int ring_ref;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) domid_t backend_id;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) int irq;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30) wait_queue_head_t read_queue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) enum status_bits {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34) VTPM_STATUS_RUNNING = 0x1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35) VTPM_STATUS_IDLE = 0x2,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36) VTPM_STATUS_RESULT = 0x4,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) VTPM_STATUS_CANCELED = 0x8,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40) static bool wait_for_tpm_stat_cond(struct tpm_chip *chip, u8 mask,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41) bool check_cancel, bool *canceled)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43) u8 status = chip->ops->status(chip);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45) *canceled = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) if ((status & mask) == mask)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47) return true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) if (check_cancel && chip->ops->req_canceled(chip, status)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) *canceled = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50) return true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52) return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55) static int wait_for_tpm_stat(struct tpm_chip *chip, u8 mask,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) unsigned long timeout, wait_queue_head_t *queue,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) bool check_cancel)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) unsigned long stop;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60) long rc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61) u8 status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62) bool canceled = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64) /* check current status */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65) status = chip->ops->status(chip);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66) if ((status & mask) == mask)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69) stop = jiffies + timeout;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) if (chip->flags & TPM_CHIP_FLAG_IRQ) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) again:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73) timeout = stop - jiffies;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) if ((long)timeout <= 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75) return -ETIME;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76) rc = wait_event_interruptible_timeout(*queue,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77) wait_for_tpm_stat_cond(chip, mask, check_cancel,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78) &canceled),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79) timeout);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80) if (rc > 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81) if (canceled)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82) return -ECANCELED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85) if (rc == -ERESTARTSYS && freezing(current)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86) clear_thread_flag(TIF_SIGPENDING);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87) goto again;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89) } else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90) do {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91) tpm_msleep(TPM_TIMEOUT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) status = chip->ops->status(chip);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93) if ((status & mask) == mask)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95) } while (time_before(jiffies, stop));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97) return -ETIME;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100) static u8 vtpm_status(struct tpm_chip *chip)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) struct tpm_private *priv = dev_get_drvdata(&chip->dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) switch (priv->shr->state) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104) case VTPM_STATE_IDLE:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) return VTPM_STATUS_IDLE | VTPM_STATUS_CANCELED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106) case VTPM_STATE_FINISH:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) return VTPM_STATUS_IDLE | VTPM_STATUS_RESULT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) case VTPM_STATE_SUBMIT:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) case VTPM_STATE_CANCEL: /* cancel requested, not yet canceled */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) return VTPM_STATUS_RUNNING;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) static bool vtpm_req_canceled(struct tpm_chip *chip, u8 status)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) return status & VTPM_STATUS_CANCELED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) static void vtpm_cancel(struct tpm_chip *chip)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) struct tpm_private *priv = dev_get_drvdata(&chip->dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124) priv->shr->state = VTPM_STATE_CANCEL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) wmb();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) notify_remote_via_evtchn(priv->evtchn);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) static unsigned int shr_data_offset(struct vtpm_shared_page *shr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131) return sizeof(*shr) + sizeof(u32) * shr->nr_extra_pages;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134) static int vtpm_send(struct tpm_chip *chip, u8 *buf, size_t count)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) struct tpm_private *priv = dev_get_drvdata(&chip->dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) struct vtpm_shared_page *shr = priv->shr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) unsigned int offset = shr_data_offset(shr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140) u32 ordinal;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) unsigned long duration;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) if (offset > PAGE_SIZE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146) if (offset + count > PAGE_SIZE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147) return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) /* Wait for completion of any existing command or cancellation */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150) if (wait_for_tpm_stat(chip, VTPM_STATUS_IDLE, chip->timeout_c,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) &priv->read_queue, true) < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) vtpm_cancel(chip);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153) return -ETIME;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156) memcpy(offset + (u8 *)shr, buf, count);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) shr->length = count;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) barrier();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) shr->state = VTPM_STATE_SUBMIT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) wmb();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) notify_remote_via_evtchn(priv->evtchn);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) ordinal = be32_to_cpu(((struct tpm_header *)buf)->ordinal);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164) duration = tpm_calc_ordinal_duration(chip, ordinal);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) if (wait_for_tpm_stat(chip, VTPM_STATUS_IDLE, duration,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) &priv->read_queue, true) < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168) /* got a signal or timeout, try to cancel */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) vtpm_cancel(chip);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170) return -ETIME;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) static int vtpm_recv(struct tpm_chip *chip, u8 *buf, size_t count)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) struct tpm_private *priv = dev_get_drvdata(&chip->dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) struct vtpm_shared_page *shr = priv->shr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180) unsigned int offset = shr_data_offset(shr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) size_t length = shr->length;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) if (shr->state == VTPM_STATE_IDLE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184) return -ECANCELED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) /* In theory the wait at the end of _send makes this one unnecessary */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) if (wait_for_tpm_stat(chip, VTPM_STATUS_RESULT, chip->timeout_c,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) &priv->read_queue, true) < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) vtpm_cancel(chip);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190) return -ETIME;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) if (offset > PAGE_SIZE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194) return -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196) if (offset + length > PAGE_SIZE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197) length = PAGE_SIZE - offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) if (length > count)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200) length = count;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202) memcpy(buf, offset + (u8 *)shr, length);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204) return length;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207) static const struct tpm_class_ops tpm_vtpm = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208) .status = vtpm_status,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209) .recv = vtpm_recv,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210) .send = vtpm_send,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) .cancel = vtpm_cancel,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212) .req_complete_mask = VTPM_STATUS_IDLE | VTPM_STATUS_RESULT,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213) .req_complete_val = VTPM_STATUS_IDLE | VTPM_STATUS_RESULT,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214) .req_canceled = vtpm_req_canceled,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) static irqreturn_t tpmif_interrupt(int dummy, void *dev_id)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) struct tpm_private *priv = dev_id;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) switch (priv->shr->state) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222) case VTPM_STATE_IDLE:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223) case VTPM_STATE_FINISH:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224) wake_up_interruptible(&priv->read_queue);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226) case VTPM_STATE_SUBMIT:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227) case VTPM_STATE_CANCEL:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228) default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) return IRQ_HANDLED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 233)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 234) static int setup_chip(struct device *dev, struct tpm_private *priv)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 235) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 236) struct tpm_chip *chip;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 237)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 238) chip = tpmm_chip_alloc(dev, &tpm_vtpm);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 239) if (IS_ERR(chip))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 240) return PTR_ERR(chip);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 241)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 242) init_waitqueue_head(&priv->read_queue);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 243)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 244) priv->chip = chip;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 245) dev_set_drvdata(&chip->dev, priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 246)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 247) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 248) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 249)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 250) /* caller must clean up in case of errors */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 251) static int setup_ring(struct xenbus_device *dev, struct tpm_private *priv)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 252) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 253) struct xenbus_transaction xbt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 254) const char *message = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 255) int rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 256) grant_ref_t gref;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 257)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 258) priv->shr = (void *)__get_free_page(GFP_KERNEL|__GFP_ZERO);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 259) if (!priv->shr) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 260) xenbus_dev_fatal(dev, -ENOMEM, "allocating shared ring");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 261) return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 262) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 263)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 264) rv = xenbus_grant_ring(dev, priv->shr, 1, &gref);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 265) if (rv < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 266) return rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 267)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 268) priv->ring_ref = gref;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 269)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 270) rv = xenbus_alloc_evtchn(dev, &priv->evtchn);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 271) if (rv)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 272) return rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 273)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 274) rv = bind_evtchn_to_irqhandler(priv->evtchn, tpmif_interrupt, 0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 275) "tpmif", priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 276) if (rv <= 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 277) xenbus_dev_fatal(dev, rv, "allocating TPM irq");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 278) return rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 279) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 280) priv->irq = rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 281)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 282) again:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 283) rv = xenbus_transaction_start(&xbt);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 284) if (rv) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 285) xenbus_dev_fatal(dev, rv, "starting transaction");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 286) return rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 287) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 288)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 289) rv = xenbus_printf(xbt, dev->nodename,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 290) "ring-ref", "%u", priv->ring_ref);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 291) if (rv) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 292) message = "writing ring-ref";
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 293) goto abort_transaction;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 294) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 295)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 296) rv = xenbus_printf(xbt, dev->nodename, "event-channel", "%u",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 297) priv->evtchn);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 298) if (rv) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 299) message = "writing event-channel";
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 300) goto abort_transaction;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 301) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 302)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 303) rv = xenbus_printf(xbt, dev->nodename, "feature-protocol-v2", "1");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 304) if (rv) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 305) message = "writing feature-protocol-v2";
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 306) goto abort_transaction;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 307) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 308)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 309) rv = xenbus_transaction_end(xbt, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 310) if (rv == -EAGAIN)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 311) goto again;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 312) if (rv) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 313) xenbus_dev_fatal(dev, rv, "completing transaction");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 314) return rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 315) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 316)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 317) xenbus_switch_state(dev, XenbusStateInitialised);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 318)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 319) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 320)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 321) abort_transaction:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 322) xenbus_transaction_end(xbt, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 323) if (message)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 324) xenbus_dev_error(dev, rv, "%s", message);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 325)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 326) return rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 327) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 328)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 329) static void ring_free(struct tpm_private *priv)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 330) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 331) if (!priv)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 332) return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 333)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 334) if (priv->ring_ref)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 335) gnttab_end_foreign_access(priv->ring_ref, 0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 336) (unsigned long)priv->shr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 337) else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 338) free_page((unsigned long)priv->shr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 339)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 340) if (priv->irq)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 341) unbind_from_irqhandler(priv->irq, priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 342)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 343) kfree(priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 344) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 345)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 346) static int tpmfront_probe(struct xenbus_device *dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 347) const struct xenbus_device_id *id)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 348) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 349) struct tpm_private *priv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 350) int rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 351)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 352) priv = kzalloc(sizeof(*priv), GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 353) if (!priv) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 354) xenbus_dev_fatal(dev, -ENOMEM, "allocating priv structure");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 355) return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 356) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 357)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 358) rv = setup_chip(&dev->dev, priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 359) if (rv) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 360) kfree(priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 361) return rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 362) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 363)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 364) rv = setup_ring(dev, priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 365) if (rv) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 366) ring_free(priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 367) return rv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 368) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 369)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 370) tpm_get_timeouts(priv->chip);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 371)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 372) return tpm_chip_register(priv->chip);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 373) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 374)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 375) static int tpmfront_remove(struct xenbus_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 376) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 377) struct tpm_chip *chip = dev_get_drvdata(&dev->dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 378) struct tpm_private *priv = dev_get_drvdata(&chip->dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 379) tpm_chip_unregister(chip);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 380) ring_free(priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 381) dev_set_drvdata(&chip->dev, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 382) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 383) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 384)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 385) static int tpmfront_resume(struct xenbus_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 386) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 387) /* A suspend/resume/migrate will interrupt a vTPM anyway */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 388) tpmfront_remove(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 389) return tpmfront_probe(dev, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 390) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 391)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 392) static void backend_changed(struct xenbus_device *dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 393) enum xenbus_state backend_state)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 394) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 395) switch (backend_state) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 396) case XenbusStateInitialised:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 397) case XenbusStateConnected:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 398) if (dev->state == XenbusStateConnected)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 399) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 400)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 401) if (!xenbus_read_unsigned(dev->otherend, "feature-protocol-v2",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 402) 0)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 403) xenbus_dev_fatal(dev, -EINVAL,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 404) "vTPM protocol 2 required");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 405) return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 406) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 407) xenbus_switch_state(dev, XenbusStateConnected);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 408) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 409)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 410) case XenbusStateClosing:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 411) case XenbusStateClosed:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 412) device_unregister(&dev->dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 413) xenbus_frontend_closed(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 414) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 415) default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 416) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 417) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 418) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 419)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 420) static const struct xenbus_device_id tpmfront_ids[] = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 421) { "vtpm" },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 422) { "" }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 423) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 424) MODULE_ALIAS("xen:vtpm");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 425)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 426) static struct xenbus_driver tpmfront_driver = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 427) .ids = tpmfront_ids,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 428) .probe = tpmfront_probe,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 429) .remove = tpmfront_remove,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 430) .resume = tpmfront_resume,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 431) .otherend_changed = backend_changed,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 432) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 433)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 434) static int __init xen_tpmfront_init(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 435) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 436) if (!xen_domain())
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 437) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 438)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 439) if (!xen_has_pv_devices())
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 440) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 441)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 442) return xenbus_register_frontend(&tpmfront_driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 443) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 444) module_init(xen_tpmfront_init);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 445)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 446) static void __exit xen_tpmfront_exit(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 447) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 448) xenbus_unregister_driver(&tpmfront_driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 449) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 450) module_exit(xen_tpmfront_exit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 451)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 452) MODULE_AUTHOR("Daniel De Graaf <dgdegra@tycho.nsa.gov>");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 453) MODULE_DESCRIPTION("Xen vTPM Driver");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 454) MODULE_LICENSE("GPL");