xref: /openbmc/linux/drivers/char/tpm/tpm_tis.c (revision e8f6f3b4)
1 /*
2  * Copyright (C) 2005, 2006 IBM Corporation
3  *
4  * Authors:
5  * Leendert van Doorn <leendert@watson.ibm.com>
6  * Kylene Hall <kjhall@us.ibm.com>
7  *
8  * Maintained by: <tpmdd-devel@lists.sourceforge.net>
9  *
10  * Device driver for TCG/TCPA TPM (trusted platform module).
11  * Specifications at www.trustedcomputinggroup.org
12  *
13  * This device driver implements the TPM interface as defined in
14  * the TCG TPM Interface Spec version 1.2, revision 1.0.
15  *
16  * This program is free software; you can redistribute it and/or
17  * modify it under the terms of the GNU General Public License as
18  * published by the Free Software Foundation, version 2 of the
19  * License.
20  */
21 #include <linux/init.h>
22 #include <linux/module.h>
23 #include <linux/moduleparam.h>
24 #include <linux/pnp.h>
25 #include <linux/slab.h>
26 #include <linux/interrupt.h>
27 #include <linux/wait.h>
28 #include <linux/acpi.h>
29 #include <linux/freezer.h>
30 #include "tpm.h"
31 
32 enum tis_access {
33 	TPM_ACCESS_VALID = 0x80,
34 	TPM_ACCESS_ACTIVE_LOCALITY = 0x20,
35 	TPM_ACCESS_REQUEST_PENDING = 0x04,
36 	TPM_ACCESS_REQUEST_USE = 0x02,
37 };
38 
39 enum tis_status {
40 	TPM_STS_VALID = 0x80,
41 	TPM_STS_COMMAND_READY = 0x40,
42 	TPM_STS_GO = 0x20,
43 	TPM_STS_DATA_AVAIL = 0x10,
44 	TPM_STS_DATA_EXPECT = 0x08,
45 };
46 
47 enum tis_int_flags {
48 	TPM_GLOBAL_INT_ENABLE = 0x80000000,
49 	TPM_INTF_BURST_COUNT_STATIC = 0x100,
50 	TPM_INTF_CMD_READY_INT = 0x080,
51 	TPM_INTF_INT_EDGE_FALLING = 0x040,
52 	TPM_INTF_INT_EDGE_RISING = 0x020,
53 	TPM_INTF_INT_LEVEL_LOW = 0x010,
54 	TPM_INTF_INT_LEVEL_HIGH = 0x008,
55 	TPM_INTF_LOCALITY_CHANGE_INT = 0x004,
56 	TPM_INTF_STS_VALID_INT = 0x002,
57 	TPM_INTF_DATA_AVAIL_INT = 0x001,
58 };
59 
60 enum tis_defaults {
61 	TIS_MEM_BASE = 0xFED40000,
62 	TIS_MEM_LEN = 0x5000,
63 	TIS_SHORT_TIMEOUT = 750,	/* ms */
64 	TIS_LONG_TIMEOUT = 2000,	/* 2 sec */
65 };
66 
67 #define	TPM_ACCESS(l)			(0x0000 | ((l) << 12))
68 #define	TPM_INT_ENABLE(l)		(0x0008 | ((l) << 12))
69 #define	TPM_INT_VECTOR(l)		(0x000C | ((l) << 12))
70 #define	TPM_INT_STATUS(l)		(0x0010 | ((l) << 12))
71 #define	TPM_INTF_CAPS(l)		(0x0014 | ((l) << 12))
72 #define	TPM_STS(l)			(0x0018 | ((l) << 12))
73 #define	TPM_DATA_FIFO(l)		(0x0024 | ((l) << 12))
74 
75 #define	TPM_DID_VID(l)			(0x0F00 | ((l) << 12))
76 #define	TPM_RID(l)			(0x0F04 | ((l) << 12))
77 
78 struct priv_data {
79 	bool irq_tested;
80 };
81 
82 static LIST_HEAD(tis_chips);
83 static DEFINE_MUTEX(tis_lock);
84 
85 #if defined(CONFIG_PNP) && defined(CONFIG_ACPI)
86 static int is_itpm(struct pnp_dev *dev)
87 {
88 	struct acpi_device *acpi = pnp_acpi_device(dev);
89 	struct acpi_hardware_id *id;
90 
91 	if (!acpi)
92 		return 0;
93 
94 	list_for_each_entry(id, &acpi->pnp.ids, list) {
95 		if (!strcmp("INTC0102", id->id))
96 			return 1;
97 	}
98 
99 	return 0;
100 }
101 #else
102 static inline int is_itpm(struct pnp_dev *dev)
103 {
104 	return 0;
105 }
106 #endif
107 
108 /* Before we attempt to access the TPM we must see that the valid bit is set.
109  * The specification says that this bit is 0 at reset and remains 0 until the
110  * 'TPM has gone through its self test and initialization and has established
111  * correct values in the other bits.' */
112 static int wait_startup(struct tpm_chip *chip, int l)
113 {
114 	unsigned long stop = jiffies + chip->vendor.timeout_a;
115 	do {
116 		if (ioread8(chip->vendor.iobase + TPM_ACCESS(l)) &
117 		    TPM_ACCESS_VALID)
118 			return 0;
119 		msleep(TPM_TIMEOUT);
120 	} while (time_before(jiffies, stop));
121 	return -1;
122 }
123 
124 static int check_locality(struct tpm_chip *chip, int l)
125 {
126 	if ((ioread8(chip->vendor.iobase + TPM_ACCESS(l)) &
127 	     (TPM_ACCESS_ACTIVE_LOCALITY | TPM_ACCESS_VALID)) ==
128 	    (TPM_ACCESS_ACTIVE_LOCALITY | TPM_ACCESS_VALID))
129 		return chip->vendor.locality = l;
130 
131 	return -1;
132 }
133 
134 static void release_locality(struct tpm_chip *chip, int l, int force)
135 {
136 	if (force || (ioread8(chip->vendor.iobase + TPM_ACCESS(l)) &
137 		      (TPM_ACCESS_REQUEST_PENDING | TPM_ACCESS_VALID)) ==
138 	    (TPM_ACCESS_REQUEST_PENDING | TPM_ACCESS_VALID))
139 		iowrite8(TPM_ACCESS_ACTIVE_LOCALITY,
140 			 chip->vendor.iobase + TPM_ACCESS(l));
141 }
142 
143 static int request_locality(struct tpm_chip *chip, int l)
144 {
145 	unsigned long stop, timeout;
146 	long rc;
147 
148 	if (check_locality(chip, l) >= 0)
149 		return l;
150 
151 	iowrite8(TPM_ACCESS_REQUEST_USE,
152 		 chip->vendor.iobase + TPM_ACCESS(l));
153 
154 	stop = jiffies + chip->vendor.timeout_a;
155 
156 	if (chip->vendor.irq) {
157 again:
158 		timeout = stop - jiffies;
159 		if ((long)timeout <= 0)
160 			return -1;
161 		rc = wait_event_interruptible_timeout(chip->vendor.int_queue,
162 						      (check_locality
163 						       (chip, l) >= 0),
164 						      timeout);
165 		if (rc > 0)
166 			return l;
167 		if (rc == -ERESTARTSYS && freezing(current)) {
168 			clear_thread_flag(TIF_SIGPENDING);
169 			goto again;
170 		}
171 	} else {
172 		/* wait for burstcount */
173 		do {
174 			if (check_locality(chip, l) >= 0)
175 				return l;
176 			msleep(TPM_TIMEOUT);
177 		}
178 		while (time_before(jiffies, stop));
179 	}
180 	return -1;
181 }
182 
183 static u8 tpm_tis_status(struct tpm_chip *chip)
184 {
185 	return ioread8(chip->vendor.iobase +
186 		       TPM_STS(chip->vendor.locality));
187 }
188 
189 static void tpm_tis_ready(struct tpm_chip *chip)
190 {
191 	/* this causes the current command to be aborted */
192 	iowrite8(TPM_STS_COMMAND_READY,
193 		 chip->vendor.iobase + TPM_STS(chip->vendor.locality));
194 }
195 
196 static int get_burstcount(struct tpm_chip *chip)
197 {
198 	unsigned long stop;
199 	int burstcnt;
200 
201 	/* wait for burstcount */
202 	/* which timeout value, spec has 2 answers (c & d) */
203 	stop = jiffies + chip->vendor.timeout_d;
204 	do {
205 		burstcnt = ioread8(chip->vendor.iobase +
206 				   TPM_STS(chip->vendor.locality) + 1);
207 		burstcnt += ioread8(chip->vendor.iobase +
208 				    TPM_STS(chip->vendor.locality) +
209 				    2) << 8;
210 		if (burstcnt)
211 			return burstcnt;
212 		msleep(TPM_TIMEOUT);
213 	} while (time_before(jiffies, stop));
214 	return -EBUSY;
215 }
216 
217 static int recv_data(struct tpm_chip *chip, u8 *buf, size_t count)
218 {
219 	int size = 0, burstcnt;
220 	while (size < count &&
221 	       wait_for_tpm_stat(chip,
222 				 TPM_STS_DATA_AVAIL | TPM_STS_VALID,
223 				 chip->vendor.timeout_c,
224 				 &chip->vendor.read_queue, true)
225 	       == 0) {
226 		burstcnt = get_burstcount(chip);
227 		for (; burstcnt > 0 && size < count; burstcnt--)
228 			buf[size++] = ioread8(chip->vendor.iobase +
229 					      TPM_DATA_FIFO(chip->vendor.
230 							    locality));
231 	}
232 	return size;
233 }
234 
235 static int tpm_tis_recv(struct tpm_chip *chip, u8 *buf, size_t count)
236 {
237 	int size = 0;
238 	int expected, status;
239 
240 	if (count < TPM_HEADER_SIZE) {
241 		size = -EIO;
242 		goto out;
243 	}
244 
245 	/* read first 10 bytes, including tag, paramsize, and result */
246 	if ((size =
247 	     recv_data(chip, buf, TPM_HEADER_SIZE)) < TPM_HEADER_SIZE) {
248 		dev_err(chip->dev, "Unable to read header\n");
249 		goto out;
250 	}
251 
252 	expected = be32_to_cpu(*(__be32 *) (buf + 2));
253 	if (expected > count) {
254 		size = -EIO;
255 		goto out;
256 	}
257 
258 	if ((size +=
259 	     recv_data(chip, &buf[TPM_HEADER_SIZE],
260 		       expected - TPM_HEADER_SIZE)) < expected) {
261 		dev_err(chip->dev, "Unable to read remainder of result\n");
262 		size = -ETIME;
263 		goto out;
264 	}
265 
266 	wait_for_tpm_stat(chip, TPM_STS_VALID, chip->vendor.timeout_c,
267 			  &chip->vendor.int_queue, false);
268 	status = tpm_tis_status(chip);
269 	if (status & TPM_STS_DATA_AVAIL) {	/* retry? */
270 		dev_err(chip->dev, "Error left over data\n");
271 		size = -EIO;
272 		goto out;
273 	}
274 
275 out:
276 	tpm_tis_ready(chip);
277 	release_locality(chip, chip->vendor.locality, 0);
278 	return size;
279 }
280 
281 static bool itpm;
282 module_param(itpm, bool, 0444);
283 MODULE_PARM_DESC(itpm, "Force iTPM workarounds (found on some Lenovo laptops)");
284 
285 /*
286  * If interrupts are used (signaled by an irq set in the vendor structure)
287  * tpm.c can skip polling for the data to be available as the interrupt is
288  * waited for here
289  */
290 static int tpm_tis_send_data(struct tpm_chip *chip, u8 *buf, size_t len)
291 {
292 	int rc, status, burstcnt;
293 	size_t count = 0;
294 
295 	if (request_locality(chip, 0) < 0)
296 		return -EBUSY;
297 
298 	status = tpm_tis_status(chip);
299 	if ((status & TPM_STS_COMMAND_READY) == 0) {
300 		tpm_tis_ready(chip);
301 		if (wait_for_tpm_stat
302 		    (chip, TPM_STS_COMMAND_READY, chip->vendor.timeout_b,
303 		     &chip->vendor.int_queue, false) < 0) {
304 			rc = -ETIME;
305 			goto out_err;
306 		}
307 	}
308 
309 	while (count < len - 1) {
310 		burstcnt = get_burstcount(chip);
311 		for (; burstcnt > 0 && count < len - 1; burstcnt--) {
312 			iowrite8(buf[count], chip->vendor.iobase +
313 				 TPM_DATA_FIFO(chip->vendor.locality));
314 			count++;
315 		}
316 
317 		wait_for_tpm_stat(chip, TPM_STS_VALID, chip->vendor.timeout_c,
318 				  &chip->vendor.int_queue, false);
319 		status = tpm_tis_status(chip);
320 		if (!itpm && (status & TPM_STS_DATA_EXPECT) == 0) {
321 			rc = -EIO;
322 			goto out_err;
323 		}
324 	}
325 
326 	/* write last byte */
327 	iowrite8(buf[count],
328 		 chip->vendor.iobase + TPM_DATA_FIFO(chip->vendor.locality));
329 	wait_for_tpm_stat(chip, TPM_STS_VALID, chip->vendor.timeout_c,
330 			  &chip->vendor.int_queue, false);
331 	status = tpm_tis_status(chip);
332 	if ((status & TPM_STS_DATA_EXPECT) != 0) {
333 		rc = -EIO;
334 		goto out_err;
335 	}
336 
337 	return 0;
338 
339 out_err:
340 	tpm_tis_ready(chip);
341 	release_locality(chip, chip->vendor.locality, 0);
342 	return rc;
343 }
344 
345 static void disable_interrupts(struct tpm_chip *chip)
346 {
347 	u32 intmask;
348 
349 	intmask =
350 	    ioread32(chip->vendor.iobase +
351 		     TPM_INT_ENABLE(chip->vendor.locality));
352 	intmask &= ~TPM_GLOBAL_INT_ENABLE;
353 	iowrite32(intmask,
354 		  chip->vendor.iobase +
355 		  TPM_INT_ENABLE(chip->vendor.locality));
356 	free_irq(chip->vendor.irq, chip);
357 	chip->vendor.irq = 0;
358 }
359 
360 /*
361  * If interrupts are used (signaled by an irq set in the vendor structure)
362  * tpm.c can skip polling for the data to be available as the interrupt is
363  * waited for here
364  */
365 static int tpm_tis_send_main(struct tpm_chip *chip, u8 *buf, size_t len)
366 {
367 	int rc;
368 	u32 ordinal;
369 
370 	rc = tpm_tis_send_data(chip, buf, len);
371 	if (rc < 0)
372 		return rc;
373 
374 	/* go and do it */
375 	iowrite8(TPM_STS_GO,
376 		 chip->vendor.iobase + TPM_STS(chip->vendor.locality));
377 
378 	if (chip->vendor.irq) {
379 		ordinal = be32_to_cpu(*((__be32 *) (buf + 6)));
380 		if (wait_for_tpm_stat
381 		    (chip, TPM_STS_DATA_AVAIL | TPM_STS_VALID,
382 		     tpm_calc_ordinal_duration(chip, ordinal),
383 		     &chip->vendor.read_queue, false) < 0) {
384 			rc = -ETIME;
385 			goto out_err;
386 		}
387 	}
388 	return len;
389 out_err:
390 	tpm_tis_ready(chip);
391 	release_locality(chip, chip->vendor.locality, 0);
392 	return rc;
393 }
394 
395 static int tpm_tis_send(struct tpm_chip *chip, u8 *buf, size_t len)
396 {
397 	int rc, irq;
398 	struct priv_data *priv = chip->vendor.priv;
399 
400 	if (!chip->vendor.irq || priv->irq_tested)
401 		return tpm_tis_send_main(chip, buf, len);
402 
403 	/* Verify receipt of the expected IRQ */
404 	irq = chip->vendor.irq;
405 	chip->vendor.irq = 0;
406 	rc = tpm_tis_send_main(chip, buf, len);
407 	chip->vendor.irq = irq;
408 	if (!priv->irq_tested)
409 		msleep(1);
410 	if (!priv->irq_tested) {
411 		disable_interrupts(chip);
412 		dev_err(chip->dev,
413 			FW_BUG "TPM interrupt not working, polling instead\n");
414 	}
415 	priv->irq_tested = true;
416 	return rc;
417 }
418 
419 struct tis_vendor_timeout_override {
420 	u32 did_vid;
421 	unsigned long timeout_us[4];
422 };
423 
424 static const struct tis_vendor_timeout_override vendor_timeout_overrides[] = {
425 	/* Atmel 3204 */
426 	{ 0x32041114, { (TIS_SHORT_TIMEOUT*1000), (TIS_LONG_TIMEOUT*1000),
427 			(TIS_SHORT_TIMEOUT*1000), (TIS_SHORT_TIMEOUT*1000) } },
428 };
429 
430 static bool tpm_tis_update_timeouts(struct tpm_chip *chip,
431 				    unsigned long *timeout_cap)
432 {
433 	int i;
434 	u32 did_vid;
435 
436 	did_vid = ioread32(chip->vendor.iobase + TPM_DID_VID(0));
437 
438 	for (i = 0; i != ARRAY_SIZE(vendor_timeout_overrides); i++) {
439 		if (vendor_timeout_overrides[i].did_vid != did_vid)
440 			continue;
441 		memcpy(timeout_cap, vendor_timeout_overrides[i].timeout_us,
442 		       sizeof(vendor_timeout_overrides[i].timeout_us));
443 		return true;
444 	}
445 
446 	return false;
447 }
448 
449 /*
450  * Early probing for iTPM with STS_DATA_EXPECT flaw.
451  * Try sending command without itpm flag set and if that
452  * fails, repeat with itpm flag set.
453  */
454 static int probe_itpm(struct tpm_chip *chip)
455 {
456 	int rc = 0;
457 	u8 cmd_getticks[] = {
458 		0x00, 0xc1, 0x00, 0x00, 0x00, 0x0a,
459 		0x00, 0x00, 0x00, 0xf1
460 	};
461 	size_t len = sizeof(cmd_getticks);
462 	bool rem_itpm = itpm;
463 	u16 vendor = ioread16(chip->vendor.iobase + TPM_DID_VID(0));
464 
465 	/* probe only iTPMS */
466 	if (vendor != TPM_VID_INTEL)
467 		return 0;
468 
469 	itpm = false;
470 
471 	rc = tpm_tis_send_data(chip, cmd_getticks, len);
472 	if (rc == 0)
473 		goto out;
474 
475 	tpm_tis_ready(chip);
476 	release_locality(chip, chip->vendor.locality, 0);
477 
478 	itpm = true;
479 
480 	rc = tpm_tis_send_data(chip, cmd_getticks, len);
481 	if (rc == 0) {
482 		dev_info(chip->dev, "Detected an iTPM.\n");
483 		rc = 1;
484 	} else
485 		rc = -EFAULT;
486 
487 out:
488 	itpm = rem_itpm;
489 	tpm_tis_ready(chip);
490 	release_locality(chip, chip->vendor.locality, 0);
491 
492 	return rc;
493 }
494 
495 static bool tpm_tis_req_canceled(struct tpm_chip *chip, u8 status)
496 {
497 	switch (chip->vendor.manufacturer_id) {
498 	case TPM_VID_WINBOND:
499 		return ((status == TPM_STS_VALID) ||
500 			(status == (TPM_STS_VALID | TPM_STS_COMMAND_READY)));
501 	case TPM_VID_STM:
502 		return (status == (TPM_STS_VALID | TPM_STS_COMMAND_READY));
503 	default:
504 		return (status == TPM_STS_COMMAND_READY);
505 	}
506 }
507 
508 static const struct tpm_class_ops tpm_tis = {
509 	.status = tpm_tis_status,
510 	.recv = tpm_tis_recv,
511 	.send = tpm_tis_send,
512 	.cancel = tpm_tis_ready,
513 	.update_timeouts = tpm_tis_update_timeouts,
514 	.req_complete_mask = TPM_STS_DATA_AVAIL | TPM_STS_VALID,
515 	.req_complete_val = TPM_STS_DATA_AVAIL | TPM_STS_VALID,
516 	.req_canceled = tpm_tis_req_canceled,
517 };
518 
519 static irqreturn_t tis_int_probe(int irq, void *dev_id)
520 {
521 	struct tpm_chip *chip = dev_id;
522 	u32 interrupt;
523 
524 	interrupt = ioread32(chip->vendor.iobase +
525 			     TPM_INT_STATUS(chip->vendor.locality));
526 
527 	if (interrupt == 0)
528 		return IRQ_NONE;
529 
530 	chip->vendor.probed_irq = irq;
531 
532 	/* Clear interrupts handled with TPM_EOI */
533 	iowrite32(interrupt,
534 		  chip->vendor.iobase +
535 		  TPM_INT_STATUS(chip->vendor.locality));
536 	return IRQ_HANDLED;
537 }
538 
539 static irqreturn_t tis_int_handler(int dummy, void *dev_id)
540 {
541 	struct tpm_chip *chip = dev_id;
542 	u32 interrupt;
543 	int i;
544 
545 	interrupt = ioread32(chip->vendor.iobase +
546 			     TPM_INT_STATUS(chip->vendor.locality));
547 
548 	if (interrupt == 0)
549 		return IRQ_NONE;
550 
551 	((struct priv_data *)chip->vendor.priv)->irq_tested = true;
552 	if (interrupt & TPM_INTF_DATA_AVAIL_INT)
553 		wake_up_interruptible(&chip->vendor.read_queue);
554 	if (interrupt & TPM_INTF_LOCALITY_CHANGE_INT)
555 		for (i = 0; i < 5; i++)
556 			if (check_locality(chip, i) >= 0)
557 				break;
558 	if (interrupt &
559 	    (TPM_INTF_LOCALITY_CHANGE_INT | TPM_INTF_STS_VALID_INT |
560 	     TPM_INTF_CMD_READY_INT))
561 		wake_up_interruptible(&chip->vendor.int_queue);
562 
563 	/* Clear interrupts handled with TPM_EOI */
564 	iowrite32(interrupt,
565 		  chip->vendor.iobase +
566 		  TPM_INT_STATUS(chip->vendor.locality));
567 	ioread32(chip->vendor.iobase + TPM_INT_STATUS(chip->vendor.locality));
568 	return IRQ_HANDLED;
569 }
570 
571 static bool interrupts = true;
572 module_param(interrupts, bool, 0444);
573 MODULE_PARM_DESC(interrupts, "Enable interrupts");
574 
575 static int tpm_tis_init(struct device *dev, resource_size_t start,
576 			resource_size_t len, unsigned int irq)
577 {
578 	u32 vendor, intfcaps, intmask;
579 	int rc, i, irq_s, irq_e, probe;
580 	struct tpm_chip *chip;
581 	struct priv_data *priv;
582 
583 	priv = devm_kzalloc(dev, sizeof(struct priv_data), GFP_KERNEL);
584 	if (priv == NULL)
585 		return -ENOMEM;
586 	if (!(chip = tpm_register_hardware(dev, &tpm_tis)))
587 		return -ENODEV;
588 	chip->vendor.priv = priv;
589 
590 	chip->vendor.iobase = ioremap(start, len);
591 	if (!chip->vendor.iobase) {
592 		rc = -EIO;
593 		goto out_err;
594 	}
595 
596 	/* Default timeouts */
597 	chip->vendor.timeout_a = msecs_to_jiffies(TIS_SHORT_TIMEOUT);
598 	chip->vendor.timeout_b = msecs_to_jiffies(TIS_LONG_TIMEOUT);
599 	chip->vendor.timeout_c = msecs_to_jiffies(TIS_SHORT_TIMEOUT);
600 	chip->vendor.timeout_d = msecs_to_jiffies(TIS_SHORT_TIMEOUT);
601 
602 	if (wait_startup(chip, 0) != 0) {
603 		rc = -ENODEV;
604 		goto out_err;
605 	}
606 
607 	if (request_locality(chip, 0) != 0) {
608 		rc = -ENODEV;
609 		goto out_err;
610 	}
611 
612 	vendor = ioread32(chip->vendor.iobase + TPM_DID_VID(0));
613 	chip->vendor.manufacturer_id = vendor;
614 
615 	dev_info(dev,
616 		 "1.2 TPM (device-id 0x%X, rev-id %d)\n",
617 		 vendor >> 16, ioread8(chip->vendor.iobase + TPM_RID(0)));
618 
619 	if (!itpm) {
620 		probe = probe_itpm(chip);
621 		if (probe < 0) {
622 			rc = -ENODEV;
623 			goto out_err;
624 		}
625 		itpm = !!probe;
626 	}
627 
628 	if (itpm)
629 		dev_info(dev, "Intel iTPM workaround enabled\n");
630 
631 
632 	/* Figure out the capabilities */
633 	intfcaps =
634 	    ioread32(chip->vendor.iobase +
635 		     TPM_INTF_CAPS(chip->vendor.locality));
636 	dev_dbg(dev, "TPM interface capabilities (0x%x):\n",
637 		intfcaps);
638 	if (intfcaps & TPM_INTF_BURST_COUNT_STATIC)
639 		dev_dbg(dev, "\tBurst Count Static\n");
640 	if (intfcaps & TPM_INTF_CMD_READY_INT)
641 		dev_dbg(dev, "\tCommand Ready Int Support\n");
642 	if (intfcaps & TPM_INTF_INT_EDGE_FALLING)
643 		dev_dbg(dev, "\tInterrupt Edge Falling\n");
644 	if (intfcaps & TPM_INTF_INT_EDGE_RISING)
645 		dev_dbg(dev, "\tInterrupt Edge Rising\n");
646 	if (intfcaps & TPM_INTF_INT_LEVEL_LOW)
647 		dev_dbg(dev, "\tInterrupt Level Low\n");
648 	if (intfcaps & TPM_INTF_INT_LEVEL_HIGH)
649 		dev_dbg(dev, "\tInterrupt Level High\n");
650 	if (intfcaps & TPM_INTF_LOCALITY_CHANGE_INT)
651 		dev_dbg(dev, "\tLocality Change Int Support\n");
652 	if (intfcaps & TPM_INTF_STS_VALID_INT)
653 		dev_dbg(dev, "\tSts Valid Int Support\n");
654 	if (intfcaps & TPM_INTF_DATA_AVAIL_INT)
655 		dev_dbg(dev, "\tData Avail Int Support\n");
656 
657 	/* INTERRUPT Setup */
658 	init_waitqueue_head(&chip->vendor.read_queue);
659 	init_waitqueue_head(&chip->vendor.int_queue);
660 
661 	intmask =
662 	    ioread32(chip->vendor.iobase +
663 		     TPM_INT_ENABLE(chip->vendor.locality));
664 
665 	intmask |= TPM_INTF_CMD_READY_INT
666 	    | TPM_INTF_LOCALITY_CHANGE_INT | TPM_INTF_DATA_AVAIL_INT
667 	    | TPM_INTF_STS_VALID_INT;
668 
669 	iowrite32(intmask,
670 		  chip->vendor.iobase +
671 		  TPM_INT_ENABLE(chip->vendor.locality));
672 	if (interrupts)
673 		chip->vendor.irq = irq;
674 	if (interrupts && !chip->vendor.irq) {
675 		irq_s =
676 		    ioread8(chip->vendor.iobase +
677 			    TPM_INT_VECTOR(chip->vendor.locality));
678 		if (irq_s) {
679 			irq_e = irq_s;
680 		} else {
681 			irq_s = 3;
682 			irq_e = 15;
683 		}
684 
685 		for (i = irq_s; i <= irq_e && chip->vendor.irq == 0; i++) {
686 			iowrite8(i, chip->vendor.iobase +
687 				 TPM_INT_VECTOR(chip->vendor.locality));
688 			if (request_irq
689 			    (i, tis_int_probe, IRQF_SHARED,
690 			     chip->vendor.miscdev.name, chip) != 0) {
691 				dev_info(chip->dev,
692 					 "Unable to request irq: %d for probe\n",
693 					 i);
694 				continue;
695 			}
696 
697 			/* Clear all existing */
698 			iowrite32(ioread32
699 				  (chip->vendor.iobase +
700 				   TPM_INT_STATUS(chip->vendor.locality)),
701 				  chip->vendor.iobase +
702 				  TPM_INT_STATUS(chip->vendor.locality));
703 
704 			/* Turn on */
705 			iowrite32(intmask | TPM_GLOBAL_INT_ENABLE,
706 				  chip->vendor.iobase +
707 				  TPM_INT_ENABLE(chip->vendor.locality));
708 
709 			chip->vendor.probed_irq = 0;
710 
711 			/* Generate Interrupts */
712 			tpm_gen_interrupt(chip);
713 
714 			chip->vendor.irq = chip->vendor.probed_irq;
715 
716 			/* free_irq will call into tis_int_probe;
717 			   clear all irqs we haven't seen while doing
718 			   tpm_gen_interrupt */
719 			iowrite32(ioread32
720 				  (chip->vendor.iobase +
721 				   TPM_INT_STATUS(chip->vendor.locality)),
722 				  chip->vendor.iobase +
723 				  TPM_INT_STATUS(chip->vendor.locality));
724 
725 			/* Turn off */
726 			iowrite32(intmask,
727 				  chip->vendor.iobase +
728 				  TPM_INT_ENABLE(chip->vendor.locality));
729 			free_irq(i, chip);
730 		}
731 	}
732 	if (chip->vendor.irq) {
733 		iowrite8(chip->vendor.irq,
734 			 chip->vendor.iobase +
735 			 TPM_INT_VECTOR(chip->vendor.locality));
736 		if (request_irq
737 		    (chip->vendor.irq, tis_int_handler, IRQF_SHARED,
738 		     chip->vendor.miscdev.name, chip) != 0) {
739 			dev_info(chip->dev,
740 				 "Unable to request irq: %d for use\n",
741 				 chip->vendor.irq);
742 			chip->vendor.irq = 0;
743 		} else {
744 			/* Clear all existing */
745 			iowrite32(ioread32
746 				  (chip->vendor.iobase +
747 				   TPM_INT_STATUS(chip->vendor.locality)),
748 				  chip->vendor.iobase +
749 				  TPM_INT_STATUS(chip->vendor.locality));
750 
751 			/* Turn on */
752 			iowrite32(intmask | TPM_GLOBAL_INT_ENABLE,
753 				  chip->vendor.iobase +
754 				  TPM_INT_ENABLE(chip->vendor.locality));
755 		}
756 	}
757 
758 	if (tpm_get_timeouts(chip)) {
759 		dev_err(dev, "Could not get TPM timeouts and durations\n");
760 		rc = -ENODEV;
761 		goto out_err;
762 	}
763 
764 	if (tpm_do_selftest(chip)) {
765 		dev_err(dev, "TPM self test failed\n");
766 		rc = -ENODEV;
767 		goto out_err;
768 	}
769 
770 	INIT_LIST_HEAD(&chip->vendor.list);
771 	mutex_lock(&tis_lock);
772 	list_add(&chip->vendor.list, &tis_chips);
773 	mutex_unlock(&tis_lock);
774 
775 
776 	return 0;
777 out_err:
778 	if (chip->vendor.iobase)
779 		iounmap(chip->vendor.iobase);
780 	tpm_remove_hardware(chip->dev);
781 	return rc;
782 }
783 
784 #ifdef CONFIG_PM_SLEEP
785 static void tpm_tis_reenable_interrupts(struct tpm_chip *chip)
786 {
787 	u32 intmask;
788 
789 	/* reenable interrupts that device may have lost or
790 	   BIOS/firmware may have disabled */
791 	iowrite8(chip->vendor.irq, chip->vendor.iobase +
792 		 TPM_INT_VECTOR(chip->vendor.locality));
793 
794 	intmask =
795 	    ioread32(chip->vendor.iobase +
796 		     TPM_INT_ENABLE(chip->vendor.locality));
797 
798 	intmask |= TPM_INTF_CMD_READY_INT
799 	    | TPM_INTF_LOCALITY_CHANGE_INT | TPM_INTF_DATA_AVAIL_INT
800 	    | TPM_INTF_STS_VALID_INT | TPM_GLOBAL_INT_ENABLE;
801 
802 	iowrite32(intmask,
803 		  chip->vendor.iobase + TPM_INT_ENABLE(chip->vendor.locality));
804 }
805 
806 static int tpm_tis_resume(struct device *dev)
807 {
808 	struct tpm_chip *chip = dev_get_drvdata(dev);
809 	int ret;
810 
811 	if (chip->vendor.irq)
812 		tpm_tis_reenable_interrupts(chip);
813 
814 	ret = tpm_pm_resume(dev);
815 	if (!ret)
816 		tpm_do_selftest(chip);
817 
818 	return ret;
819 }
820 #endif
821 
822 static SIMPLE_DEV_PM_OPS(tpm_tis_pm, tpm_pm_suspend, tpm_tis_resume);
823 
824 #ifdef CONFIG_PNP
825 static int tpm_tis_pnp_init(struct pnp_dev *pnp_dev,
826 				      const struct pnp_device_id *pnp_id)
827 {
828 	resource_size_t start, len;
829 	unsigned int irq = 0;
830 
831 	start = pnp_mem_start(pnp_dev, 0);
832 	len = pnp_mem_len(pnp_dev, 0);
833 
834 	if (pnp_irq_valid(pnp_dev, 0))
835 		irq = pnp_irq(pnp_dev, 0);
836 	else
837 		interrupts = false;
838 
839 	if (is_itpm(pnp_dev))
840 		itpm = true;
841 
842 	return tpm_tis_init(&pnp_dev->dev, start, len, irq);
843 }
844 
845 static struct pnp_device_id tpm_pnp_tbl[] = {
846 	{"PNP0C31", 0},		/* TPM */
847 	{"ATM1200", 0},		/* Atmel */
848 	{"IFX0102", 0},		/* Infineon */
849 	{"BCM0101", 0},		/* Broadcom */
850 	{"BCM0102", 0},		/* Broadcom */
851 	{"NSC1200", 0},		/* National */
852 	{"ICO0102", 0},		/* Intel */
853 	/* Add new here */
854 	{"", 0},		/* User Specified */
855 	{"", 0}			/* Terminator */
856 };
857 MODULE_DEVICE_TABLE(pnp, tpm_pnp_tbl);
858 
859 static void tpm_tis_pnp_remove(struct pnp_dev *dev)
860 {
861 	struct tpm_chip *chip = pnp_get_drvdata(dev);
862 
863 	tpm_dev_vendor_release(chip);
864 
865 	kfree(chip);
866 }
867 
868 
869 static struct pnp_driver tis_pnp_driver = {
870 	.name = "tpm_tis",
871 	.id_table = tpm_pnp_tbl,
872 	.probe = tpm_tis_pnp_init,
873 	.remove = tpm_tis_pnp_remove,
874 	.driver	= {
875 		.pm = &tpm_tis_pm,
876 	},
877 };
878 
879 #define TIS_HID_USR_IDX sizeof(tpm_pnp_tbl)/sizeof(struct pnp_device_id) -2
880 module_param_string(hid, tpm_pnp_tbl[TIS_HID_USR_IDX].id,
881 		    sizeof(tpm_pnp_tbl[TIS_HID_USR_IDX].id), 0444);
882 MODULE_PARM_DESC(hid, "Set additional specific HID for this driver to probe");
883 #endif
884 
885 static struct platform_driver tis_drv = {
886 	.driver = {
887 		.name = "tpm_tis",
888 		.pm		= &tpm_tis_pm,
889 	},
890 };
891 
892 static struct platform_device *pdev;
893 
894 static bool force;
895 module_param(force, bool, 0444);
896 MODULE_PARM_DESC(force, "Force device probe rather than using ACPI entry");
897 static int __init init_tis(void)
898 {
899 	int rc;
900 #ifdef CONFIG_PNP
901 	if (!force)
902 		return pnp_register_driver(&tis_pnp_driver);
903 #endif
904 
905 	rc = platform_driver_register(&tis_drv);
906 	if (rc < 0)
907 		return rc;
908 	pdev = platform_device_register_simple("tpm_tis", -1, NULL, 0);
909 	if (IS_ERR(pdev)) {
910 		rc = PTR_ERR(pdev);
911 		goto err_dev;
912 	}
913 	rc = tpm_tis_init(&pdev->dev, TIS_MEM_BASE, TIS_MEM_LEN, 0);
914 	if (rc)
915 		goto err_init;
916 	return 0;
917 err_init:
918 	platform_device_unregister(pdev);
919 err_dev:
920 	platform_driver_unregister(&tis_drv);
921 	return rc;
922 }
923 
924 static void __exit cleanup_tis(void)
925 {
926 	struct tpm_vendor_specific *i, *j;
927 	struct tpm_chip *chip;
928 	mutex_lock(&tis_lock);
929 	list_for_each_entry_safe(i, j, &tis_chips, list) {
930 		chip = to_tpm_chip(i);
931 		tpm_remove_hardware(chip->dev);
932 		iowrite32(~TPM_GLOBAL_INT_ENABLE &
933 			  ioread32(chip->vendor.iobase +
934 				   TPM_INT_ENABLE(chip->vendor.
935 						  locality)),
936 			  chip->vendor.iobase +
937 			  TPM_INT_ENABLE(chip->vendor.locality));
938 		release_locality(chip, chip->vendor.locality, 1);
939 		if (chip->vendor.irq)
940 			free_irq(chip->vendor.irq, chip);
941 		iounmap(i->iobase);
942 		list_del(&i->list);
943 	}
944 	mutex_unlock(&tis_lock);
945 #ifdef CONFIG_PNP
946 	if (!force) {
947 		pnp_unregister_driver(&tis_pnp_driver);
948 		return;
949 	}
950 #endif
951 	platform_device_unregister(pdev);
952 	platform_driver_unregister(&tis_drv);
953 }
954 
955 module_init(init_tis);
956 module_exit(cleanup_tis);
957 MODULE_AUTHOR("Leendert van Doorn (leendert@watson.ibm.com)");
958 MODULE_DESCRIPTION("TPM Driver");
959 MODULE_VERSION("2.0");
960 MODULE_LICENSE("GPL");
961