Add the rt linux 4.1.3-rt3 as base
[kvmfornfv.git] / kernel / drivers / w1 / masters / omap_hdq.c
1 /*
2  * drivers/w1/masters/omap_hdq.c
3  *
4  * Copyright (C) 2007,2012 Texas Instruments, Inc.
5  *
6  * This file is licensed under the terms of the GNU General Public License
7  * version 2. This program is licensed "as is" without any warranty of any
8  * kind, whether express or implied.
9  *
10  */
11 #include <linux/kernel.h>
12 #include <linux/module.h>
13 #include <linux/platform_device.h>
14 #include <linux/interrupt.h>
15 #include <linux/slab.h>
16 #include <linux/err.h>
17 #include <linux/io.h>
18 #include <linux/sched.h>
19 #include <linux/pm_runtime.h>
20
21 #include "../w1.h"
22 #include "../w1_int.h"
23
24 #define MOD_NAME        "OMAP_HDQ:"
25
26 #define OMAP_HDQ_REVISION                       0x00
27 #define OMAP_HDQ_TX_DATA                        0x04
28 #define OMAP_HDQ_RX_DATA                        0x08
29 #define OMAP_HDQ_CTRL_STATUS                    0x0c
30 #define OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK      (1<<6)
31 #define OMAP_HDQ_CTRL_STATUS_CLOCKENABLE        (1<<5)
32 #define OMAP_HDQ_CTRL_STATUS_GO                 (1<<4)
33 #define OMAP_HDQ_CTRL_STATUS_INITIALIZATION     (1<<2)
34 #define OMAP_HDQ_CTRL_STATUS_DIR                (1<<1)
35 #define OMAP_HDQ_CTRL_STATUS_MODE               (1<<0)
36 #define OMAP_HDQ_INT_STATUS                     0x10
37 #define OMAP_HDQ_INT_STATUS_TXCOMPLETE          (1<<2)
38 #define OMAP_HDQ_INT_STATUS_RXCOMPLETE          (1<<1)
39 #define OMAP_HDQ_INT_STATUS_TIMEOUT             (1<<0)
40 #define OMAP_HDQ_SYSCONFIG                      0x14
41 #define OMAP_HDQ_SYSCONFIG_SOFTRESET            (1<<1)
42 #define OMAP_HDQ_SYSCONFIG_AUTOIDLE             (1<<0)
43 #define OMAP_HDQ_SYSSTATUS                      0x18
44 #define OMAP_HDQ_SYSSTATUS_RESETDONE            (1<<0)
45
46 #define OMAP_HDQ_FLAG_CLEAR                     0
47 #define OMAP_HDQ_FLAG_SET                       1
48 #define OMAP_HDQ_TIMEOUT                        (HZ/5)
49
50 #define OMAP_HDQ_MAX_USER                       4
51
52 static DECLARE_WAIT_QUEUE_HEAD(hdq_wait_queue);
53 static int w1_id;
54
55 struct hdq_data {
56         struct device           *dev;
57         void __iomem            *hdq_base;
58         /* lock status update */
59         struct  mutex           hdq_mutex;
60         int                     hdq_usecount;
61         u8                      hdq_irqstatus;
62         /* device lock */
63         spinlock_t              hdq_spinlock;
64         /*
65          * Used to control the call to omap_hdq_get and omap_hdq_put.
66          * HDQ Protocol: Write the CMD|REG_address first, followed by
67          * the data wrire or read.
68          */
69         int                     init_trans;
70 };
71
72 static int omap_hdq_probe(struct platform_device *pdev);
73 static int omap_hdq_remove(struct platform_device *pdev);
74
75 static const struct of_device_id omap_hdq_dt_ids[] = {
76         { .compatible = "ti,omap3-1w" },
77         {}
78 };
79 MODULE_DEVICE_TABLE(of, omap_hdq_dt_ids);
80
81 static struct platform_driver omap_hdq_driver = {
82         .probe =        omap_hdq_probe,
83         .remove =       omap_hdq_remove,
84         .driver =       {
85                 .name = "omap_hdq",
86                 .of_match_table = omap_hdq_dt_ids,
87         },
88 };
89
90 static u8 omap_w1_read_byte(void *_hdq);
91 static void omap_w1_write_byte(void *_hdq, u8 byte);
92 static u8 omap_w1_reset_bus(void *_hdq);
93 static void omap_w1_search_bus(void *_hdq, struct w1_master *master_dev,
94                 u8 search_type, w1_slave_found_callback slave_found);
95
96
97 static struct w1_bus_master omap_w1_master = {
98         .read_byte      = omap_w1_read_byte,
99         .write_byte     = omap_w1_write_byte,
100         .reset_bus      = omap_w1_reset_bus,
101         .search         = omap_w1_search_bus,
102 };
103
104 /* HDQ register I/O routines */
105 static inline u8 hdq_reg_in(struct hdq_data *hdq_data, u32 offset)
106 {
107         return __raw_readl(hdq_data->hdq_base + offset);
108 }
109
110 static inline void hdq_reg_out(struct hdq_data *hdq_data, u32 offset, u8 val)
111 {
112         __raw_writel(val, hdq_data->hdq_base + offset);
113 }
114
115 static inline u8 hdq_reg_merge(struct hdq_data *hdq_data, u32 offset,
116                         u8 val, u8 mask)
117 {
118         u8 new_val = (__raw_readl(hdq_data->hdq_base + offset) & ~mask)
119                         | (val & mask);
120         __raw_writel(new_val, hdq_data->hdq_base + offset);
121
122         return new_val;
123 }
124
125 /*
126  * Wait for one or more bits in flag change.
127  * HDQ_FLAG_SET: wait until any bit in the flag is set.
128  * HDQ_FLAG_CLEAR: wait until all bits in the flag are cleared.
129  * return 0 on success and -ETIMEDOUT in the case of timeout.
130  */
131 static int hdq_wait_for_flag(struct hdq_data *hdq_data, u32 offset,
132                 u8 flag, u8 flag_set, u8 *status)
133 {
134         int ret = 0;
135         unsigned long timeout = jiffies + OMAP_HDQ_TIMEOUT;
136
137         if (flag_set == OMAP_HDQ_FLAG_CLEAR) {
138                 /* wait for the flag clear */
139                 while (((*status = hdq_reg_in(hdq_data, offset)) & flag)
140                         && time_before(jiffies, timeout)) {
141                         schedule_timeout_uninterruptible(1);
142                 }
143                 if (*status & flag)
144                         ret = -ETIMEDOUT;
145         } else if (flag_set == OMAP_HDQ_FLAG_SET) {
146                 /* wait for the flag set */
147                 while (!((*status = hdq_reg_in(hdq_data, offset)) & flag)
148                         && time_before(jiffies, timeout)) {
149                         schedule_timeout_uninterruptible(1);
150                 }
151                 if (!(*status & flag))
152                         ret = -ETIMEDOUT;
153         } else
154                 return -EINVAL;
155
156         return ret;
157 }
158
159 /* write out a byte and fill *status with HDQ_INT_STATUS */
160 static int hdq_write_byte(struct hdq_data *hdq_data, u8 val, u8 *status)
161 {
162         int ret;
163         u8 tmp_status;
164         unsigned long irqflags;
165
166         *status = 0;
167
168         spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags);
169         /* clear interrupt flags via a dummy read */
170         hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
171         /* ISR loads it with new INT_STATUS */
172         hdq_data->hdq_irqstatus = 0;
173         spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags);
174
175         hdq_reg_out(hdq_data, OMAP_HDQ_TX_DATA, val);
176
177         /* set the GO bit */
178         hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS, OMAP_HDQ_CTRL_STATUS_GO,
179                 OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_GO);
180         /* wait for the TXCOMPLETE bit */
181         ret = wait_event_timeout(hdq_wait_queue,
182                 hdq_data->hdq_irqstatus, OMAP_HDQ_TIMEOUT);
183         if (ret == 0) {
184                 dev_dbg(hdq_data->dev, "TX wait elapsed\n");
185                 ret = -ETIMEDOUT;
186                 goto out;
187         }
188
189         *status = hdq_data->hdq_irqstatus;
190         /* check irqstatus */
191         if (!(*status & OMAP_HDQ_INT_STATUS_TXCOMPLETE)) {
192                 dev_dbg(hdq_data->dev, "timeout waiting for"
193                         " TXCOMPLETE/RXCOMPLETE, %x", *status);
194                 ret = -ETIMEDOUT;
195                 goto out;
196         }
197
198         /* wait for the GO bit return to zero */
199         ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_CTRL_STATUS,
200                         OMAP_HDQ_CTRL_STATUS_GO,
201                         OMAP_HDQ_FLAG_CLEAR, &tmp_status);
202         if (ret) {
203                 dev_dbg(hdq_data->dev, "timeout waiting GO bit"
204                         " return to zero, %x", tmp_status);
205         }
206
207 out:
208         return ret;
209 }
210
211 /* HDQ Interrupt service routine */
212 static irqreturn_t hdq_isr(int irq, void *_hdq)
213 {
214         struct hdq_data *hdq_data = _hdq;
215         unsigned long irqflags;
216
217         spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags);
218         hdq_data->hdq_irqstatus = hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
219         spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags);
220         dev_dbg(hdq_data->dev, "hdq_isr: %x", hdq_data->hdq_irqstatus);
221
222         if (hdq_data->hdq_irqstatus &
223                 (OMAP_HDQ_INT_STATUS_TXCOMPLETE | OMAP_HDQ_INT_STATUS_RXCOMPLETE
224                 | OMAP_HDQ_INT_STATUS_TIMEOUT)) {
225                 /* wake up sleeping process */
226                 wake_up(&hdq_wait_queue);
227         }
228
229         return IRQ_HANDLED;
230 }
231
232 /* HDQ Mode: always return success */
233 static u8 omap_w1_reset_bus(void *_hdq)
234 {
235         return 0;
236 }
237
238 /* W1 search callback function */
239 static void omap_w1_search_bus(void *_hdq, struct w1_master *master_dev,
240                 u8 search_type, w1_slave_found_callback slave_found)
241 {
242         u64 module_id, rn_le, cs, id;
243
244         if (w1_id)
245                 module_id = w1_id;
246         else
247                 module_id = 0x1;
248
249         rn_le = cpu_to_le64(module_id);
250         /*
251          * HDQ might not obey truly the 1-wire spec.
252          * So calculate CRC based on module parameter.
253          */
254         cs = w1_calc_crc8((u8 *)&rn_le, 7);
255         id = (cs << 56) | module_id;
256
257         slave_found(master_dev, id);
258 }
259
260 static int _omap_hdq_reset(struct hdq_data *hdq_data)
261 {
262         int ret;
263         u8 tmp_status;
264
265         hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG, OMAP_HDQ_SYSCONFIG_SOFTRESET);
266         /*
267          * Select HDQ mode & enable clocks.
268          * It is observed that INT flags can't be cleared via a read and GO/INIT
269          * won't return to zero if interrupt is disabled. So we always enable
270          * interrupt.
271          */
272         hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS,
273                 OMAP_HDQ_CTRL_STATUS_CLOCKENABLE |
274                 OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK);
275
276         /* wait for reset to complete */
277         ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_SYSSTATUS,
278                 OMAP_HDQ_SYSSTATUS_RESETDONE, OMAP_HDQ_FLAG_SET, &tmp_status);
279         if (ret)
280                 dev_dbg(hdq_data->dev, "timeout waiting HDQ reset, %x",
281                                 tmp_status);
282         else {
283                 hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS,
284                         OMAP_HDQ_CTRL_STATUS_CLOCKENABLE |
285                         OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK);
286                 hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG,
287                         OMAP_HDQ_SYSCONFIG_AUTOIDLE);
288         }
289
290         return ret;
291 }
292
293 /* Issue break pulse to the device */
294 static int omap_hdq_break(struct hdq_data *hdq_data)
295 {
296         int ret = 0;
297         u8 tmp_status;
298         unsigned long irqflags;
299
300         ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
301         if (ret < 0) {
302                 dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
303                 ret = -EINTR;
304                 goto rtn;
305         }
306
307         spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags);
308         /* clear interrupt flags via a dummy read */
309         hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
310         /* ISR loads it with new INT_STATUS */
311         hdq_data->hdq_irqstatus = 0;
312         spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags);
313
314         /* set the INIT and GO bit */
315         hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS,
316                 OMAP_HDQ_CTRL_STATUS_INITIALIZATION | OMAP_HDQ_CTRL_STATUS_GO,
317                 OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_INITIALIZATION |
318                 OMAP_HDQ_CTRL_STATUS_GO);
319
320         /* wait for the TIMEOUT bit */
321         ret = wait_event_timeout(hdq_wait_queue,
322                 hdq_data->hdq_irqstatus, OMAP_HDQ_TIMEOUT);
323         if (ret == 0) {
324                 dev_dbg(hdq_data->dev, "break wait elapsed\n");
325                 ret = -EINTR;
326                 goto out;
327         }
328
329         tmp_status = hdq_data->hdq_irqstatus;
330         /* check irqstatus */
331         if (!(tmp_status & OMAP_HDQ_INT_STATUS_TIMEOUT)) {
332                 dev_dbg(hdq_data->dev, "timeout waiting for TIMEOUT, %x",
333                                 tmp_status);
334                 ret = -ETIMEDOUT;
335                 goto out;
336         }
337         /*
338          * wait for both INIT and GO bits rerurn to zero.
339          * zero wait time expected for interrupt mode.
340          */
341         ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_CTRL_STATUS,
342                         OMAP_HDQ_CTRL_STATUS_INITIALIZATION |
343                         OMAP_HDQ_CTRL_STATUS_GO, OMAP_HDQ_FLAG_CLEAR,
344                         &tmp_status);
345         if (ret)
346                 dev_dbg(hdq_data->dev, "timeout waiting INIT&GO bits"
347                         " return to zero, %x", tmp_status);
348
349 out:
350         mutex_unlock(&hdq_data->hdq_mutex);
351 rtn:
352         return ret;
353 }
354
355 static int hdq_read_byte(struct hdq_data *hdq_data, u8 *val)
356 {
357         int ret = 0;
358         u8 status;
359
360         ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
361         if (ret < 0) {
362                 ret = -EINTR;
363                 goto rtn;
364         }
365
366         if (!hdq_data->hdq_usecount) {
367                 ret = -EINVAL;
368                 goto out;
369         }
370
371         if (!(hdq_data->hdq_irqstatus & OMAP_HDQ_INT_STATUS_RXCOMPLETE)) {
372                 hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS,
373                         OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_GO,
374                         OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_GO);
375                 /*
376                  * The RX comes immediately after TX.
377                  */
378                 wait_event_timeout(hdq_wait_queue,
379                                    (hdq_data->hdq_irqstatus
380                                     & OMAP_HDQ_INT_STATUS_RXCOMPLETE),
381                                    OMAP_HDQ_TIMEOUT);
382
383                 hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS, 0,
384                         OMAP_HDQ_CTRL_STATUS_DIR);
385                 status = hdq_data->hdq_irqstatus;
386                 /* check irqstatus */
387                 if (!(status & OMAP_HDQ_INT_STATUS_RXCOMPLETE)) {
388                         dev_dbg(hdq_data->dev, "timeout waiting for"
389                                 " RXCOMPLETE, %x", status);
390                         ret = -ETIMEDOUT;
391                         goto out;
392                 }
393         }
394         /* the data is ready. Read it in! */
395         *val = hdq_reg_in(hdq_data, OMAP_HDQ_RX_DATA);
396 out:
397         mutex_unlock(&hdq_data->hdq_mutex);
398 rtn:
399         return ret;
400
401 }
402
403 /* Enable clocks and set the controller to HDQ mode */
404 static int omap_hdq_get(struct hdq_data *hdq_data)
405 {
406         int ret = 0;
407
408         ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
409         if (ret < 0) {
410                 ret = -EINTR;
411                 goto rtn;
412         }
413
414         if (OMAP_HDQ_MAX_USER == hdq_data->hdq_usecount) {
415                 dev_dbg(hdq_data->dev, "attempt to exceed the max use count");
416                 ret = -EINVAL;
417                 goto out;
418         } else {
419                 hdq_data->hdq_usecount++;
420                 try_module_get(THIS_MODULE);
421                 if (1 == hdq_data->hdq_usecount) {
422
423                         pm_runtime_get_sync(hdq_data->dev);
424
425                         /* make sure HDQ is out of reset */
426                         if (!(hdq_reg_in(hdq_data, OMAP_HDQ_SYSSTATUS) &
427                                 OMAP_HDQ_SYSSTATUS_RESETDONE)) {
428                                 ret = _omap_hdq_reset(hdq_data);
429                                 if (ret)
430                                         /* back up the count */
431                                         hdq_data->hdq_usecount--;
432                         } else {
433                                 /* select HDQ mode & enable clocks */
434                                 hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS,
435                                         OMAP_HDQ_CTRL_STATUS_CLOCKENABLE |
436                                         OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK);
437                                 hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG,
438                                         OMAP_HDQ_SYSCONFIG_AUTOIDLE);
439                                 hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
440                         }
441                 }
442         }
443
444 out:
445         mutex_unlock(&hdq_data->hdq_mutex);
446 rtn:
447         return ret;
448 }
449
450 /* Disable clocks to the module */
451 static int omap_hdq_put(struct hdq_data *hdq_data)
452 {
453         int ret = 0;
454
455         ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
456         if (ret < 0)
457                 return -EINTR;
458
459         if (0 == hdq_data->hdq_usecount) {
460                 dev_dbg(hdq_data->dev, "attempt to decrement use count"
461                         " when it is zero");
462                 ret = -EINVAL;
463         } else {
464                 hdq_data->hdq_usecount--;
465                 module_put(THIS_MODULE);
466                 if (0 == hdq_data->hdq_usecount)
467                         pm_runtime_put_sync(hdq_data->dev);
468         }
469         mutex_unlock(&hdq_data->hdq_mutex);
470
471         return ret;
472 }
473
474 /* Read a byte of data from the device */
475 static u8 omap_w1_read_byte(void *_hdq)
476 {
477         struct hdq_data *hdq_data = _hdq;
478         u8 val = 0;
479         int ret;
480
481         ret = hdq_read_byte(hdq_data, &val);
482         if (ret) {
483                 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
484                 if (ret < 0) {
485                         dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
486                         return -EINTR;
487                 }
488                 hdq_data->init_trans = 0;
489                 mutex_unlock(&hdq_data->hdq_mutex);
490                 omap_hdq_put(hdq_data);
491                 return -1;
492         }
493
494         /* Write followed by a read, release the module */
495         if (hdq_data->init_trans) {
496                 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
497                 if (ret < 0) {
498                         dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
499                         return -EINTR;
500                 }
501                 hdq_data->init_trans = 0;
502                 mutex_unlock(&hdq_data->hdq_mutex);
503                 omap_hdq_put(hdq_data);
504         }
505
506         return val;
507 }
508
509 /* Write a byte of data to the device */
510 static void omap_w1_write_byte(void *_hdq, u8 byte)
511 {
512         struct hdq_data *hdq_data = _hdq;
513         int ret;
514         u8 status;
515
516         /* First write to initialize the transfer */
517         if (hdq_data->init_trans == 0)
518                 omap_hdq_get(hdq_data);
519
520         ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
521         if (ret < 0) {
522                 dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
523                 return;
524         }
525         hdq_data->init_trans++;
526         mutex_unlock(&hdq_data->hdq_mutex);
527
528         ret = hdq_write_byte(hdq_data, byte, &status);
529         if (ret < 0) {
530                 dev_dbg(hdq_data->dev, "TX failure:Ctrl status %x\n", status);
531                 return;
532         }
533
534         /* Second write, data transferred. Release the module */
535         if (hdq_data->init_trans > 1) {
536                 omap_hdq_put(hdq_data);
537                 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
538                 if (ret < 0) {
539                         dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
540                         return;
541                 }
542                 hdq_data->init_trans = 0;
543                 mutex_unlock(&hdq_data->hdq_mutex);
544         }
545 }
546
547 static int omap_hdq_probe(struct platform_device *pdev)
548 {
549         struct device *dev = &pdev->dev;
550         struct hdq_data *hdq_data;
551         struct resource *res;
552         int ret, irq;
553         u8 rev;
554
555         hdq_data = devm_kzalloc(dev, sizeof(*hdq_data), GFP_KERNEL);
556         if (!hdq_data) {
557                 dev_dbg(&pdev->dev, "unable to allocate memory\n");
558                 return -ENOMEM;
559         }
560
561         hdq_data->dev = dev;
562         platform_set_drvdata(pdev, hdq_data);
563
564         res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
565         hdq_data->hdq_base = devm_ioremap_resource(dev, res);
566         if (IS_ERR(hdq_data->hdq_base))
567                 return PTR_ERR(hdq_data->hdq_base);
568
569         hdq_data->hdq_usecount = 0;
570         mutex_init(&hdq_data->hdq_mutex);
571
572         pm_runtime_enable(&pdev->dev);
573         pm_runtime_get_sync(&pdev->dev);
574
575         rev = hdq_reg_in(hdq_data, OMAP_HDQ_REVISION);
576         dev_info(&pdev->dev, "OMAP HDQ Hardware Rev %c.%c. Driver in %s mode\n",
577                 (rev >> 4) + '0', (rev & 0x0f) + '0', "Interrupt");
578
579         spin_lock_init(&hdq_data->hdq_spinlock);
580
581         irq = platform_get_irq(pdev, 0);
582         if (irq < 0) {
583                 ret = -ENXIO;
584                 goto err_irq;
585         }
586
587         ret = devm_request_irq(dev, irq, hdq_isr, 0, "omap_hdq", hdq_data);
588         if (ret < 0) {
589                 dev_dbg(&pdev->dev, "could not request irq\n");
590                 goto err_irq;
591         }
592
593         omap_hdq_break(hdq_data);
594
595         pm_runtime_put_sync(&pdev->dev);
596
597         omap_w1_master.data = hdq_data;
598
599         ret = w1_add_master_device(&omap_w1_master);
600         if (ret) {
601                 dev_dbg(&pdev->dev, "Failure in registering w1 master\n");
602                 goto err_w1;
603         }
604
605         return 0;
606
607 err_irq:
608         pm_runtime_put_sync(&pdev->dev);
609 err_w1:
610         pm_runtime_disable(&pdev->dev);
611
612         return ret;
613 }
614
615 static int omap_hdq_remove(struct platform_device *pdev)
616 {
617         struct hdq_data *hdq_data = platform_get_drvdata(pdev);
618
619         mutex_lock(&hdq_data->hdq_mutex);
620
621         if (hdq_data->hdq_usecount) {
622                 dev_dbg(&pdev->dev, "removed when use count is not zero\n");
623                 mutex_unlock(&hdq_data->hdq_mutex);
624                 return -EBUSY;
625         }
626
627         mutex_unlock(&hdq_data->hdq_mutex);
628
629         /* remove module dependency */
630         pm_runtime_disable(&pdev->dev);
631
632         return 0;
633 }
634
635 module_platform_driver(omap_hdq_driver);
636
637 module_param(w1_id, int, S_IRUSR);
638 MODULE_PARM_DESC(w1_id, "1-wire id for the slave detection");
639
640 MODULE_AUTHOR("Texas Instruments");
641 MODULE_DESCRIPTION("HDQ driver Library");
642 MODULE_LICENSE("GPL");