1 | // SPDX-License-Identifier: GPL-2.0-only |
2 | /* |
3 | * Copyright (C) 2012 Avionic Design GmbH |
4 | * Copyright (C) 2012-2013, NVIDIA Corporation |
5 | */ |
6 | |
7 | #include <linux/debugfs.h> |
8 | #include <linux/dma-mapping.h> |
9 | #include <linux/host1x.h> |
10 | #include <linux/of.h> |
11 | #include <linux/seq_file.h> |
12 | #include <linux/slab.h> |
13 | #include <linux/of_device.h> |
14 | |
15 | #include "bus.h" |
16 | #include "dev.h" |
17 | |
18 | static DEFINE_MUTEX(clients_lock); |
19 | static LIST_HEAD(clients); |
20 | |
21 | static DEFINE_MUTEX(drivers_lock); |
22 | static LIST_HEAD(drivers); |
23 | |
24 | static DEFINE_MUTEX(devices_lock); |
25 | static LIST_HEAD(devices); |
26 | |
27 | struct host1x_subdev { |
28 | struct host1x_client *client; |
29 | struct device_node *np; |
30 | struct list_head list; |
31 | }; |
32 | |
33 | /** |
34 | * host1x_subdev_add() - add a new subdevice with an associated device node |
35 | * @device: host1x device to add the subdevice to |
36 | * @driver: host1x driver containing the subdevices |
37 | * @np: device node |
38 | */ |
39 | static int host1x_subdev_add(struct host1x_device *device, |
40 | struct host1x_driver *driver, |
41 | struct device_node *np) |
42 | { |
43 | struct host1x_subdev *subdev; |
44 | struct device_node *child; |
45 | int err; |
46 | |
47 | subdev = kzalloc(size: sizeof(*subdev), GFP_KERNEL); |
48 | if (!subdev) |
49 | return -ENOMEM; |
50 | |
51 | INIT_LIST_HEAD(list: &subdev->list); |
52 | subdev->np = of_node_get(node: np); |
53 | |
54 | mutex_lock(&device->subdevs_lock); |
55 | list_add_tail(new: &subdev->list, head: &device->subdevs); |
56 | mutex_unlock(lock: &device->subdevs_lock); |
57 | |
58 | /* recursively add children */ |
59 | for_each_child_of_node(np, child) { |
60 | if (of_match_node(matches: driver->subdevs, node: child) && |
61 | of_device_is_available(device: child)) { |
62 | err = host1x_subdev_add(device, driver, np: child); |
63 | if (err < 0) { |
64 | /* XXX cleanup? */ |
65 | of_node_put(node: child); |
66 | return err; |
67 | } |
68 | } |
69 | } |
70 | |
71 | return 0; |
72 | } |
73 | |
74 | /** |
75 | * host1x_subdev_del() - remove subdevice |
76 | * @subdev: subdevice to remove |
77 | */ |
78 | static void host1x_subdev_del(struct host1x_subdev *subdev) |
79 | { |
80 | list_del(entry: &subdev->list); |
81 | of_node_put(node: subdev->np); |
82 | kfree(objp: subdev); |
83 | } |
84 | |
85 | /** |
86 | * host1x_device_parse_dt() - scan device tree and add matching subdevices |
87 | * @device: host1x logical device |
88 | * @driver: host1x driver |
89 | */ |
90 | static int host1x_device_parse_dt(struct host1x_device *device, |
91 | struct host1x_driver *driver) |
92 | { |
93 | struct device_node *np; |
94 | int err; |
95 | |
96 | for_each_child_of_node(device->dev.parent->of_node, np) { |
97 | if (of_match_node(matches: driver->subdevs, node: np) && |
98 | of_device_is_available(device: np)) { |
99 | err = host1x_subdev_add(device, driver, np); |
100 | if (err < 0) { |
101 | of_node_put(node: np); |
102 | return err; |
103 | } |
104 | } |
105 | } |
106 | |
107 | return 0; |
108 | } |
109 | |
110 | static void host1x_subdev_register(struct host1x_device *device, |
111 | struct host1x_subdev *subdev, |
112 | struct host1x_client *client) |
113 | { |
114 | int err; |
115 | |
116 | /* |
117 | * Move the subdevice to the list of active (registered) subdevices |
118 | * and associate it with a client. At the same time, associate the |
119 | * client with its parent device. |
120 | */ |
121 | mutex_lock(&device->subdevs_lock); |
122 | mutex_lock(&device->clients_lock); |
123 | list_move_tail(list: &client->list, head: &device->clients); |
124 | list_move_tail(list: &subdev->list, head: &device->active); |
125 | client->host = &device->dev; |
126 | subdev->client = client; |
127 | mutex_unlock(lock: &device->clients_lock); |
128 | mutex_unlock(lock: &device->subdevs_lock); |
129 | |
130 | if (list_empty(head: &device->subdevs)) { |
131 | err = device_add(dev: &device->dev); |
132 | if (err < 0) |
133 | dev_err(&device->dev, "failed to add: %d\n" , err); |
134 | else |
135 | device->registered = true; |
136 | } |
137 | } |
138 | |
139 | static void __host1x_subdev_unregister(struct host1x_device *device, |
140 | struct host1x_subdev *subdev) |
141 | { |
142 | struct host1x_client *client = subdev->client; |
143 | |
144 | /* |
145 | * If all subdevices have been activated, we're about to remove the |
146 | * first active subdevice, so unload the driver first. |
147 | */ |
148 | if (list_empty(head: &device->subdevs)) { |
149 | if (device->registered) { |
150 | device->registered = false; |
151 | device_del(dev: &device->dev); |
152 | } |
153 | } |
154 | |
155 | /* |
156 | * Move the subdevice back to the list of idle subdevices and remove |
157 | * it from list of clients. |
158 | */ |
159 | mutex_lock(&device->clients_lock); |
160 | subdev->client = NULL; |
161 | client->host = NULL; |
162 | list_move_tail(list: &subdev->list, head: &device->subdevs); |
163 | /* |
164 | * XXX: Perhaps don't do this here, but rather explicitly remove it |
165 | * when the device is about to be deleted. |
166 | * |
167 | * This is somewhat complicated by the fact that this function is |
168 | * used to remove the subdevice when a client is unregistered but |
169 | * also when the composite device is about to be removed. |
170 | */ |
171 | list_del_init(entry: &client->list); |
172 | mutex_unlock(lock: &device->clients_lock); |
173 | } |
174 | |
175 | static void host1x_subdev_unregister(struct host1x_device *device, |
176 | struct host1x_subdev *subdev) |
177 | { |
178 | mutex_lock(&device->subdevs_lock); |
179 | __host1x_subdev_unregister(device, subdev); |
180 | mutex_unlock(lock: &device->subdevs_lock); |
181 | } |
182 | |
183 | /** |
184 | * host1x_device_init() - initialize a host1x logical device |
185 | * @device: host1x logical device |
186 | * |
187 | * The driver for the host1x logical device can call this during execution of |
188 | * its &host1x_driver.probe implementation to initialize each of its clients. |
189 | * The client drivers access the subsystem specific driver data using the |
190 | * &host1x_client.parent field and driver data associated with it (usually by |
191 | * calling dev_get_drvdata()). |
192 | */ |
193 | int host1x_device_init(struct host1x_device *device) |
194 | { |
195 | struct host1x_client *client; |
196 | int err; |
197 | |
198 | mutex_lock(&device->clients_lock); |
199 | |
200 | list_for_each_entry(client, &device->clients, list) { |
201 | if (client->ops && client->ops->early_init) { |
202 | err = client->ops->early_init(client); |
203 | if (err < 0) { |
204 | dev_err(&device->dev, "failed to early initialize %s: %d\n" , |
205 | dev_name(client->dev), err); |
206 | goto teardown_late; |
207 | } |
208 | } |
209 | } |
210 | |
211 | list_for_each_entry(client, &device->clients, list) { |
212 | if (client->ops && client->ops->init) { |
213 | err = client->ops->init(client); |
214 | if (err < 0) { |
215 | dev_err(&device->dev, |
216 | "failed to initialize %s: %d\n" , |
217 | dev_name(client->dev), err); |
218 | goto teardown; |
219 | } |
220 | } |
221 | } |
222 | |
223 | mutex_unlock(lock: &device->clients_lock); |
224 | |
225 | return 0; |
226 | |
227 | teardown: |
228 | list_for_each_entry_continue_reverse(client, &device->clients, list) |
229 | if (client->ops->exit) |
230 | client->ops->exit(client); |
231 | |
232 | /* reset client to end of list for late teardown */ |
233 | client = list_entry(&device->clients, struct host1x_client, list); |
234 | |
235 | teardown_late: |
236 | list_for_each_entry_continue_reverse(client, &device->clients, list) |
237 | if (client->ops->late_exit) |
238 | client->ops->late_exit(client); |
239 | |
240 | mutex_unlock(lock: &device->clients_lock); |
241 | return err; |
242 | } |
243 | EXPORT_SYMBOL(host1x_device_init); |
244 | |
245 | /** |
246 | * host1x_device_exit() - uninitialize host1x logical device |
247 | * @device: host1x logical device |
248 | * |
249 | * When the driver for a host1x logical device is unloaded, it can call this |
250 | * function to tear down each of its clients. Typically this is done after a |
251 | * subsystem-specific data structure is removed and the functionality can no |
252 | * longer be used. |
253 | */ |
254 | int host1x_device_exit(struct host1x_device *device) |
255 | { |
256 | struct host1x_client *client; |
257 | int err; |
258 | |
259 | mutex_lock(&device->clients_lock); |
260 | |
261 | list_for_each_entry_reverse(client, &device->clients, list) { |
262 | if (client->ops && client->ops->exit) { |
263 | err = client->ops->exit(client); |
264 | if (err < 0) { |
265 | dev_err(&device->dev, |
266 | "failed to cleanup %s: %d\n" , |
267 | dev_name(client->dev), err); |
268 | mutex_unlock(lock: &device->clients_lock); |
269 | return err; |
270 | } |
271 | } |
272 | } |
273 | |
274 | list_for_each_entry_reverse(client, &device->clients, list) { |
275 | if (client->ops && client->ops->late_exit) { |
276 | err = client->ops->late_exit(client); |
277 | if (err < 0) { |
278 | dev_err(&device->dev, "failed to late cleanup %s: %d\n" , |
279 | dev_name(client->dev), err); |
280 | mutex_unlock(lock: &device->clients_lock); |
281 | return err; |
282 | } |
283 | } |
284 | } |
285 | |
286 | mutex_unlock(lock: &device->clients_lock); |
287 | |
288 | return 0; |
289 | } |
290 | EXPORT_SYMBOL(host1x_device_exit); |
291 | |
292 | static int host1x_add_client(struct host1x *host1x, |
293 | struct host1x_client *client) |
294 | { |
295 | struct host1x_device *device; |
296 | struct host1x_subdev *subdev; |
297 | |
298 | mutex_lock(&host1x->devices_lock); |
299 | |
300 | list_for_each_entry(device, &host1x->devices, list) { |
301 | list_for_each_entry(subdev, &device->subdevs, list) { |
302 | if (subdev->np == client->dev->of_node) { |
303 | host1x_subdev_register(device, subdev, client); |
304 | mutex_unlock(lock: &host1x->devices_lock); |
305 | return 0; |
306 | } |
307 | } |
308 | } |
309 | |
310 | mutex_unlock(lock: &host1x->devices_lock); |
311 | return -ENODEV; |
312 | } |
313 | |
314 | static int host1x_del_client(struct host1x *host1x, |
315 | struct host1x_client *client) |
316 | { |
317 | struct host1x_device *device, *dt; |
318 | struct host1x_subdev *subdev; |
319 | |
320 | mutex_lock(&host1x->devices_lock); |
321 | |
322 | list_for_each_entry_safe(device, dt, &host1x->devices, list) { |
323 | list_for_each_entry(subdev, &device->active, list) { |
324 | if (subdev->client == client) { |
325 | host1x_subdev_unregister(device, subdev); |
326 | mutex_unlock(lock: &host1x->devices_lock); |
327 | return 0; |
328 | } |
329 | } |
330 | } |
331 | |
332 | mutex_unlock(lock: &host1x->devices_lock); |
333 | return -ENODEV; |
334 | } |
335 | |
336 | static int host1x_device_match(struct device *dev, struct device_driver *drv) |
337 | { |
338 | return strcmp(dev_name(dev), drv->name) == 0; |
339 | } |
340 | |
341 | /* |
342 | * Note that this is really only needed for backwards compatibility |
343 | * with libdrm, which parses this information from sysfs and will |
344 | * fail if it can't find the OF_FULLNAME, specifically. |
345 | */ |
346 | static int host1x_device_uevent(const struct device *dev, |
347 | struct kobj_uevent_env *env) |
348 | { |
349 | of_device_uevent(dev: dev->parent, env); |
350 | |
351 | return 0; |
352 | } |
353 | |
354 | static int host1x_dma_configure(struct device *dev) |
355 | { |
356 | return of_dma_configure(dev, np: dev->of_node, force_dma: true); |
357 | } |
358 | |
359 | static const struct dev_pm_ops host1x_device_pm_ops = { |
360 | .suspend = pm_generic_suspend, |
361 | .resume = pm_generic_resume, |
362 | .freeze = pm_generic_freeze, |
363 | .thaw = pm_generic_thaw, |
364 | .poweroff = pm_generic_poweroff, |
365 | .restore = pm_generic_restore, |
366 | }; |
367 | |
368 | struct bus_type host1x_bus_type = { |
369 | .name = "host1x" , |
370 | .match = host1x_device_match, |
371 | .uevent = host1x_device_uevent, |
372 | .dma_configure = host1x_dma_configure, |
373 | .pm = &host1x_device_pm_ops, |
374 | }; |
375 | |
376 | static void __host1x_device_del(struct host1x_device *device) |
377 | { |
378 | struct host1x_subdev *subdev, *sd; |
379 | struct host1x_client *client, *cl; |
380 | |
381 | mutex_lock(&device->subdevs_lock); |
382 | |
383 | /* unregister subdevices */ |
384 | list_for_each_entry_safe(subdev, sd, &device->active, list) { |
385 | /* |
386 | * host1x_subdev_unregister() will remove the client from |
387 | * any lists, so we'll need to manually add it back to the |
388 | * list of idle clients. |
389 | * |
390 | * XXX: Alternatively, perhaps don't remove the client from |
391 | * any lists in host1x_subdev_unregister() and instead do |
392 | * that explicitly from host1x_unregister_client()? |
393 | */ |
394 | client = subdev->client; |
395 | |
396 | __host1x_subdev_unregister(device, subdev); |
397 | |
398 | /* add the client to the list of idle clients */ |
399 | mutex_lock(&clients_lock); |
400 | list_add_tail(new: &client->list, head: &clients); |
401 | mutex_unlock(lock: &clients_lock); |
402 | } |
403 | |
404 | /* remove subdevices */ |
405 | list_for_each_entry_safe(subdev, sd, &device->subdevs, list) |
406 | host1x_subdev_del(subdev); |
407 | |
408 | mutex_unlock(lock: &device->subdevs_lock); |
409 | |
410 | /* move clients to idle list */ |
411 | mutex_lock(&clients_lock); |
412 | mutex_lock(&device->clients_lock); |
413 | |
414 | list_for_each_entry_safe(client, cl, &device->clients, list) |
415 | list_move_tail(list: &client->list, head: &clients); |
416 | |
417 | mutex_unlock(lock: &device->clients_lock); |
418 | mutex_unlock(lock: &clients_lock); |
419 | |
420 | /* finally remove the device */ |
421 | list_del_init(entry: &device->list); |
422 | } |
423 | |
424 | static void host1x_device_release(struct device *dev) |
425 | { |
426 | struct host1x_device *device = to_host1x_device(dev); |
427 | |
428 | __host1x_device_del(device); |
429 | kfree(objp: device); |
430 | } |
431 | |
432 | static int host1x_device_add(struct host1x *host1x, |
433 | struct host1x_driver *driver) |
434 | { |
435 | struct host1x_client *client, *tmp; |
436 | struct host1x_subdev *subdev; |
437 | struct host1x_device *device; |
438 | int err; |
439 | |
440 | device = kzalloc(size: sizeof(*device), GFP_KERNEL); |
441 | if (!device) |
442 | return -ENOMEM; |
443 | |
444 | device_initialize(dev: &device->dev); |
445 | |
446 | mutex_init(&device->subdevs_lock); |
447 | INIT_LIST_HEAD(list: &device->subdevs); |
448 | INIT_LIST_HEAD(list: &device->active); |
449 | mutex_init(&device->clients_lock); |
450 | INIT_LIST_HEAD(list: &device->clients); |
451 | INIT_LIST_HEAD(list: &device->list); |
452 | device->driver = driver; |
453 | |
454 | device->dev.coherent_dma_mask = host1x->dev->coherent_dma_mask; |
455 | device->dev.dma_mask = &device->dev.coherent_dma_mask; |
456 | dev_set_name(dev: &device->dev, name: "%s" , driver->driver.name); |
457 | device->dev.release = host1x_device_release; |
458 | device->dev.bus = &host1x_bus_type; |
459 | device->dev.parent = host1x->dev; |
460 | |
461 | of_dma_configure(dev: &device->dev, np: host1x->dev->of_node, force_dma: true); |
462 | |
463 | device->dev.dma_parms = &device->dma_parms; |
464 | dma_set_max_seg_size(dev: &device->dev, UINT_MAX); |
465 | |
466 | err = host1x_device_parse_dt(device, driver); |
467 | if (err < 0) { |
468 | kfree(objp: device); |
469 | return err; |
470 | } |
471 | |
472 | list_add_tail(new: &device->list, head: &host1x->devices); |
473 | |
474 | mutex_lock(&clients_lock); |
475 | |
476 | list_for_each_entry_safe(client, tmp, &clients, list) { |
477 | list_for_each_entry(subdev, &device->subdevs, list) { |
478 | if (subdev->np == client->dev->of_node) { |
479 | host1x_subdev_register(device, subdev, client); |
480 | break; |
481 | } |
482 | } |
483 | } |
484 | |
485 | mutex_unlock(lock: &clients_lock); |
486 | |
487 | return 0; |
488 | } |
489 | |
490 | /* |
491 | * Removes a device by first unregistering any subdevices and then removing |
492 | * itself from the list of devices. |
493 | * |
494 | * This function must be called with the host1x->devices_lock held. |
495 | */ |
496 | static void host1x_device_del(struct host1x *host1x, |
497 | struct host1x_device *device) |
498 | { |
499 | if (device->registered) { |
500 | device->registered = false; |
501 | device_del(dev: &device->dev); |
502 | } |
503 | |
504 | put_device(dev: &device->dev); |
505 | } |
506 | |
507 | static void host1x_attach_driver(struct host1x *host1x, |
508 | struct host1x_driver *driver) |
509 | { |
510 | struct host1x_device *device; |
511 | int err; |
512 | |
513 | mutex_lock(&host1x->devices_lock); |
514 | |
515 | list_for_each_entry(device, &host1x->devices, list) { |
516 | if (device->driver == driver) { |
517 | mutex_unlock(lock: &host1x->devices_lock); |
518 | return; |
519 | } |
520 | } |
521 | |
522 | err = host1x_device_add(host1x, driver); |
523 | if (err < 0) |
524 | dev_err(host1x->dev, "failed to allocate device: %d\n" , err); |
525 | |
526 | mutex_unlock(lock: &host1x->devices_lock); |
527 | } |
528 | |
529 | static void host1x_detach_driver(struct host1x *host1x, |
530 | struct host1x_driver *driver) |
531 | { |
532 | struct host1x_device *device, *tmp; |
533 | |
534 | mutex_lock(&host1x->devices_lock); |
535 | |
536 | list_for_each_entry_safe(device, tmp, &host1x->devices, list) |
537 | if (device->driver == driver) |
538 | host1x_device_del(host1x, device); |
539 | |
540 | mutex_unlock(lock: &host1x->devices_lock); |
541 | } |
542 | |
543 | static int host1x_devices_show(struct seq_file *s, void *data) |
544 | { |
545 | struct host1x *host1x = s->private; |
546 | struct host1x_device *device; |
547 | |
548 | mutex_lock(&host1x->devices_lock); |
549 | |
550 | list_for_each_entry(device, &host1x->devices, list) { |
551 | struct host1x_subdev *subdev; |
552 | |
553 | seq_printf(m: s, fmt: "%s\n" , dev_name(dev: &device->dev)); |
554 | |
555 | mutex_lock(&device->subdevs_lock); |
556 | |
557 | list_for_each_entry(subdev, &device->active, list) |
558 | seq_printf(m: s, fmt: " %pOFf: %s\n" , subdev->np, |
559 | dev_name(dev: subdev->client->dev)); |
560 | |
561 | list_for_each_entry(subdev, &device->subdevs, list) |
562 | seq_printf(m: s, fmt: " %pOFf:\n" , subdev->np); |
563 | |
564 | mutex_unlock(lock: &device->subdevs_lock); |
565 | } |
566 | |
567 | mutex_unlock(lock: &host1x->devices_lock); |
568 | |
569 | return 0; |
570 | } |
571 | DEFINE_SHOW_ATTRIBUTE(host1x_devices); |
572 | |
573 | /** |
574 | * host1x_register() - register a host1x controller |
575 | * @host1x: host1x controller |
576 | * |
577 | * The host1x controller driver uses this to register a host1x controller with |
578 | * the infrastructure. Note that all Tegra SoC generations have only ever come |
579 | * with a single host1x instance, so this function is somewhat academic. |
580 | */ |
581 | int host1x_register(struct host1x *host1x) |
582 | { |
583 | struct host1x_driver *driver; |
584 | |
585 | mutex_lock(&devices_lock); |
586 | list_add_tail(new: &host1x->list, head: &devices); |
587 | mutex_unlock(lock: &devices_lock); |
588 | |
589 | mutex_lock(&drivers_lock); |
590 | |
591 | list_for_each_entry(driver, &drivers, list) |
592 | host1x_attach_driver(host1x, driver); |
593 | |
594 | mutex_unlock(lock: &drivers_lock); |
595 | |
596 | debugfs_create_file(name: "devices" , S_IRUGO, parent: host1x->debugfs, data: host1x, |
597 | fops: &host1x_devices_fops); |
598 | |
599 | return 0; |
600 | } |
601 | |
602 | /** |
603 | * host1x_unregister() - unregister a host1x controller |
604 | * @host1x: host1x controller |
605 | * |
606 | * The host1x controller driver uses this to remove a host1x controller from |
607 | * the infrastructure. |
608 | */ |
609 | int host1x_unregister(struct host1x *host1x) |
610 | { |
611 | struct host1x_driver *driver; |
612 | |
613 | mutex_lock(&drivers_lock); |
614 | |
615 | list_for_each_entry(driver, &drivers, list) |
616 | host1x_detach_driver(host1x, driver); |
617 | |
618 | mutex_unlock(lock: &drivers_lock); |
619 | |
620 | mutex_lock(&devices_lock); |
621 | list_del_init(entry: &host1x->list); |
622 | mutex_unlock(lock: &devices_lock); |
623 | |
624 | return 0; |
625 | } |
626 | |
627 | static int host1x_device_probe(struct device *dev) |
628 | { |
629 | struct host1x_driver *driver = to_host1x_driver(driver: dev->driver); |
630 | struct host1x_device *device = to_host1x_device(dev); |
631 | |
632 | if (driver->probe) |
633 | return driver->probe(device); |
634 | |
635 | return 0; |
636 | } |
637 | |
638 | static int host1x_device_remove(struct device *dev) |
639 | { |
640 | struct host1x_driver *driver = to_host1x_driver(driver: dev->driver); |
641 | struct host1x_device *device = to_host1x_device(dev); |
642 | |
643 | if (driver->remove) |
644 | return driver->remove(device); |
645 | |
646 | return 0; |
647 | } |
648 | |
649 | static void host1x_device_shutdown(struct device *dev) |
650 | { |
651 | struct host1x_driver *driver = to_host1x_driver(driver: dev->driver); |
652 | struct host1x_device *device = to_host1x_device(dev); |
653 | |
654 | if (driver->shutdown) |
655 | driver->shutdown(device); |
656 | } |
657 | |
658 | /** |
659 | * host1x_driver_register_full() - register a host1x driver |
660 | * @driver: host1x driver |
661 | * @owner: owner module |
662 | * |
663 | * Drivers for host1x logical devices call this function to register a driver |
664 | * with the infrastructure. Note that since these drive logical devices, the |
665 | * registration of the driver actually triggers tho logical device creation. |
666 | * A logical device will be created for each host1x instance. |
667 | */ |
668 | int host1x_driver_register_full(struct host1x_driver *driver, |
669 | struct module *owner) |
670 | { |
671 | struct host1x *host1x; |
672 | |
673 | INIT_LIST_HEAD(list: &driver->list); |
674 | |
675 | mutex_lock(&drivers_lock); |
676 | list_add_tail(new: &driver->list, head: &drivers); |
677 | mutex_unlock(lock: &drivers_lock); |
678 | |
679 | mutex_lock(&devices_lock); |
680 | |
681 | list_for_each_entry(host1x, &devices, list) |
682 | host1x_attach_driver(host1x, driver); |
683 | |
684 | mutex_unlock(lock: &devices_lock); |
685 | |
686 | driver->driver.bus = &host1x_bus_type; |
687 | driver->driver.owner = owner; |
688 | driver->driver.probe = host1x_device_probe; |
689 | driver->driver.remove = host1x_device_remove; |
690 | driver->driver.shutdown = host1x_device_shutdown; |
691 | |
692 | return driver_register(drv: &driver->driver); |
693 | } |
694 | EXPORT_SYMBOL(host1x_driver_register_full); |
695 | |
696 | /** |
697 | * host1x_driver_unregister() - unregister a host1x driver |
698 | * @driver: host1x driver |
699 | * |
700 | * Unbinds the driver from each of the host1x logical devices that it is |
701 | * bound to, effectively removing the subsystem devices that they represent. |
702 | */ |
703 | void host1x_driver_unregister(struct host1x_driver *driver) |
704 | { |
705 | struct host1x *host1x; |
706 | |
707 | driver_unregister(drv: &driver->driver); |
708 | |
709 | mutex_lock(&devices_lock); |
710 | |
711 | list_for_each_entry(host1x, &devices, list) |
712 | host1x_detach_driver(host1x, driver); |
713 | |
714 | mutex_unlock(lock: &devices_lock); |
715 | |
716 | mutex_lock(&drivers_lock); |
717 | list_del_init(entry: &driver->list); |
718 | mutex_unlock(lock: &drivers_lock); |
719 | } |
720 | EXPORT_SYMBOL(host1x_driver_unregister); |
721 | |
722 | /** |
723 | * __host1x_client_init() - initialize a host1x client |
724 | * @client: host1x client |
725 | * @key: lock class key for the client-specific mutex |
726 | */ |
727 | void __host1x_client_init(struct host1x_client *client, struct lock_class_key *key) |
728 | { |
729 | host1x_bo_cache_init(cache: &client->cache); |
730 | INIT_LIST_HEAD(list: &client->list); |
731 | __mutex_init(lock: &client->lock, name: "host1x client lock" , key); |
732 | client->usecount = 0; |
733 | } |
734 | EXPORT_SYMBOL(__host1x_client_init); |
735 | |
736 | /** |
737 | * host1x_client_exit() - uninitialize a host1x client |
738 | * @client: host1x client |
739 | */ |
740 | void host1x_client_exit(struct host1x_client *client) |
741 | { |
742 | mutex_destroy(lock: &client->lock); |
743 | } |
744 | EXPORT_SYMBOL(host1x_client_exit); |
745 | |
746 | /** |
747 | * __host1x_client_register() - register a host1x client |
748 | * @client: host1x client |
749 | * |
750 | * Registers a host1x client with each host1x controller instance. Note that |
751 | * each client will only match their parent host1x controller and will only be |
752 | * associated with that instance. Once all clients have been registered with |
753 | * their parent host1x controller, the infrastructure will set up the logical |
754 | * device and call host1x_device_init(), which will in turn call each client's |
755 | * &host1x_client_ops.init implementation. |
756 | */ |
757 | int __host1x_client_register(struct host1x_client *client) |
758 | { |
759 | struct host1x *host1x; |
760 | int err; |
761 | |
762 | mutex_lock(&devices_lock); |
763 | |
764 | list_for_each_entry(host1x, &devices, list) { |
765 | err = host1x_add_client(host1x, client); |
766 | if (!err) { |
767 | mutex_unlock(lock: &devices_lock); |
768 | return 0; |
769 | } |
770 | } |
771 | |
772 | mutex_unlock(lock: &devices_lock); |
773 | |
774 | mutex_lock(&clients_lock); |
775 | list_add_tail(new: &client->list, head: &clients); |
776 | mutex_unlock(lock: &clients_lock); |
777 | |
778 | return 0; |
779 | } |
780 | EXPORT_SYMBOL(__host1x_client_register); |
781 | |
782 | /** |
783 | * host1x_client_unregister() - unregister a host1x client |
784 | * @client: host1x client |
785 | * |
786 | * Removes a host1x client from its host1x controller instance. If a logical |
787 | * device has already been initialized, it will be torn down. |
788 | */ |
789 | void host1x_client_unregister(struct host1x_client *client) |
790 | { |
791 | struct host1x_client *c; |
792 | struct host1x *host1x; |
793 | int err; |
794 | |
795 | mutex_lock(&devices_lock); |
796 | |
797 | list_for_each_entry(host1x, &devices, list) { |
798 | err = host1x_del_client(host1x, client); |
799 | if (!err) { |
800 | mutex_unlock(lock: &devices_lock); |
801 | return; |
802 | } |
803 | } |
804 | |
805 | mutex_unlock(lock: &devices_lock); |
806 | mutex_lock(&clients_lock); |
807 | |
808 | list_for_each_entry(c, &clients, list) { |
809 | if (c == client) { |
810 | list_del_init(entry: &c->list); |
811 | break; |
812 | } |
813 | } |
814 | |
815 | mutex_unlock(lock: &clients_lock); |
816 | |
817 | host1x_bo_cache_destroy(cache: &client->cache); |
818 | } |
819 | EXPORT_SYMBOL(host1x_client_unregister); |
820 | |
821 | int host1x_client_suspend(struct host1x_client *client) |
822 | { |
823 | int err = 0; |
824 | |
825 | mutex_lock(&client->lock); |
826 | |
827 | if (client->usecount == 1) { |
828 | if (client->ops && client->ops->suspend) { |
829 | err = client->ops->suspend(client); |
830 | if (err < 0) |
831 | goto unlock; |
832 | } |
833 | } |
834 | |
835 | client->usecount--; |
836 | dev_dbg(client->dev, "use count: %u\n" , client->usecount); |
837 | |
838 | if (client->parent) { |
839 | err = host1x_client_suspend(client: client->parent); |
840 | if (err < 0) |
841 | goto resume; |
842 | } |
843 | |
844 | goto unlock; |
845 | |
846 | resume: |
847 | if (client->usecount == 0) |
848 | if (client->ops && client->ops->resume) |
849 | client->ops->resume(client); |
850 | |
851 | client->usecount++; |
852 | unlock: |
853 | mutex_unlock(lock: &client->lock); |
854 | return err; |
855 | } |
856 | EXPORT_SYMBOL(host1x_client_suspend); |
857 | |
858 | int host1x_client_resume(struct host1x_client *client) |
859 | { |
860 | int err = 0; |
861 | |
862 | mutex_lock(&client->lock); |
863 | |
864 | if (client->parent) { |
865 | err = host1x_client_resume(client: client->parent); |
866 | if (err < 0) |
867 | goto unlock; |
868 | } |
869 | |
870 | if (client->usecount == 0) { |
871 | if (client->ops && client->ops->resume) { |
872 | err = client->ops->resume(client); |
873 | if (err < 0) |
874 | goto suspend; |
875 | } |
876 | } |
877 | |
878 | client->usecount++; |
879 | dev_dbg(client->dev, "use count: %u\n" , client->usecount); |
880 | |
881 | goto unlock; |
882 | |
883 | suspend: |
884 | if (client->parent) |
885 | host1x_client_suspend(client->parent); |
886 | unlock: |
887 | mutex_unlock(lock: &client->lock); |
888 | return err; |
889 | } |
890 | EXPORT_SYMBOL(host1x_client_resume); |
891 | |
892 | struct host1x_bo_mapping *host1x_bo_pin(struct device *dev, struct host1x_bo *bo, |
893 | enum dma_data_direction dir, |
894 | struct host1x_bo_cache *cache) |
895 | { |
896 | struct host1x_bo_mapping *mapping; |
897 | |
898 | if (cache) { |
899 | mutex_lock(&cache->lock); |
900 | |
901 | list_for_each_entry(mapping, &cache->mappings, entry) { |
902 | if (mapping->bo == bo && mapping->direction == dir) { |
903 | kref_get(kref: &mapping->ref); |
904 | goto unlock; |
905 | } |
906 | } |
907 | } |
908 | |
909 | mapping = bo->ops->pin(dev, bo, dir); |
910 | if (IS_ERR(ptr: mapping)) |
911 | goto unlock; |
912 | |
913 | spin_lock(lock: &mapping->bo->lock); |
914 | list_add_tail(new: &mapping->list, head: &bo->mappings); |
915 | spin_unlock(lock: &mapping->bo->lock); |
916 | |
917 | if (cache) { |
918 | INIT_LIST_HEAD(list: &mapping->entry); |
919 | mapping->cache = cache; |
920 | |
921 | list_add_tail(new: &mapping->entry, head: &cache->mappings); |
922 | |
923 | /* bump reference count to track the copy in the cache */ |
924 | kref_get(kref: &mapping->ref); |
925 | } |
926 | |
927 | unlock: |
928 | if (cache) |
929 | mutex_unlock(lock: &cache->lock); |
930 | |
931 | return mapping; |
932 | } |
933 | EXPORT_SYMBOL(host1x_bo_pin); |
934 | |
935 | static void __host1x_bo_unpin(struct kref *ref) |
936 | { |
937 | struct host1x_bo_mapping *mapping = to_host1x_bo_mapping(ref); |
938 | |
939 | /* |
940 | * When the last reference of the mapping goes away, make sure to remove the mapping from |
941 | * the cache. |
942 | */ |
943 | if (mapping->cache) |
944 | list_del(entry: &mapping->entry); |
945 | |
946 | spin_lock(lock: &mapping->bo->lock); |
947 | list_del(entry: &mapping->list); |
948 | spin_unlock(lock: &mapping->bo->lock); |
949 | |
950 | mapping->bo->ops->unpin(mapping); |
951 | } |
952 | |
953 | void host1x_bo_unpin(struct host1x_bo_mapping *mapping) |
954 | { |
955 | struct host1x_bo_cache *cache = mapping->cache; |
956 | |
957 | if (cache) |
958 | mutex_lock(&cache->lock); |
959 | |
960 | kref_put(kref: &mapping->ref, release: __host1x_bo_unpin); |
961 | |
962 | if (cache) |
963 | mutex_unlock(lock: &cache->lock); |
964 | } |
965 | EXPORT_SYMBOL(host1x_bo_unpin); |
966 | |