877 lines
20 KiB
C
877 lines
20 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* V4L2 asynchronous subdevice registration API
|
|
*
|
|
* Copyright (C) 2012-2013, Guennadi Liakhovetski <g.liakhovetski@gmx.de>
|
|
*/
|
|
|
|
#include <linux/debugfs.h>
|
|
#include <linux/device.h>
|
|
#include <linux/err.h>
|
|
#include <linux/i2c.h>
|
|
#include <linux/list.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/module.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/of.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/seq_file.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/types.h>
|
|
|
|
#include <media/v4l2-async.h>
|
|
#include <media/v4l2-device.h>
|
|
#include <media/v4l2-fwnode.h>
|
|
#include <media/v4l2-subdev.h>
|
|
|
|
static int v4l2_async_nf_call_bound(struct v4l2_async_notifier *n,
|
|
struct v4l2_subdev *subdev,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
if (!n->ops || !n->ops->bound)
|
|
return 0;
|
|
|
|
return n->ops->bound(n, subdev, asd);
|
|
}
|
|
|
|
static void v4l2_async_nf_call_unbind(struct v4l2_async_notifier *n,
|
|
struct v4l2_subdev *subdev,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
if (!n->ops || !n->ops->unbind)
|
|
return;
|
|
|
|
n->ops->unbind(n, subdev, asd);
|
|
}
|
|
|
|
static int v4l2_async_nf_call_complete(struct v4l2_async_notifier *n)
|
|
{
|
|
if (!n->ops || !n->ops->complete)
|
|
return 0;
|
|
|
|
return n->ops->complete(n);
|
|
}
|
|
|
|
static bool match_i2c(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_subdev *sd, struct v4l2_async_subdev *asd)
|
|
{
|
|
#if IS_ENABLED(CONFIG_I2C)
|
|
struct i2c_client *client = i2c_verify_client(sd->dev);
|
|
|
|
return client &&
|
|
asd->match.i2c.adapter_id == client->adapter->nr &&
|
|
asd->match.i2c.address == client->addr;
|
|
#else
|
|
return false;
|
|
#endif
|
|
}
|
|
|
|
static bool match_fwnode(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_subdev *sd, struct v4l2_async_subdev *asd)
|
|
{
|
|
struct fwnode_handle *other_fwnode;
|
|
struct fwnode_handle *dev_fwnode;
|
|
bool asd_fwnode_is_ep;
|
|
bool sd_fwnode_is_ep;
|
|
struct device *dev;
|
|
|
|
/*
|
|
* Both the subdev and the async subdev can provide either an endpoint
|
|
* fwnode or a device fwnode. Start with the simple case of direct
|
|
* fwnode matching.
|
|
*/
|
|
if (sd->fwnode == asd->match.fwnode)
|
|
return true;
|
|
|
|
/*
|
|
* Check the same situation for any possible secondary assigned to the
|
|
* subdev's fwnode
|
|
*/
|
|
if (!IS_ERR_OR_NULL(sd->fwnode->secondary) &&
|
|
sd->fwnode->secondary == asd->match.fwnode)
|
|
return true;
|
|
|
|
/*
|
|
* Otherwise, check if the sd fwnode and the asd fwnode refer to an
|
|
* endpoint or a device. If they're of the same type, there's no match.
|
|
* Technically speaking this checks if the nodes refer to a connected
|
|
* endpoint, which is the simplest check that works for both OF and
|
|
* ACPI. This won't make a difference, as drivers should not try to
|
|
* match unconnected endpoints.
|
|
*/
|
|
sd_fwnode_is_ep = fwnode_graph_is_endpoint(sd->fwnode);
|
|
asd_fwnode_is_ep = fwnode_graph_is_endpoint(asd->match.fwnode);
|
|
|
|
if (sd_fwnode_is_ep == asd_fwnode_is_ep)
|
|
return false;
|
|
|
|
/*
|
|
* The sd and asd fwnodes are of different types. Get the device fwnode
|
|
* parent of the endpoint fwnode, and compare it with the other fwnode.
|
|
*/
|
|
if (sd_fwnode_is_ep) {
|
|
dev_fwnode = fwnode_graph_get_port_parent(sd->fwnode);
|
|
other_fwnode = asd->match.fwnode;
|
|
} else {
|
|
dev_fwnode = fwnode_graph_get_port_parent(asd->match.fwnode);
|
|
other_fwnode = sd->fwnode;
|
|
}
|
|
|
|
fwnode_handle_put(dev_fwnode);
|
|
|
|
if (dev_fwnode != other_fwnode)
|
|
return false;
|
|
|
|
/*
|
|
* We have a heterogeneous match. Retrieve the struct device of the side
|
|
* that matched on a device fwnode to print its driver name.
|
|
*/
|
|
if (sd_fwnode_is_ep)
|
|
dev = notifier->v4l2_dev ? notifier->v4l2_dev->dev
|
|
: notifier->sd->dev;
|
|
else
|
|
dev = sd->dev;
|
|
|
|
if (dev && dev->driver) {
|
|
if (sd_fwnode_is_ep)
|
|
dev_warn(dev, "Driver %s uses device fwnode, incorrect match may occur\n",
|
|
dev->driver->name);
|
|
dev_notice(dev, "Consider updating driver %s to match on endpoints\n",
|
|
dev->driver->name);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static LIST_HEAD(subdev_list);
|
|
static LIST_HEAD(notifier_list);
|
|
static DEFINE_MUTEX(list_lock);
|
|
|
|
static struct v4l2_async_subdev *
|
|
v4l2_async_find_match(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_subdev *sd)
|
|
{
|
|
bool (*match)(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_subdev *sd, struct v4l2_async_subdev *asd);
|
|
struct v4l2_async_subdev *asd;
|
|
|
|
list_for_each_entry(asd, ¬ifier->waiting, list) {
|
|
/* bus_type has been verified valid before */
|
|
switch (asd->match_type) {
|
|
case V4L2_ASYNC_MATCH_I2C:
|
|
match = match_i2c;
|
|
break;
|
|
case V4L2_ASYNC_MATCH_FWNODE:
|
|
match = match_fwnode;
|
|
break;
|
|
default:
|
|
/* Cannot happen, unless someone breaks us */
|
|
WARN_ON(true);
|
|
return NULL;
|
|
}
|
|
|
|
/* match cannot be NULL here */
|
|
if (match(notifier, sd, asd))
|
|
return asd;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/* Compare two async sub-device descriptors for equivalence */
|
|
static bool asd_equal(struct v4l2_async_subdev *asd_x,
|
|
struct v4l2_async_subdev *asd_y)
|
|
{
|
|
if (asd_x->match_type != asd_y->match_type)
|
|
return false;
|
|
|
|
switch (asd_x->match_type) {
|
|
case V4L2_ASYNC_MATCH_I2C:
|
|
return asd_x->match.i2c.adapter_id ==
|
|
asd_y->match.i2c.adapter_id &&
|
|
asd_x->match.i2c.address ==
|
|
asd_y->match.i2c.address;
|
|
case V4L2_ASYNC_MATCH_FWNODE:
|
|
return asd_x->match.fwnode == asd_y->match.fwnode;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
/* Find the sub-device notifier registered by a sub-device driver. */
|
|
static struct v4l2_async_notifier *
|
|
v4l2_async_find_subdev_notifier(struct v4l2_subdev *sd)
|
|
{
|
|
struct v4l2_async_notifier *n;
|
|
|
|
list_for_each_entry(n, ¬ifier_list, list)
|
|
if (n->sd == sd)
|
|
return n;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/* Get v4l2_device related to the notifier if one can be found. */
|
|
static struct v4l2_device *
|
|
v4l2_async_nf_find_v4l2_dev(struct v4l2_async_notifier *notifier)
|
|
{
|
|
while (notifier->parent)
|
|
notifier = notifier->parent;
|
|
|
|
return notifier->v4l2_dev;
|
|
}
|
|
|
|
/*
|
|
* Return true if all child sub-device notifiers are complete, false otherwise.
|
|
*/
|
|
static bool
|
|
v4l2_async_nf_can_complete(struct v4l2_async_notifier *notifier)
|
|
{
|
|
struct v4l2_subdev *sd;
|
|
|
|
if (!list_empty(¬ifier->waiting))
|
|
return false;
|
|
|
|
list_for_each_entry(sd, ¬ifier->done, async_list) {
|
|
struct v4l2_async_notifier *subdev_notifier =
|
|
v4l2_async_find_subdev_notifier(sd);
|
|
|
|
if (subdev_notifier &&
|
|
!v4l2_async_nf_can_complete(subdev_notifier))
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
/*
|
|
* Complete the master notifier if possible. This is done when all async
|
|
* sub-devices have been bound; v4l2_device is also available then.
|
|
*/
|
|
static int
|
|
v4l2_async_nf_try_complete(struct v4l2_async_notifier *notifier)
|
|
{
|
|
/* Quick check whether there are still more sub-devices here. */
|
|
if (!list_empty(¬ifier->waiting))
|
|
return 0;
|
|
|
|
/* Check the entire notifier tree; find the root notifier first. */
|
|
while (notifier->parent)
|
|
notifier = notifier->parent;
|
|
|
|
/* This is root if it has v4l2_dev. */
|
|
if (!notifier->v4l2_dev)
|
|
return 0;
|
|
|
|
/* Is everything ready? */
|
|
if (!v4l2_async_nf_can_complete(notifier))
|
|
return 0;
|
|
|
|
return v4l2_async_nf_call_complete(notifier);
|
|
}
|
|
|
|
static int
|
|
v4l2_async_nf_try_all_subdevs(struct v4l2_async_notifier *notifier);
|
|
|
|
static int v4l2_async_match_notify(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_device *v4l2_dev,
|
|
struct v4l2_subdev *sd,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
struct v4l2_async_notifier *subdev_notifier;
|
|
int ret;
|
|
|
|
ret = v4l2_device_register_subdev(v4l2_dev, sd);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
ret = v4l2_async_nf_call_bound(notifier, sd, asd);
|
|
if (ret < 0) {
|
|
v4l2_device_unregister_subdev(sd);
|
|
return ret;
|
|
}
|
|
|
|
/* Remove from the waiting list */
|
|
list_del(&asd->list);
|
|
sd->asd = asd;
|
|
sd->notifier = notifier;
|
|
|
|
/* Move from the global subdevice list to notifier's done */
|
|
list_move(&sd->async_list, ¬ifier->done);
|
|
|
|
/*
|
|
* See if the sub-device has a notifier. If not, return here.
|
|
*/
|
|
subdev_notifier = v4l2_async_find_subdev_notifier(sd);
|
|
if (!subdev_notifier || subdev_notifier->parent)
|
|
return 0;
|
|
|
|
/*
|
|
* Proceed with checking for the sub-device notifier's async
|
|
* sub-devices, and return the result. The error will be handled by the
|
|
* caller.
|
|
*/
|
|
subdev_notifier->parent = notifier;
|
|
|
|
return v4l2_async_nf_try_all_subdevs(subdev_notifier);
|
|
}
|
|
|
|
/* Test all async sub-devices in a notifier for a match. */
|
|
static int
|
|
v4l2_async_nf_try_all_subdevs(struct v4l2_async_notifier *notifier)
|
|
{
|
|
struct v4l2_device *v4l2_dev =
|
|
v4l2_async_nf_find_v4l2_dev(notifier);
|
|
struct v4l2_subdev *sd;
|
|
|
|
if (!v4l2_dev)
|
|
return 0;
|
|
|
|
again:
|
|
list_for_each_entry(sd, &subdev_list, async_list) {
|
|
struct v4l2_async_subdev *asd;
|
|
int ret;
|
|
|
|
asd = v4l2_async_find_match(notifier, sd);
|
|
if (!asd)
|
|
continue;
|
|
|
|
ret = v4l2_async_match_notify(notifier, v4l2_dev, sd, asd);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
/*
|
|
* v4l2_async_match_notify() may lead to registering a
|
|
* new notifier and thus changing the async subdevs
|
|
* list. In order to proceed safely from here, restart
|
|
* parsing the list from the beginning.
|
|
*/
|
|
goto again;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void v4l2_async_cleanup(struct v4l2_subdev *sd)
|
|
{
|
|
v4l2_device_unregister_subdev(sd);
|
|
/*
|
|
* Subdevice driver will reprobe and put the subdev back
|
|
* onto the list
|
|
*/
|
|
list_del_init(&sd->async_list);
|
|
sd->asd = NULL;
|
|
}
|
|
|
|
/* Unbind all sub-devices in the notifier tree. */
|
|
static void
|
|
v4l2_async_nf_unbind_all_subdevs(struct v4l2_async_notifier *notifier)
|
|
{
|
|
struct v4l2_subdev *sd, *tmp;
|
|
|
|
list_for_each_entry_safe(sd, tmp, ¬ifier->done, async_list) {
|
|
struct v4l2_async_notifier *subdev_notifier =
|
|
v4l2_async_find_subdev_notifier(sd);
|
|
|
|
if (subdev_notifier)
|
|
v4l2_async_nf_unbind_all_subdevs(subdev_notifier);
|
|
|
|
v4l2_async_nf_call_unbind(notifier, sd, sd->asd);
|
|
v4l2_async_cleanup(sd);
|
|
|
|
list_move(&sd->async_list, &subdev_list);
|
|
}
|
|
|
|
notifier->parent = NULL;
|
|
}
|
|
|
|
/* See if an async sub-device can be found in a notifier's lists. */
|
|
static bool
|
|
__v4l2_async_nf_has_async_subdev(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
struct v4l2_async_subdev *asd_y;
|
|
struct v4l2_subdev *sd;
|
|
|
|
list_for_each_entry(asd_y, ¬ifier->waiting, list)
|
|
if (asd_equal(asd, asd_y))
|
|
return true;
|
|
|
|
list_for_each_entry(sd, ¬ifier->done, async_list) {
|
|
if (WARN_ON(!sd->asd))
|
|
continue;
|
|
|
|
if (asd_equal(asd, sd->asd))
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
/*
|
|
* Find out whether an async sub-device was set up already or
|
|
* whether it exists in a given notifier before @this_index.
|
|
* If @this_index < 0, search the notifier's entire @asd_list.
|
|
*/
|
|
static bool
|
|
v4l2_async_nf_has_async_subdev(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_async_subdev *asd, int this_index)
|
|
{
|
|
struct v4l2_async_subdev *asd_y;
|
|
int j = 0;
|
|
|
|
lockdep_assert_held(&list_lock);
|
|
|
|
/* Check that an asd is not being added more than once. */
|
|
list_for_each_entry(asd_y, ¬ifier->asd_list, asd_list) {
|
|
if (this_index >= 0 && j++ >= this_index)
|
|
break;
|
|
if (asd_equal(asd, asd_y))
|
|
return true;
|
|
}
|
|
|
|
/* Check that an asd does not exist in other notifiers. */
|
|
list_for_each_entry(notifier, ¬ifier_list, list)
|
|
if (__v4l2_async_nf_has_async_subdev(notifier, asd))
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
static int v4l2_async_nf_asd_valid(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_async_subdev *asd,
|
|
int this_index)
|
|
{
|
|
struct device *dev =
|
|
notifier->v4l2_dev ? notifier->v4l2_dev->dev : NULL;
|
|
|
|
if (!asd)
|
|
return -EINVAL;
|
|
|
|
switch (asd->match_type) {
|
|
case V4L2_ASYNC_MATCH_I2C:
|
|
case V4L2_ASYNC_MATCH_FWNODE:
|
|
if (v4l2_async_nf_has_async_subdev(notifier, asd, this_index)) {
|
|
dev_dbg(dev, "subdev descriptor already listed in this or other notifiers\n");
|
|
return -EEXIST;
|
|
}
|
|
break;
|
|
default:
|
|
dev_err(dev, "Invalid match type %u on %p\n",
|
|
asd->match_type, asd);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void v4l2_async_nf_init(struct v4l2_async_notifier *notifier)
|
|
{
|
|
INIT_LIST_HEAD(¬ifier->asd_list);
|
|
}
|
|
EXPORT_SYMBOL(v4l2_async_nf_init);
|
|
|
|
static int __v4l2_async_nf_register(struct v4l2_async_notifier *notifier)
|
|
{
|
|
struct v4l2_async_subdev *asd;
|
|
int ret, i = 0;
|
|
|
|
INIT_LIST_HEAD(¬ifier->waiting);
|
|
INIT_LIST_HEAD(¬ifier->done);
|
|
|
|
mutex_lock(&list_lock);
|
|
|
|
list_for_each_entry(asd, ¬ifier->asd_list, asd_list) {
|
|
ret = v4l2_async_nf_asd_valid(notifier, asd, i++);
|
|
if (ret)
|
|
goto err_unlock;
|
|
|
|
list_add_tail(&asd->list, ¬ifier->waiting);
|
|
}
|
|
|
|
ret = v4l2_async_nf_try_all_subdevs(notifier);
|
|
if (ret < 0)
|
|
goto err_unbind;
|
|
|
|
ret = v4l2_async_nf_try_complete(notifier);
|
|
if (ret < 0)
|
|
goto err_unbind;
|
|
|
|
/* Keep also completed notifiers on the list */
|
|
list_add(¬ifier->list, ¬ifier_list);
|
|
|
|
mutex_unlock(&list_lock);
|
|
|
|
return 0;
|
|
|
|
err_unbind:
|
|
/*
|
|
* On failure, unbind all sub-devices registered through this notifier.
|
|
*/
|
|
v4l2_async_nf_unbind_all_subdevs(notifier);
|
|
|
|
err_unlock:
|
|
mutex_unlock(&list_lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int v4l2_async_nf_register(struct v4l2_device *v4l2_dev,
|
|
struct v4l2_async_notifier *notifier)
|
|
{
|
|
int ret;
|
|
|
|
if (WARN_ON(!v4l2_dev || notifier->sd))
|
|
return -EINVAL;
|
|
|
|
notifier->v4l2_dev = v4l2_dev;
|
|
|
|
ret = __v4l2_async_nf_register(notifier);
|
|
if (ret)
|
|
notifier->v4l2_dev = NULL;
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(v4l2_async_nf_register);
|
|
|
|
int v4l2_async_subdev_nf_register(struct v4l2_subdev *sd,
|
|
struct v4l2_async_notifier *notifier)
|
|
{
|
|
int ret;
|
|
|
|
if (WARN_ON(!sd || notifier->v4l2_dev))
|
|
return -EINVAL;
|
|
|
|
notifier->sd = sd;
|
|
|
|
ret = __v4l2_async_nf_register(notifier);
|
|
if (ret)
|
|
notifier->sd = NULL;
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(v4l2_async_subdev_nf_register);
|
|
|
|
static void
|
|
__v4l2_async_nf_unregister(struct v4l2_async_notifier *notifier)
|
|
{
|
|
if (!notifier || (!notifier->v4l2_dev && !notifier->sd))
|
|
return;
|
|
|
|
v4l2_async_nf_unbind_all_subdevs(notifier);
|
|
|
|
notifier->sd = NULL;
|
|
notifier->v4l2_dev = NULL;
|
|
|
|
list_del(¬ifier->list);
|
|
}
|
|
|
|
void v4l2_async_nf_unregister(struct v4l2_async_notifier *notifier)
|
|
{
|
|
mutex_lock(&list_lock);
|
|
|
|
__v4l2_async_nf_unregister(notifier);
|
|
|
|
mutex_unlock(&list_lock);
|
|
}
|
|
EXPORT_SYMBOL(v4l2_async_nf_unregister);
|
|
|
|
static void __v4l2_async_nf_cleanup(struct v4l2_async_notifier *notifier)
|
|
{
|
|
struct v4l2_async_subdev *asd, *tmp;
|
|
|
|
if (!notifier || !notifier->asd_list.next)
|
|
return;
|
|
|
|
list_for_each_entry_safe(asd, tmp, ¬ifier->asd_list, asd_list) {
|
|
switch (asd->match_type) {
|
|
case V4L2_ASYNC_MATCH_FWNODE:
|
|
fwnode_handle_put(asd->match.fwnode);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
list_del(&asd->asd_list);
|
|
kfree(asd);
|
|
}
|
|
}
|
|
|
|
void v4l2_async_nf_cleanup(struct v4l2_async_notifier *notifier)
|
|
{
|
|
mutex_lock(&list_lock);
|
|
|
|
__v4l2_async_nf_cleanup(notifier);
|
|
|
|
mutex_unlock(&list_lock);
|
|
}
|
|
EXPORT_SYMBOL_GPL(v4l2_async_nf_cleanup);
|
|
|
|
int __v4l2_async_nf_add_subdev(struct v4l2_async_notifier *notifier,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
int ret;
|
|
|
|
mutex_lock(&list_lock);
|
|
|
|
ret = v4l2_async_nf_asd_valid(notifier, asd, -1);
|
|
if (ret)
|
|
goto unlock;
|
|
|
|
list_add_tail(&asd->asd_list, ¬ifier->asd_list);
|
|
|
|
unlock:
|
|
mutex_unlock(&list_lock);
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_GPL(__v4l2_async_nf_add_subdev);
|
|
|
|
struct v4l2_async_subdev *
|
|
__v4l2_async_nf_add_fwnode(struct v4l2_async_notifier *notifier,
|
|
struct fwnode_handle *fwnode,
|
|
unsigned int asd_struct_size)
|
|
{
|
|
struct v4l2_async_subdev *asd;
|
|
int ret;
|
|
|
|
asd = kzalloc(asd_struct_size, GFP_KERNEL);
|
|
if (!asd)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
asd->match_type = V4L2_ASYNC_MATCH_FWNODE;
|
|
asd->match.fwnode = fwnode_handle_get(fwnode);
|
|
|
|
ret = __v4l2_async_nf_add_subdev(notifier, asd);
|
|
if (ret) {
|
|
fwnode_handle_put(fwnode);
|
|
kfree(asd);
|
|
return ERR_PTR(ret);
|
|
}
|
|
|
|
return asd;
|
|
}
|
|
EXPORT_SYMBOL_GPL(__v4l2_async_nf_add_fwnode);
|
|
|
|
struct v4l2_async_subdev *
|
|
__v4l2_async_nf_add_fwnode_remote(struct v4l2_async_notifier *notif,
|
|
struct fwnode_handle *endpoint,
|
|
unsigned int asd_struct_size)
|
|
{
|
|
struct v4l2_async_subdev *asd;
|
|
struct fwnode_handle *remote;
|
|
|
|
remote = fwnode_graph_get_remote_port_parent(endpoint);
|
|
if (!remote)
|
|
return ERR_PTR(-ENOTCONN);
|
|
|
|
asd = __v4l2_async_nf_add_fwnode(notif, remote, asd_struct_size);
|
|
/*
|
|
* Calling __v4l2_async_nf_add_fwnode grabs a refcount,
|
|
* so drop the one we got in fwnode_graph_get_remote_port_parent.
|
|
*/
|
|
fwnode_handle_put(remote);
|
|
return asd;
|
|
}
|
|
EXPORT_SYMBOL_GPL(__v4l2_async_nf_add_fwnode_remote);
|
|
|
|
struct v4l2_async_subdev *
|
|
__v4l2_async_nf_add_i2c(struct v4l2_async_notifier *notifier, int adapter_id,
|
|
unsigned short address, unsigned int asd_struct_size)
|
|
{
|
|
struct v4l2_async_subdev *asd;
|
|
int ret;
|
|
|
|
asd = kzalloc(asd_struct_size, GFP_KERNEL);
|
|
if (!asd)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
asd->match_type = V4L2_ASYNC_MATCH_I2C;
|
|
asd->match.i2c.adapter_id = adapter_id;
|
|
asd->match.i2c.address = address;
|
|
|
|
ret = __v4l2_async_nf_add_subdev(notifier, asd);
|
|
if (ret) {
|
|
kfree(asd);
|
|
return ERR_PTR(ret);
|
|
}
|
|
|
|
return asd;
|
|
}
|
|
EXPORT_SYMBOL_GPL(__v4l2_async_nf_add_i2c);
|
|
|
|
int v4l2_async_register_subdev(struct v4l2_subdev *sd)
|
|
{
|
|
struct v4l2_async_notifier *subdev_notifier;
|
|
struct v4l2_async_notifier *notifier;
|
|
int ret;
|
|
|
|
/*
|
|
* No reference taken. The reference is held by the device
|
|
* (struct v4l2_subdev.dev), and async sub-device does not
|
|
* exist independently of the device at any point of time.
|
|
*/
|
|
if (!sd->fwnode && sd->dev)
|
|
sd->fwnode = dev_fwnode(sd->dev);
|
|
|
|
mutex_lock(&list_lock);
|
|
|
|
INIT_LIST_HEAD(&sd->async_list);
|
|
|
|
list_for_each_entry(notifier, ¬ifier_list, list) {
|
|
struct v4l2_device *v4l2_dev =
|
|
v4l2_async_nf_find_v4l2_dev(notifier);
|
|
struct v4l2_async_subdev *asd;
|
|
|
|
if (!v4l2_dev)
|
|
continue;
|
|
|
|
asd = v4l2_async_find_match(notifier, sd);
|
|
if (!asd)
|
|
continue;
|
|
|
|
ret = v4l2_async_match_notify(notifier, v4l2_dev, sd, asd);
|
|
if (ret)
|
|
goto err_unbind;
|
|
|
|
ret = v4l2_async_nf_try_complete(notifier);
|
|
if (ret)
|
|
goto err_unbind;
|
|
|
|
goto out_unlock;
|
|
}
|
|
|
|
/* None matched, wait for hot-plugging */
|
|
list_add(&sd->async_list, &subdev_list);
|
|
|
|
out_unlock:
|
|
mutex_unlock(&list_lock);
|
|
|
|
return 0;
|
|
|
|
err_unbind:
|
|
/*
|
|
* Complete failed. Unbind the sub-devices bound through registering
|
|
* this async sub-device.
|
|
*/
|
|
subdev_notifier = v4l2_async_find_subdev_notifier(sd);
|
|
if (subdev_notifier)
|
|
v4l2_async_nf_unbind_all_subdevs(subdev_notifier);
|
|
|
|
if (sd->asd)
|
|
v4l2_async_nf_call_unbind(notifier, sd, sd->asd);
|
|
v4l2_async_cleanup(sd);
|
|
|
|
mutex_unlock(&list_lock);
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(v4l2_async_register_subdev);
|
|
|
|
void v4l2_async_unregister_subdev(struct v4l2_subdev *sd)
|
|
{
|
|
if (!sd->async_list.next)
|
|
return;
|
|
|
|
mutex_lock(&list_lock);
|
|
|
|
__v4l2_async_nf_unregister(sd->subdev_notifier);
|
|
__v4l2_async_nf_cleanup(sd->subdev_notifier);
|
|
kfree(sd->subdev_notifier);
|
|
sd->subdev_notifier = NULL;
|
|
|
|
if (sd->asd) {
|
|
struct v4l2_async_notifier *notifier = sd->notifier;
|
|
|
|
list_add(&sd->asd->list, ¬ifier->waiting);
|
|
|
|
v4l2_async_nf_call_unbind(notifier, sd, sd->asd);
|
|
}
|
|
|
|
v4l2_async_cleanup(sd);
|
|
|
|
mutex_unlock(&list_lock);
|
|
}
|
|
EXPORT_SYMBOL(v4l2_async_unregister_subdev);
|
|
|
|
static void print_waiting_subdev(struct seq_file *s,
|
|
struct v4l2_async_subdev *asd)
|
|
{
|
|
switch (asd->match_type) {
|
|
case V4L2_ASYNC_MATCH_I2C:
|
|
seq_printf(s, " [i2c] dev=%d-%04x\n", asd->match.i2c.adapter_id,
|
|
asd->match.i2c.address);
|
|
break;
|
|
case V4L2_ASYNC_MATCH_FWNODE: {
|
|
struct fwnode_handle *devnode, *fwnode = asd->match.fwnode;
|
|
|
|
devnode = fwnode_graph_is_endpoint(fwnode) ?
|
|
fwnode_graph_get_port_parent(fwnode) :
|
|
fwnode_handle_get(fwnode);
|
|
|
|
seq_printf(s, " [fwnode] dev=%s, node=%pfw\n",
|
|
devnode->dev ? dev_name(devnode->dev) : "nil",
|
|
fwnode);
|
|
|
|
fwnode_handle_put(devnode);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
static const char *
|
|
v4l2_async_nf_name(struct v4l2_async_notifier *notifier)
|
|
{
|
|
if (notifier->v4l2_dev)
|
|
return notifier->v4l2_dev->name;
|
|
else if (notifier->sd)
|
|
return notifier->sd->name;
|
|
else
|
|
return "nil";
|
|
}
|
|
|
|
static int pending_subdevs_show(struct seq_file *s, void *data)
|
|
{
|
|
struct v4l2_async_notifier *notif;
|
|
struct v4l2_async_subdev *asd;
|
|
|
|
mutex_lock(&list_lock);
|
|
|
|
list_for_each_entry(notif, ¬ifier_list, list) {
|
|
seq_printf(s, "%s:\n", v4l2_async_nf_name(notif));
|
|
list_for_each_entry(asd, ¬if->waiting, list)
|
|
print_waiting_subdev(s, asd);
|
|
}
|
|
|
|
mutex_unlock(&list_lock);
|
|
|
|
return 0;
|
|
}
|
|
DEFINE_SHOW_ATTRIBUTE(pending_subdevs);
|
|
|
|
static struct dentry *v4l2_async_debugfs_dir;
|
|
|
|
static int __init v4l2_async_init(void)
|
|
{
|
|
v4l2_async_debugfs_dir = debugfs_create_dir("v4l2-async", NULL);
|
|
debugfs_create_file("pending_async_subdevices", 0444,
|
|
v4l2_async_debugfs_dir, NULL,
|
|
&pending_subdevs_fops);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void __exit v4l2_async_exit(void)
|
|
{
|
|
debugfs_remove_recursive(v4l2_async_debugfs_dir);
|
|
}
|
|
|
|
subsys_initcall(v4l2_async_init);
|
|
module_exit(v4l2_async_exit);
|
|
|
|
MODULE_AUTHOR("Guennadi Liakhovetski <g.liakhovetski@gmx.de>");
|
|
MODULE_AUTHOR("Sakari Ailus <sakari.ailus@linux.intel.com>");
|
|
MODULE_AUTHOR("Ezequiel Garcia <ezequiel@collabora.com>");
|
|
MODULE_LICENSE("GPL");
|