Module Name:    src
Committed By:   riastradh
Date:           Mon Aug 27 14:13:16 UTC 2018

Modified Files:
        src/sys/external/bsd/drm2/linux: linux_fence.c

Log Message:
Lotsa kasserts about referencedness of frences.


To generate a diff of this commit:
cvs rdiff -u -r1.5 -r1.6 src/sys/external/bsd/drm2/linux/linux_fence.c

Please note that diffs are not public domain; they are subject to the
copyright notices on the relevant files.

Modified files:

Index: src/sys/external/bsd/drm2/linux/linux_fence.c
diff -u src/sys/external/bsd/drm2/linux/linux_fence.c:1.5 src/sys/external/bsd/drm2/linux/linux_fence.c:1.6
--- src/sys/external/bsd/drm2/linux/linux_fence.c:1.5	Mon Aug 27 14:01:01 2018
+++ src/sys/external/bsd/drm2/linux/linux_fence.c	Mon Aug 27 14:13:16 2018
@@ -1,4 +1,4 @@
-/*	$NetBSD: linux_fence.c,v 1.5 2018/08/27 14:01:01 riastradh Exp $	*/
+/*	$NetBSD: linux_fence.c,v 1.6 2018/08/27 14:13:16 riastradh Exp $	*/
 
 /*-
  * Copyright (c) 2018 The NetBSD Foundation, Inc.
@@ -30,7 +30,7 @@
  */
 
 #include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: linux_fence.c,v 1.5 2018/08/27 14:01:01 riastradh Exp $");
+__KERNEL_RCSID(0, "$NetBSD: linux_fence.c,v 1.6 2018/08/27 14:13:16 riastradh Exp $");
 
 #include <sys/atomic.h>
 #include <sys/condvar.h>
@@ -44,6 +44,19 @@ __KERNEL_RCSID(0, "$NetBSD: linux_fence.
 #include <linux/spinlock.h>
 
 /*
+ * fence_referenced_p(fence)
+ *
+ *	True if fence has a positive reference count.  True after
+ *	fence_init; after the last fence_put, this becomes false.
+ */
+static inline bool
+fence_referenced_p(struct fence *fence)
+{
+
+	return kref_referenced_p(&fence->refcount);
+}
+
+/*
  * fence_init(fence, ops, lock, context, seqno)
  *
  *	Initialize fence.  Caller should call fence_destroy when done,
@@ -74,6 +87,8 @@ void
 fence_destroy(struct fence *fence)
 {
 
+	KASSERT(!fence_referenced_p(fence));
+
 	KASSERT(TAILQ_EMPTY(&fence->f_callbacks));
 	cv_destroy(&fence->f_cv);
 }
@@ -83,6 +98,8 @@ fence_free_cb(struct rcu_head *rcu)
 {
 	struct fence *fence = container_of(rcu, struct fence, f_rcu);
 
+	KASSERT(!fence_referenced_p(fence));
+
 	fence_destroy(fence);
 	kfree(fence);
 }
@@ -103,6 +120,8 @@ void
 fence_free(struct fence *fence)
 {
 
+	KASSERT(!fence_referenced_p(fence));
+
 	call_rcu(&fence->f_rcu, &fence_free_cb);
 }
 
@@ -176,6 +195,8 @@ fence_release(struct kref *refcount)
 {
 	struct fence *fence = container_of(refcount, struct fence, refcount);
 
+	KASSERT(!fence_referenced_p(fence));
+
 	if (fence->ops->release)
 		(*fence->ops->release)(fence);
 	else
@@ -192,6 +213,7 @@ void
 fence_put(struct fence *fence)
 {
 
+	KASSERT(fence_referenced_p(fence));
 	kref_put(&fence->refcount, &fence_release);
 }
 
@@ -210,6 +232,7 @@ static int
 fence_ensure_signal_enabled(struct fence *fence)
 {
 
+	KASSERT(fence_referenced_p(fence));
 	KASSERT(spin_is_locked(fence->lock));
 
 	/* If the fence was already signalled, fail with -ENOENT.  */
@@ -251,6 +274,8 @@ fence_add_callback(struct fence *fence, 
 {
 	int ret;
 
+	KASSERT(fence_referenced_p(fence));
+
 	/* Optimistically try to skip the lock if it's already signalled.  */
 	if (fence->flags & (1u << FENCE_FLAG_SIGNALED_BIT)) {
 		ret = -ENOENT;
@@ -288,6 +313,8 @@ fence_remove_callback(struct fence *fenc
 {
 	bool onqueue;
 
+	KASSERT(fence_referenced_p(fence));
+
 	spin_lock(fence->lock);
 	onqueue = fcb->fcb_onqueue;
 	if (onqueue) {
@@ -311,6 +338,8 @@ void
 fence_enable_sw_signaling(struct fence *fence)
 {
 
+	KASSERT(fence_referenced_p(fence));
+
 	spin_lock(fence->lock);
 	(void)fence_ensure_signal_enabled(fence);
 	spin_unlock(fence->lock);
@@ -330,6 +359,8 @@ fence_is_signaled(struct fence *fence)
 {
 	bool signaled;
 
+	KASSERT(fence_referenced_p(fence));
+
 	spin_lock(fence->lock);
 	signaled = fence_is_signaled_locked(fence);
 	spin_unlock(fence->lock);
@@ -347,6 +378,7 @@ bool
 fence_is_signaled_locked(struct fence *fence)
 {
 
+	KASSERT(fence_referenced_p(fence));
 	KASSERT(spin_is_locked(fence->lock));
 
 	/* Check whether we already set the signalled bit.  */
@@ -383,6 +415,8 @@ fence_signal(struct fence *fence)
 {
 	int ret;
 
+	KASSERT(fence_referenced_p(fence));
+
 	spin_lock(fence->lock);
 	ret = fence_signal_locked(fence);
 	spin_unlock(fence->lock);
@@ -401,6 +435,7 @@ fence_signal_locked(struct fence *fence)
 {
 	struct fence_cb *fcb, *next;
 
+	KASSERT(fence_referenced_p(fence));
 	KASSERT(spin_is_locked(fence->lock));
 
 	/* If it's been signalled, fail; otherwise set the signalled bit.  */
@@ -435,6 +470,8 @@ wait_any_cb(struct fence *fence, struct 
 {
 	struct wait_any *cb = container_of(fcb, struct wait_any, fcb);
 
+	KASSERT(fence_referenced_p(fence));
+
 	mutex_enter(&cb->common->lock);
 	cb->common->done = true;
 	cv_broadcast(&cb->common->cv);
@@ -472,6 +509,7 @@ fence_wait_any_timeout(struct fence **fe
 	/* Add a callback to each of the fences, or stop here if we can't.  */
 	for (i = 0; i < nfences; i++) {
 		cb[i].common = &common;
+		KASSERT(fence_referenced_p(fences[i]));
 		ret = fence_add_callback(fences[i], &cb[i].fcb, &wait_any_cb);
 		if (ret)
 			goto out1;
@@ -559,8 +597,10 @@ long
 fence_wait_timeout(struct fence *fence, bool intr, long timeout)
 {
 
+	KASSERT(fence_referenced_p(fence));
 	KASSERT(timeout >= 0);
 	KASSERT(timeout < MAX_SCHEDULE_TIMEOUT);
+
 	return (*fence->ops->wait)(fence, intr, timeout);
 }
 
@@ -577,6 +617,8 @@ fence_wait(struct fence *fence, bool int
 {
 	long ret;
 
+	KASSERT(fence_referenced_p(fence));
+
 	ret = (*fence->ops->wait)(fence, intr, MAX_SCHEDULE_TIMEOUT);
 	KASSERT(ret != 0);
 
@@ -600,6 +642,7 @@ fence_default_wait(struct fence *fence, 
 	kmutex_t *lock = &fence->lock->sl_lock;
 	long ret = 0;
 
+	KASSERT(fence_referenced_p(fence));
 	KASSERT(timeout >= 0);
 	KASSERT(timeout <= MAX_SCHEDULE_TIMEOUT);
 

Reply via email to