vmwgfx was its only user and always sets it to the same..

Signed-off-by: Maarten Lankhorst <maarten.lankhorst at canonical.com>
---
 drivers/gpu/drm/ttm/ttm_bo.c           |   13 ++++---------
 drivers/gpu/drm/ttm/ttm_bo_util.c      |    1 -
 drivers/gpu/drm/ttm/ttm_execbuf_util.c |    1 -
 include/drm/ttm/ttm_bo_api.h           |    2 --
 include/drm/ttm/ttm_execbuf_util.h     |    3 ---
 5 files changed, 4 insertions(+), 16 deletions(-)

diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c
index d9d8541..7673a7f 100644
--- a/drivers/gpu/drm/ttm/ttm_bo.c
+++ b/drivers/gpu/drm/ttm/ttm_bo.c
@@ -500,7 +500,6 @@ static void ttm_bo_cleanup_refs_or_queue(struct 
ttm_buffer_object *bo)
        struct ttm_bo_global *glob = bo->glob;
        struct ttm_bo_driver *driver;
        void *sync_obj = NULL;
-       void *sync_obj_arg;
        int put_count;
        int ret;

@@ -536,7 +535,6 @@ queue:
        driver = bdev->driver;
        if (bo->sync_obj)
                sync_obj = driver->sync_obj_ref(bo->sync_obj);
-       sync_obj_arg = bo->sync_obj_arg;

        kref_get(&bo->list_kref);
        list_add_tail(&bo->ddestroy, &bdev->ddestroy);
@@ -544,7 +542,7 @@ queue:
        spin_unlock(&bdev->fence_lock);

        if (sync_obj) {
-               driver->sync_obj_flush(sync_obj, sync_obj_arg);
+               driver->sync_obj_flush(sync_obj, NULL);
                driver->sync_obj_unref(&sync_obj);
        }
        schedule_delayed_work(&bdev->wq,
@@ -1705,7 +1703,6 @@ int ttm_bo_wait(struct ttm_buffer_object *bo,
        struct ttm_bo_driver *driver = bo->bdev->driver;
        struct ttm_bo_device *bdev = bo->bdev;
        void *sync_obj;
-       void *sync_obj_arg;
        int ret = 0;

        if (likely(bo->sync_obj == NULL))
@@ -1713,7 +1710,7 @@ int ttm_bo_wait(struct ttm_buffer_object *bo,

        while (bo->sync_obj) {

-               if (driver->sync_obj_signaled(bo->sync_obj, bo->sync_obj_arg)) {
+               if (driver->sync_obj_signaled(bo->sync_obj, NULL)) {
                        void *tmp_obj = bo->sync_obj;
                        bo->sync_obj = NULL;
                        clear_bit(TTM_BO_PRIV_FLAG_MOVING, &bo->priv_flags);
@@ -1727,9 +1724,8 @@ int ttm_bo_wait(struct ttm_buffer_object *bo,
                        return -EBUSY;

                sync_obj = driver->sync_obj_ref(bo->sync_obj);
-               sync_obj_arg = bo->sync_obj_arg;
                spin_unlock(&bdev->fence_lock);
-               ret = driver->sync_obj_wait(sync_obj, sync_obj_arg,
+               ret = driver->sync_obj_wait(sync_obj, NULL,
                                            lazy, interruptible);
                if (unlikely(ret != 0)) {
                        driver->sync_obj_unref(&sync_obj);
@@ -1737,8 +1733,7 @@ int ttm_bo_wait(struct ttm_buffer_object *bo,
                        return ret;
                }
                spin_lock(&bdev->fence_lock);
-               if (likely(bo->sync_obj == sync_obj &&
-                          bo->sync_obj_arg == sync_obj_arg)) {
+               if (likely(bo->sync_obj == sync_obj)) {
                        void *tmp_obj = bo->sync_obj;
                        bo->sync_obj = NULL;
                        clear_bit(TTM_BO_PRIV_FLAG_MOVING,
diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c 
b/drivers/gpu/drm/ttm/ttm_bo_util.c
index c521a82..2fb810c 100644
--- a/drivers/gpu/drm/ttm/ttm_bo_util.c
+++ b/drivers/gpu/drm/ttm/ttm_bo_util.c
@@ -629,7 +629,6 @@ int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
                bo->sync_obj = NULL;
        }
        bo->sync_obj = driver->sync_obj_ref(sync_obj);
-       bo->sync_obj_arg = sync_obj_arg;
        if (evict) {
                ret = ttm_bo_wait(bo, false, false, false);
                spin_unlock(&bdev->fence_lock);
diff --git a/drivers/gpu/drm/ttm/ttm_execbuf_util.c 
b/drivers/gpu/drm/ttm/ttm_execbuf_util.c
index e244d92..721886e 100644
--- a/drivers/gpu/drm/ttm/ttm_execbuf_util.c
+++ b/drivers/gpu/drm/ttm/ttm_execbuf_util.c
@@ -214,7 +214,6 @@ void ttm_eu_fence_buffer_objects(struct list_head *list, 
void *sync_obj)
                bo = entry->bo;
                entry->old_sync_obj = bo->sync_obj;
                bo->sync_obj = driver->sync_obj_ref(sync_obj);
-               bo->sync_obj_arg = entry->new_sync_obj_arg;
                ttm_bo_unreserve_locked(bo);
                entry->reserved = false;
        }
diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h
index 1d71f6b..816d9b1 100644
--- a/include/drm/ttm/ttm_bo_api.h
+++ b/include/drm/ttm/ttm_bo_api.h
@@ -172,7 +172,6 @@ struct ttm_tt;
  * @seq_valid: The value of @val_seq is valid. This value is protected by
  * the bo_device::lru_lock.
  * @reserved: Deadlock-free lock used for synchronization state transitions.
- * @sync_obj_arg: Opaque argument to synchronization object function.
  * @sync_obj: Pointer to a synchronization object.
  * @priv_flags: Flags describing buffer object internal state.
  * @vm_rb: Rb node for the vm rb tree.
@@ -249,7 +248,6 @@ struct ttm_buffer_object {
         * checking NULL while reserved but not holding the mentioned lock.
         */

-       void *sync_obj_arg;
        void *sync_obj;
        unsigned long priv_flags;

diff --git a/include/drm/ttm/ttm_execbuf_util.h 
b/include/drm/ttm/ttm_execbuf_util.h
index 1926cae..547e19f 100644
--- a/include/drm/ttm/ttm_execbuf_util.h
+++ b/include/drm/ttm/ttm_execbuf_util.h
@@ -39,8 +39,6 @@
  *
  * @head:           list head for thread-private list.
  * @bo:             refcounted buffer object pointer.
- * @new_sync_obj_arg: New sync_obj_arg for @bo, to be used once
- * adding a new sync object.
  * @reserved:       Indicates whether @bo has been reserved for validation.
  * @removed:        Indicates whether @bo has been removed from lru lists.
  * @put_count:      Number of outstanding references on bo::list_kref.
@@ -50,7 +48,6 @@
 struct ttm_validate_buffer {
        struct list_head head;
        struct ttm_buffer_object *bo;
-       void *new_sync_obj_arg;
        bool reserved;
        bool removed;
        int put_count;

Reply via email to