From patchwork Wed Oct 24 13:40:23 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Sharat Masetty X-Patchwork-Id: 10654627 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id A0CA714BB for ; Wed, 24 Oct 2018 13:40:40 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 86FCE2A87C for ; Wed, 24 Oct 2018 13:40:40 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id 848572A875; Wed, 24 Oct 2018 13:40:40 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-7.7 required=2.0 tests=BAYES_00,DKIM_INVALID, DKIM_SIGNED,MAILING_LIST_MULTI,RCVD_IN_DNSWL_HI autolearn=ham version=3.3.1 Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id E30172A8A6 for ; Wed, 24 Oct 2018 13:40:39 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726865AbeJXWIs (ORCPT ); Wed, 24 Oct 2018 18:08:48 -0400 Received: from smtp.codeaurora.org ([198.145.29.96]:46010 "EHLO smtp.codeaurora.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726449AbeJXWIs (ORCPT ); Wed, 24 Oct 2018 18:08:48 -0400 Received: by smtp.codeaurora.org (Postfix, from userid 1000) id 46D6560C54; Wed, 24 Oct 2018 13:40:38 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=codeaurora.org; s=default; t=1540388438; bh=e6A0LdXVVEwzMS7rM2ifcq0rjR4+Hh9PIkeQlyYJUYA=; h=From:To:Cc:Subject:Date:From; b=Uu4rs608U8Bx2dhJ64dayOfrBD6O5H4D5NgM1od/ChDTmWz6KcIjzxIRJIrdvU1sG 7I3tjH2zizgAQOyaDNq1n4N2bXAYj4TDO1qRGwAVALBAPcbiVrbd960sNBNWzfKu3S aNw4XjoeqgLUtNKlJDTRl7ZlhoKlOHT/hpPjWRIw= Received: from smasetty-linux.qualcomm.com (blr-c-bdr-fw-01_globalnat_allzones-outside.qualcomm.com [103.229.19.19]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-SHA256 (128/128 bits)) (No client certificate requested) (Authenticated sender: smasetty@smtp.codeaurora.org) by smtp.codeaurora.org (Postfix) with ESMTPSA id A042B60316; Wed, 24 Oct 2018 13:40:35 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=codeaurora.org; s=default; t=1540388437; bh=e6A0LdXVVEwzMS7rM2ifcq0rjR4+Hh9PIkeQlyYJUYA=; h=From:To:Cc:Subject:Date:From; b=V2q5C2cah8nHJMB5EYiY35jpnr+hj6mMVdBc/rR+hXJgigameiDwKOPtGaoM15PqU TCXV8e2QuhrmQwmfBZeDgHPdBOKAgDIvnRLkH5rCOPsscWGpRrjReeRihx1IyPC3aX 1Y3hYFVoNSDB+wxmFaUYMAX0rcPCqyMcd09BVLE0= DMARC-Filter: OpenDMARC Filter v1.3.2 smtp.codeaurora.org A042B60316 Authentication-Results: pdx-caf-mail.web.codeaurora.org; dmarc=none (p=none dis=none) header.from=codeaurora.org Authentication-Results: pdx-caf-mail.web.codeaurora.org; spf=none smtp.mailfrom=smasetty@codeaurora.org From: Sharat Masetty To: freedreno@lists.freedesktop.org Cc: dri-devel@lists.freedesktop.org, linux-arm-msm@vger.kernel.org, jcrouse@codeaurora.org, Christian.Koenig@amd.com, Sharat Masetty Subject: [RFC PATCH] drm/scheduler: Add drm_sched_job_cleanup Date: Wed, 24 Oct 2018 19:10:23 +0530 Message-Id: <1540388423-30236-1-git-send-email-smasetty@codeaurora.org> X-Mailer: git-send-email 1.9.1 Sender: linux-arm-msm-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-arm-msm@vger.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP Hi, Can you please review this and share your thoughts on this approach. We primarily need a clean way to release the finished fence in case the job was not pushed to the queue. So, this patch adds the new API to clean up the resources allocated in sched_job_init() Additionally I also move the fence_put(finished_fence) from the scheduler to the drivers handler of free_job(). The drivers get to use this new API. This is done so that the layer creating the sched object is the one freeing up the resources as well. Signed-off-by: Sharat Masetty --- drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c | 2 +- drivers/gpu/drm/amd/amdgpu/amdgpu_job.c | 2 ++ drivers/gpu/drm/etnaviv/etnaviv_sched.c | 2 ++ drivers/gpu/drm/msm/msm_sched.c | 8 +++----- drivers/gpu/drm/scheduler/gpu_scheduler.c | 11 +++++++++-- drivers/gpu/drm/v3d/v3d_sched.c | 2 ++ include/drm/gpu_scheduler.h | 1 + 7 files changed, 20 insertions(+), 8 deletions(-) diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c index 9c85a90..9a83152 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c @@ -1198,7 +1198,7 @@ static int amdgpu_cs_submit(struct amdgpu_cs_parser *p, r = amdgpu_ctx_add_fence(p->ctx, ring, p->fence, &seq); if (r) { dma_fence_put(p->fence); - dma_fence_put(&job->base.s_fence->finished); + drm_sched_job_cleanup(&job->base); amdgpu_job_free(job); amdgpu_mn_unlock(p->mn); return r; diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c index 2bd5676..5d252d4 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c @@ -100,6 +100,8 @@ static void amdgpu_job_free_cb(struct drm_sched_job *s_job) { struct amdgpu_job *job = container_of(s_job, struct amdgpu_job, base); + drm_sched_job_cleanup(s_job); + amdgpu_ring_priority_put(job->ring, s_job->s_priority); dma_fence_put(job->fence); amdgpu_sync_free(&job->sync); diff --git a/drivers/gpu/drm/etnaviv/etnaviv_sched.c b/drivers/gpu/drm/etnaviv/etnaviv_sched.c index 50d6b88..30398c5 100644 --- a/drivers/gpu/drm/etnaviv/etnaviv_sched.c +++ b/drivers/gpu/drm/etnaviv/etnaviv_sched.c @@ -127,6 +127,8 @@ static void etnaviv_sched_free_job(struct drm_sched_job *sched_job) { struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); + drm_sched_job_cleanup(sched_job); + etnaviv_submit_put(submit); } diff --git a/drivers/gpu/drm/msm/msm_sched.c b/drivers/gpu/drm/msm/msm_sched.c index 70b7713..748acc6 100644 --- a/drivers/gpu/drm/msm/msm_sched.c +++ b/drivers/gpu/drm/msm/msm_sched.c @@ -235,6 +235,8 @@ static void msm_sched_free_job(struct drm_sched_job *sched_job) struct msm_gpu *gpu = submit->gpu; int i; + drm_sched_job_cleanup(sched_job); + mutex_lock(&gpu->dev->struct_mutex); for (i = 0; i < submit->nr_bos; i++) { @@ -300,11 +302,7 @@ int msm_sched_job_init(struct drm_sched_job *sched_job) mutex_unlock(&ring->fence_idr_lock); if (submit->out_fence_id < 0) { - /* - * TODO: The scheduler's finished fence leaks here since the - * job will not be pushed to the queue. Need to update scheduler - * to fix this cleanly(?) - */ + drm_sched_job_cleanup(sched_job); dma_fence_put(submit->out_fence); submit->out_fence = NULL; return -ENOMEM; diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler.c b/drivers/gpu/drm/scheduler/gpu_scheduler.c index f5534ff..bbf9315 100644 --- a/drivers/gpu/drm/scheduler/gpu_scheduler.c +++ b/drivers/gpu/drm/scheduler/gpu_scheduler.c @@ -214,7 +214,6 @@ static void drm_sched_entity_kill_jobs_cb(struct dma_fence *f, finish_cb); drm_sched_fence_finished(job->s_fence); WARN_ON(job->s_fence->parent); - dma_fence_put(&job->s_fence->finished); job->sched->ops->free_job(job); } @@ -480,7 +479,6 @@ static void drm_sched_job_finish(struct work_struct *work) drm_sched_queue_delayed_work(next); } spin_unlock(&sched->job_list_lock); - dma_fence_put(&s_job->s_fence->finished); sched->ops->free_job(s_job); } @@ -636,6 +634,15 @@ int drm_sched_job_init(struct drm_sched_job *job, EXPORT_SYMBOL(drm_sched_job_init); /** + * Cleanup sched_job resources + */ +void drm_sched_job_cleanup(struct drm_sched_job *job) +{ + dma_fence_put(&job->s_fence->finished); +} +EXPORT_SYMBOL(drm_sched_job_cleanup); + +/** * Return ture if we can push more jobs to the hw. */ static bool drm_sched_ready(struct drm_gpu_scheduler *sched) diff --git a/drivers/gpu/drm/v3d/v3d_sched.c b/drivers/gpu/drm/v3d/v3d_sched.c index b07bece..4abd73c 100644 --- a/drivers/gpu/drm/v3d/v3d_sched.c +++ b/drivers/gpu/drm/v3d/v3d_sched.c @@ -35,6 +35,8 @@ { struct v3d_job *job = to_v3d_job(sched_job); + drm_sched_job_cleanup(sched_job); + v3d_exec_put(job->exec); } diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h index 5c59c38..2ec3ddf 100644 --- a/include/drm/gpu_scheduler.h +++ b/include/drm/gpu_scheduler.h @@ -213,6 +213,7 @@ int drm_sched_job_init(struct drm_sched_job *job, struct drm_gpu_scheduler *sched, struct drm_sched_entity *entity, void *owner); +void drm_sched_job_cleanup(struct drm_sched_job *job); void drm_sched_hw_job_reset(struct drm_gpu_scheduler *sched, struct drm_sched_job *job); void drm_sched_job_recovery(struct drm_gpu_scheduler *sched);