/*
* Copyright (c) 2014-2018, NVIDIA CORPORATION. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#include "fence_gk20a.h"
#include <nvgpu/semaphore.h>
#include <nvgpu/kmem.h>
#include <nvgpu/soc.h>
#include <nvgpu/nvhost.h>
#include <nvgpu/barrier.h>
#include <nvgpu/os_fence.h>
#include <nvgpu/channel.h>
#include "gk20a.h"
struct gk20a_fence_ops {
int (*wait)(struct gk20a_fence *, long timeout);
bool (*is_expired)(struct gk20a_fence *);
void *(*free)(struct nvgpu_ref *);
};
static void gk20a_fence_free(struct nvgpu_ref *ref)
{
struct gk20a_fence *f =
container_of(ref, struct gk20a_fence, ref);
struct gk20a *g = f->g;
if (nvgpu_os_fence_is_initialized(&f->os_fence)) {
f->os_fence.ops->drop_ref(&f->os_fence);
}
if (f->semaphore) {
nvgpu_semaphore_put(f->semaphore);
}
if (f->allocator) {
if (nvgpu_alloc_initialized(f->allocator)) {
nvgpu_free(f->allocator, (u64)(uintptr_t)f);
}
} else {
nvgpu_kfree(g, f);
}
}
void gk20a_fence_put(struct gk20a_fence *f)
{
if (f) {
nvgpu_ref_put(&f->ref, gk20a_fence_free);
}
}
struct gk20a_fence *gk20a_fence_get(struct gk20a_fence *f)
{
if (f) {
nvgpu_ref_get(&f->ref);
}
return f;
}
inline bool gk20a_fence_is_valid(struct gk20a_fence *f)
{
bool valid = f->valid;
nvgpu_smp_rmb();
return valid;
}
int gk20a_fence_install_fd(struct gk20a_fence *f, int fd)
{
if (!f || !gk20a_fence_is_valid(f) ||
!nvgpu_os_fence_is_initialized(&f->os_fence)) {
return -EINVAL;
}
f->os_fence.ops->install_fence(&f->os_fence, fd);
return 0;
}
int gk20a_fence_wait(struct gk20a *g, struct gk20a_fence *f,
unsigned long timeout)
{
if (f && gk20a_fence_is_valid(f)) {
if (!nvgpu_platform_is_silicon(g)) {
timeout = MAX_SCHEDULE_TIMEOUT;
}
return f->ops->wait(f, timeout);
}
return 0;
}
bool gk20a_fence_is_expired(struct gk20a_fence *f)
{
if (f && gk20a_fence_is_valid(f) && f->ops) {
return f->ops->is_expired(f);
} else {
return true;
}
}
int gk20a_alloc_fence_pool(struct channel_gk20a *c, unsigned int count)
{
int err;
size_t size;
struct gk20a_fence *fence_pool = NULL;
size = sizeof(struct gk20a_fence);
if (count <= UINT_MAX / size) {
size = count * size;
fence_pool = nvgpu_vzalloc(c->g, size);
}
if (!fence_pool) {
return -ENOMEM;
}
err = nvgpu_lockless_allocator_init(c->g, &c->fence_allocator,
"fence_pool", (size_t)fence_pool, size,
sizeof(struct gk20a_fence), 0);
if (err) {
goto fail;
}
return 0;
fail:
nvgpu_vfree(c->g, fence_pool);
return err;
}
void gk20a_free_fence_pool(struct channel_gk20a *c)
{
if (nvgpu_alloc_initialized(&c->fence_allocator)) {
struct gk20a_fence *fence_pool;
fence_pool = (struct gk20a_fence *)(uintptr_t)
nvgpu_alloc_base(&c->fence_allocator);
nvgpu_alloc_destroy(&c->fence_allocator);
nvgpu_vfree(c->g, fence_pool);
}
}
struct gk20a_fence *gk20a_alloc_fence(struct channel_gk20a *c)
{
struct gk20a_fence *fence = NULL;
if (channel_gk20a_is_prealloc_enabled(c)) {
if (nvgpu_alloc_initialized(&c->fence_allocator)) {
fence = (struct gk20a_fence *)(uintptr_t)
nvgpu_alloc(&c->fence_allocator,
sizeof(struct gk20a_fence));
/* clear the node and reset the allocator pointer */
if (fence) {
memset(fence, 0, sizeof(*fence));
fence->allocator = &c->fence_allocator;
}
}
} else {
fence = nvgpu_kzalloc(c->g, sizeof(struct gk20a_fence));
}
if (fence) {
nvgpu_ref_init(&fence->ref);
fence->g = c->g;
}
return fence;
}
void gk20a_init_fence(struct gk20a_fence *f,
const struct gk20a_fence_ops *ops,
struct nvgpu_os_fence os_fence)
{
if (!f) {
return;
}
f->ops = ops;
f->syncpt_id = -1;
f->semaphore = NULL;
f->os_fence = os_fence;
}
/* Fences that are backed by GPU semaphores: */
static int nvgpu_semaphore_fence_wait(struct gk20a_fence *f, long timeout)
{
if (!nvgpu_semaphore_is_acquired(f->semaphore)) {
return 0;
}
return NVGPU_COND_WAIT_INTERRUPTIBLE(
f->semaphore_wq,
!nvgpu_semaphore_is_acquired(f->semaphore),
timeout);
}
static bool nvgpu_semaphore_fence_is_expired(struct gk20a_fence *f)
{
return !nvgpu_semaphore_is_acquired(f->semaphore);
}
static const struct gk20a_fence_ops nvgpu_semaphore_fence_ops = {
.wait = &nvgpu_semaphore_fence_wait,
.is_expired = &nvgpu_semaphore_fence_is_expired,
};
/* This function takes ownership of the semaphore as well as the os_fence */
int gk20a_fence_from_semaphore(
struct gk20a_fence *fence_out,
struct nvgpu_semaphore *semaphore,
struct nvgpu_cond *semaphore_wq,
struct nvgpu_os_fence os_fence)
{
struct gk20a_fence *f = fence_out;
gk20a_init_fence(f, &nvgpu_semaphore_fence_ops, os_fence);
if (!f) {
return -EINVAL;
}
f->semaphore = semaphore;
f->semaphore_wq = semaphore_wq;
/* commit previous writes before setting the valid flag */
nvgpu_smp_wmb();
f->valid = true;
return 0;
}
#ifdef CONFIG_TEGRA_GK20A_NVHOST
/* Fences that are backed by host1x syncpoints: */
static int gk20a_syncpt_fence_wait(struct gk20a_fence *f, long timeout)
{
return nvgpu_nvhost_syncpt_wait_timeout_ext(
f->nvhost_dev, f->syncpt_id, f->syncpt_value,
(u32)timeout, NULL, NULL);
}
static bool gk20a_syncpt_fence_is_expired(struct gk20a_fence *f)
{
/*
* In cases we don't register a notifier, we can't expect the
* syncpt value to be updated. For this case, we force a read
* of the value from HW, and then check for expiration.
*/
if (!nvgpu_nvhost_syncpt_is_expired_ext(f->nvhost_dev, f->syncpt_id,
f->syncpt_value)) {
u32 val;
if (!nvgpu_nvhost_syncpt_read_ext_check(f->nvhost_dev,
f->syncpt_id, &val)) {
return nvgpu_nvhost_syncpt_is_expired_ext(
f->nvhost_dev,
f->syncpt_id, f->syncpt_value);
}
}
return true;
}
static const struct gk20a_fence_ops gk20a_syncpt_fence_ops = {
.wait = &gk20a_syncpt_fence_wait,
.is_expired = &gk20a_syncpt_fence_is_expired,
};
/* This function takes the ownership of the os_fence */
int gk20a_fence_from_syncpt(
struct gk20a_fence *fence_out,
struct nvgpu_nvhost_dev *nvhost_dev,
u32 id, u32 value, struct nvgpu_os_fence os_fence)
{
struct gk20a_fence *f = fence_out;
gk20a_init_fence(f, &gk20a_syncpt_fence_ops, os_fence);
if (!f)
return -EINVAL;
f->nvhost_dev = nvhost_dev;
f->syncpt_id = id;
f->syncpt_value = value;
/* commit previous writes before setting the valid flag */
nvgpu_smp_wmb();
f->valid = true;
return 0;
}
#else
int gk20a_fence_from_syncpt(
struct gk20a_fence *fence_out,
struct nvgpu_nvhost_dev *nvhost_dev,
u32 id, u32 value, struct nvgpu_os_fence os_fence)
{
return -EINVAL;
}
#endif