sync with OpenBSD -current
This commit is contained in:
parent
ee68147dcd
commit
1cefe29c7e
1651 changed files with 283292 additions and 68089 deletions
|
@ -42,7 +42,6 @@
|
|||
#include "amdgpu_reset.h"
|
||||
|
||||
/*
|
||||
* Fences
|
||||
* Fences mark an event in the GPUs pipeline and are used
|
||||
* for GPU/CPU synchronization. When the fence is written,
|
||||
* it is expected that all buffers associated with that fence
|
||||
|
@ -55,6 +54,7 @@ struct amdgpu_fence {
|
|||
|
||||
/* RB, DMA, etc. */
|
||||
struct amdgpu_ring *ring;
|
||||
ktime_t start_timestamp;
|
||||
};
|
||||
|
||||
static struct pool amdgpu_fence_slab;
|
||||
|
@ -148,7 +148,7 @@ static u32 amdgpu_fence_read(struct amdgpu_ring *ring)
|
|||
* Returns 0 on success, -ENOMEM on failure.
|
||||
*/
|
||||
int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **f, struct amdgpu_job *job,
|
||||
unsigned flags)
|
||||
unsigned int flags)
|
||||
{
|
||||
struct amdgpu_device *adev = ring->adev;
|
||||
struct dma_fence *fence;
|
||||
|
@ -186,11 +186,11 @@ int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **f, struct amd
|
|||
adev->fence_context + ring->idx, seq);
|
||||
/* Against remove in amdgpu_job_{free, free_cb} */
|
||||
dma_fence_get(fence);
|
||||
}
|
||||
else
|
||||
} else {
|
||||
dma_fence_init(fence, &amdgpu_fence_ops,
|
||||
&ring->fence_drv.lock,
|
||||
adev->fence_context + ring->idx, seq);
|
||||
}
|
||||
}
|
||||
|
||||
amdgpu_ring_emit_fence(ring, ring->fence_drv.gpu_addr,
|
||||
|
@ -212,6 +212,8 @@ int amdgpu_fence_emit(struct amdgpu_ring *ring, struct dma_fence **f, struct amd
|
|||
}
|
||||
}
|
||||
|
||||
to_amdgpu_fence(fence)->start_timestamp = ktime_get();
|
||||
|
||||
/* This function can't be called concurrently anyway, otherwise
|
||||
* emitting the fence would mess up the hardware ring buffer.
|
||||
*/
|
||||
|
@ -386,14 +388,11 @@ signed long amdgpu_fence_wait_polling(struct amdgpu_ring *ring,
|
|||
uint32_t wait_seq,
|
||||
signed long timeout)
|
||||
{
|
||||
uint32_t seq;
|
||||
|
||||
do {
|
||||
seq = amdgpu_fence_read(ring);
|
||||
udelay(5);
|
||||
timeout -= 5;
|
||||
} while ((int32_t)(wait_seq - seq) > 0 && timeout > 0);
|
||||
|
||||
while ((int32_t)(wait_seq - amdgpu_fence_read(ring)) > 0 && timeout > 0) {
|
||||
udelay(2);
|
||||
timeout -= 2;
|
||||
}
|
||||
return timeout > 0 ? timeout : 0;
|
||||
}
|
||||
/**
|
||||
|
@ -405,7 +404,7 @@ signed long amdgpu_fence_wait_polling(struct amdgpu_ring *ring,
|
|||
* Returns the number of emitted fences on the ring. Used by the
|
||||
* dynpm code to ring track activity.
|
||||
*/
|
||||
unsigned amdgpu_fence_count_emitted(struct amdgpu_ring *ring)
|
||||
unsigned int amdgpu_fence_count_emitted(struct amdgpu_ring *ring)
|
||||
{
|
||||
uint64_t emitted;
|
||||
|
||||
|
@ -418,6 +417,57 @@ unsigned amdgpu_fence_count_emitted(struct amdgpu_ring *ring)
|
|||
return lower_32_bits(emitted);
|
||||
}
|
||||
|
||||
/**
|
||||
* amdgpu_fence_last_unsignaled_time_us - the time fence emitted until now
|
||||
* @ring: ring the fence is associated with
|
||||
*
|
||||
* Find the earliest fence unsignaled until now, calculate the time delta
|
||||
* between the time fence emitted and now.
|
||||
*/
|
||||
u64 amdgpu_fence_last_unsignaled_time_us(struct amdgpu_ring *ring)
|
||||
{
|
||||
struct amdgpu_fence_driver *drv = &ring->fence_drv;
|
||||
struct dma_fence *fence;
|
||||
uint32_t last_seq, sync_seq;
|
||||
|
||||
last_seq = atomic_read(&ring->fence_drv.last_seq);
|
||||
sync_seq = READ_ONCE(ring->fence_drv.sync_seq);
|
||||
if (last_seq == sync_seq)
|
||||
return 0;
|
||||
|
||||
++last_seq;
|
||||
last_seq &= drv->num_fences_mask;
|
||||
fence = drv->fences[last_seq];
|
||||
if (!fence)
|
||||
return 0;
|
||||
|
||||
return ktime_us_delta(ktime_get(),
|
||||
to_amdgpu_fence(fence)->start_timestamp);
|
||||
}
|
||||
|
||||
/**
|
||||
* amdgpu_fence_update_start_timestamp - update the timestamp of the fence
|
||||
* @ring: ring the fence is associated with
|
||||
* @seq: the fence seq number to update.
|
||||
* @timestamp: the start timestamp to update.
|
||||
*
|
||||
* The function called at the time the fence and related ib is about to
|
||||
* resubmit to gpu in MCBP scenario. Thus we do not consider race condition
|
||||
* with amdgpu_fence_process to modify the same fence.
|
||||
*/
|
||||
void amdgpu_fence_update_start_timestamp(struct amdgpu_ring *ring, uint32_t seq, ktime_t timestamp)
|
||||
{
|
||||
struct amdgpu_fence_driver *drv = &ring->fence_drv;
|
||||
struct dma_fence *fence;
|
||||
|
||||
seq &= drv->num_fences_mask;
|
||||
fence = drv->fences[seq];
|
||||
if (!fence)
|
||||
return;
|
||||
|
||||
to_amdgpu_fence(fence)->start_timestamp = timestamp;
|
||||
}
|
||||
|
||||
/**
|
||||
* amdgpu_fence_driver_start_ring - make the fence driver
|
||||
* ready for use on the requested ring.
|
||||
|
@ -433,7 +483,7 @@ unsigned amdgpu_fence_count_emitted(struct amdgpu_ring *ring)
|
|||
*/
|
||||
int amdgpu_fence_driver_start_ring(struct amdgpu_ring *ring,
|
||||
struct amdgpu_irq_src *irq_src,
|
||||
unsigned irq_type)
|
||||
unsigned int irq_type)
|
||||
{
|
||||
struct amdgpu_device *adev = ring->adev;
|
||||
uint64_t index;
|
||||
|
@ -443,7 +493,7 @@ int amdgpu_fence_driver_start_ring(struct amdgpu_ring *ring,
|
|||
ring->fence_drv.gpu_addr = ring->fence_gpu_addr;
|
||||
} else {
|
||||
/* put fence directly behind firmware */
|
||||
index = roundup2(adev->uvd.fw->size, 8);
|
||||
index = ALIGN(adev->uvd.fw->size, 8);
|
||||
ring->fence_drv.cpu_addr = adev->uvd.inst[ring->me].cpu_addr + index;
|
||||
ring->fence_drv.gpu_addr = adev->uvd.inst[ring->me].gpu_addr + index;
|
||||
}
|
||||
|
@ -656,6 +706,7 @@ void amdgpu_fence_driver_hw_init(struct amdgpu_device *adev)
|
|||
|
||||
for (i = 0; i < AMDGPU_MAX_RINGS; i++) {
|
||||
struct amdgpu_ring *ring = adev->rings[i];
|
||||
|
||||
if (!ring || !ring->fence_drv.initialized)
|
||||
continue;
|
||||
|
||||
|
@ -697,6 +748,30 @@ void amdgpu_fence_driver_clear_job_fences(struct amdgpu_ring *ring)
|
|||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* amdgpu_fence_driver_set_error - set error code on fences
|
||||
* @ring: the ring which contains the fences
|
||||
* @error: the error code to set
|
||||
*
|
||||
* Set an error code to all the fences pending on the ring.
|
||||
*/
|
||||
void amdgpu_fence_driver_set_error(struct amdgpu_ring *ring, int error)
|
||||
{
|
||||
struct amdgpu_fence_driver *drv = &ring->fence_drv;
|
||||
unsigned long flags;
|
||||
|
||||
spin_lock_irqsave(&drv->lock, flags);
|
||||
for (unsigned int i = 0; i <= drv->num_fences_mask; ++i) {
|
||||
struct dma_fence *fence;
|
||||
|
||||
fence = rcu_dereference_protected(drv->fences[i],
|
||||
lockdep_is_held(&drv->lock));
|
||||
if (fence && !dma_fence_is_signaled_locked(fence))
|
||||
dma_fence_set_error(fence, error);
|
||||
}
|
||||
spin_unlock_irqrestore(&drv->lock, flags);
|
||||
}
|
||||
|
||||
/**
|
||||
* amdgpu_fence_driver_force_completion - force signal latest fence of ring
|
||||
*
|
||||
|
@ -705,6 +780,7 @@ void amdgpu_fence_driver_clear_job_fences(struct amdgpu_ring *ring)
|
|||
*/
|
||||
void amdgpu_fence_driver_force_completion(struct amdgpu_ring *ring)
|
||||
{
|
||||
amdgpu_fence_driver_set_error(ring, -ECANCELED);
|
||||
amdgpu_fence_write(ring, ring->fence_drv.sync_seq);
|
||||
amdgpu_fence_process(ring);
|
||||
}
|
||||
|
@ -843,11 +919,12 @@ static const struct dma_fence_ops amdgpu_job_fence_ops = {
|
|||
#if defined(CONFIG_DEBUG_FS)
|
||||
static int amdgpu_debugfs_fence_info_show(struct seq_file *m, void *unused)
|
||||
{
|
||||
struct amdgpu_device *adev = (struct amdgpu_device *)m->private;
|
||||
struct amdgpu_device *adev = m->private;
|
||||
int i;
|
||||
|
||||
for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
|
||||
struct amdgpu_ring *ring = adev->rings[i];
|
||||
|
||||
if (!ring || !ring->fence_drv.initialized)
|
||||
continue;
|
||||
|
||||
|
@ -921,6 +998,7 @@ static void amdgpu_debugfs_reset_work(struct work_struct *work)
|
|||
reset_work);
|
||||
|
||||
struct amdgpu_reset_context reset_context;
|
||||
|
||||
memset(&reset_context, 0, sizeof(reset_context));
|
||||
|
||||
reset_context.method = AMD_RESET_METHOD_NONE;
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue