2023-01-25 15:15:43 +11:00
|
|
|
use crate::error;
|
2023-01-24 18:02:27 +11:00
|
|
|
use std::cell::RefCell;
|
|
|
|
use std::marker::PhantomData;
|
2023-02-01 11:23:57 +11:00
|
|
|
use std::ops::Deref;
|
2023-02-02 10:09:34 +11:00
|
|
|
use std::rc::Rc;
|
2023-02-06 13:49:55 +11:00
|
|
|
use bitvec::{bits, bitvec};
|
|
|
|
use bitvec::boxed::BitBox;
|
2023-02-06 10:34:30 +11:00
|
|
|
|
2023-02-06 12:05:22 +11:00
|
|
|
use crate::error::FilterChainError;
|
2023-02-06 08:17:23 +11:00
|
|
|
use windows::Win32::Graphics::Direct3D12::{
|
|
|
|
ID3D12DescriptorHeap, ID3D12Device, D3D12_CPU_DESCRIPTOR_HANDLE, D3D12_DESCRIPTOR_HEAP_DESC,
|
|
|
|
D3D12_DESCRIPTOR_HEAP_FLAG_NONE, D3D12_DESCRIPTOR_HEAP_FLAG_SHADER_VISIBLE,
|
|
|
|
D3D12_DESCRIPTOR_HEAP_TYPE, D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV,
|
|
|
|
D3D12_DESCRIPTOR_HEAP_TYPE_RTV, D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER,
|
|
|
|
D3D12_GPU_DESCRIPTOR_HANDLE,
|
|
|
|
};
|
2023-01-24 18:02:27 +11:00
|
|
|
|
|
|
|
#[const_trait]
|
|
|
|
pub trait D3D12HeapType {
|
|
|
|
fn get_desc(size: usize) -> D3D12_DESCRIPTOR_HEAP_DESC;
|
|
|
|
}
|
|
|
|
|
2023-01-25 15:15:43 +11:00
|
|
|
pub trait D3D12ShaderVisibleHeapType: D3D12HeapType {}
|
2023-02-01 09:50:47 +11:00
|
|
|
#[derive(Clone)]
|
2023-01-25 15:15:43 +11:00
|
|
|
pub struct SamplerPaletteHeap;
|
2023-02-01 09:50:47 +11:00
|
|
|
|
|
|
|
#[derive(Clone)]
|
|
|
|
pub struct CpuStagingHeap;
|
|
|
|
|
2023-02-01 16:16:06 +11:00
|
|
|
#[derive(Clone)]
|
|
|
|
pub struct RenderTargetHeap;
|
|
|
|
|
2023-02-01 09:50:47 +11:00
|
|
|
#[derive(Clone)]
|
2023-01-27 09:57:54 +11:00
|
|
|
pub struct ResourceWorkHeap;
|
2023-01-25 15:15:43 +11:00
|
|
|
|
2023-02-01 10:59:49 +11:00
|
|
|
#[derive(Clone)]
|
|
|
|
pub struct SamplerWorkHeap;
|
|
|
|
|
2023-01-25 15:15:43 +11:00
|
|
|
impl const D3D12HeapType for SamplerPaletteHeap {
|
|
|
|
// sampler palettes just get set directly
|
2023-01-24 18:02:27 +11:00
|
|
|
fn get_desc(size: usize) -> D3D12_DESCRIPTOR_HEAP_DESC {
|
|
|
|
D3D12_DESCRIPTOR_HEAP_DESC {
|
|
|
|
Type: D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER,
|
|
|
|
NumDescriptors: size as u32,
|
2023-02-01 10:59:49 +11:00
|
|
|
Flags: D3D12_DESCRIPTOR_HEAP_FLAG_NONE,
|
2023-01-24 18:02:27 +11:00
|
|
|
NodeMask: 0,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-01 09:50:47 +11:00
|
|
|
impl const D3D12HeapType for CpuStagingHeap {
|
2023-01-25 15:15:43 +11:00
|
|
|
// Lut texture heaps are CPU only and get bound to the descriptor heap of the shader.
|
|
|
|
fn get_desc(size: usize) -> D3D12_DESCRIPTOR_HEAP_DESC {
|
|
|
|
D3D12_DESCRIPTOR_HEAP_DESC {
|
|
|
|
Type: D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV,
|
|
|
|
NumDescriptors: size as u32,
|
|
|
|
Flags: D3D12_DESCRIPTOR_HEAP_FLAG_NONE,
|
|
|
|
NodeMask: 0,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-01 16:16:06 +11:00
|
|
|
impl const D3D12HeapType for RenderTargetHeap {
|
|
|
|
// Lut texture heaps are CPU only and get bound to the descriptor heap of the shader.
|
|
|
|
fn get_desc(size: usize) -> D3D12_DESCRIPTOR_HEAP_DESC {
|
|
|
|
D3D12_DESCRIPTOR_HEAP_DESC {
|
|
|
|
Type: D3D12_DESCRIPTOR_HEAP_TYPE_RTV,
|
|
|
|
NumDescriptors: size as u32,
|
|
|
|
Flags: D3D12_DESCRIPTOR_HEAP_FLAG_NONE,
|
|
|
|
NodeMask: 0,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-27 09:57:54 +11:00
|
|
|
impl D3D12ShaderVisibleHeapType for ResourceWorkHeap {}
|
|
|
|
impl const D3D12HeapType for ResourceWorkHeap {
|
|
|
|
// Lut texture heaps are CPU only and get bound to the descriptor heap of the shader.
|
|
|
|
fn get_desc(size: usize) -> D3D12_DESCRIPTOR_HEAP_DESC {
|
|
|
|
D3D12_DESCRIPTOR_HEAP_DESC {
|
|
|
|
Type: D3D12_DESCRIPTOR_HEAP_TYPE_CBV_SRV_UAV,
|
|
|
|
NumDescriptors: size as u32,
|
|
|
|
Flags: D3D12_DESCRIPTOR_HEAP_FLAG_SHADER_VISIBLE,
|
|
|
|
NodeMask: 0,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-01 10:59:49 +11:00
|
|
|
impl D3D12ShaderVisibleHeapType for SamplerWorkHeap {}
|
|
|
|
impl const D3D12HeapType for SamplerWorkHeap {
|
|
|
|
// Lut texture heaps are CPU only and get bound to the descriptor heap of the shader.
|
|
|
|
fn get_desc(size: usize) -> D3D12_DESCRIPTOR_HEAP_DESC {
|
|
|
|
D3D12_DESCRIPTOR_HEAP_DESC {
|
|
|
|
Type: D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER,
|
|
|
|
NumDescriptors: size as u32,
|
|
|
|
Flags: D3D12_DESCRIPTOR_HEAP_FLAG_SHADER_VISIBLE,
|
|
|
|
NodeMask: 0,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-02 10:09:34 +11:00
|
|
|
pub type D3D12DescriptorHeapSlot<T> = Rc<D3D12DescriptorHeapSlotInner<T>>;
|
|
|
|
|
|
|
|
pub struct D3D12DescriptorHeapSlotInner<T> {
|
2023-01-24 18:02:27 +11:00
|
|
|
cpu_handle: D3D12_CPU_DESCRIPTOR_HANDLE,
|
2023-01-25 15:15:43 +11:00
|
|
|
gpu_handle: Option<D3D12_GPU_DESCRIPTOR_HANDLE>,
|
2023-02-02 10:09:34 +11:00
|
|
|
heap: Rc<RefCell<D3D12DescriptorHeapInner>>,
|
2023-01-24 18:02:27 +11:00
|
|
|
slot: usize,
|
2023-01-25 15:15:43 +11:00
|
|
|
_pd: PhantomData<T>,
|
2023-01-24 18:02:27 +11:00
|
|
|
}
|
|
|
|
|
2023-02-02 10:09:34 +11:00
|
|
|
impl<T> D3D12DescriptorHeapSlotInner<T> {
|
2023-01-24 18:02:27 +11:00
|
|
|
/// Get the index of the resource within the heap.
|
|
|
|
pub fn index(&self) -> usize {
|
|
|
|
self.slot
|
|
|
|
}
|
2023-02-01 11:23:57 +11:00
|
|
|
|
|
|
|
/// unsafe because type must match
|
2023-02-02 10:09:34 +11:00
|
|
|
pub unsafe fn copy_descriptor(&self, source: D3D12_CPU_DESCRIPTOR_HANDLE) {
|
2023-02-01 11:23:57 +11:00
|
|
|
unsafe {
|
2023-02-06 08:17:23 +11:00
|
|
|
let heap = self.heap.deref().borrow();
|
|
|
|
|
|
|
|
heap.device
|
|
|
|
.CopyDescriptorsSimple(1, self.cpu_handle, source, heap.ty)
|
2023-02-01 11:23:57 +11:00
|
|
|
}
|
|
|
|
}
|
2023-01-24 18:02:27 +11:00
|
|
|
}
|
|
|
|
|
2023-02-02 10:09:34 +11:00
|
|
|
impl<T> AsRef<D3D12_CPU_DESCRIPTOR_HANDLE> for D3D12DescriptorHeapSlotInner<T> {
|
2023-01-24 18:02:27 +11:00
|
|
|
fn as_ref(&self) -> &D3D12_CPU_DESCRIPTOR_HANDLE {
|
|
|
|
&self.cpu_handle
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-25 15:15:43 +11:00
|
|
|
impl<T: D3D12ShaderVisibleHeapType> AsRef<D3D12_GPU_DESCRIPTOR_HANDLE>
|
2023-02-02 10:09:34 +11:00
|
|
|
for D3D12DescriptorHeapSlotInner<T>
|
2023-01-25 15:15:43 +11:00
|
|
|
{
|
2023-01-24 18:02:27 +11:00
|
|
|
fn as_ref(&self) -> &D3D12_GPU_DESCRIPTOR_HANDLE {
|
2023-02-06 11:58:51 +11:00
|
|
|
/// SAFETY: D3D12ShaderVisibleHeapType must have a GPU handle.
|
2023-01-25 15:15:43 +11:00
|
|
|
self.gpu_handle.as_ref().unwrap()
|
2023-01-24 18:02:27 +11:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-01-28 17:38:55 +11:00
|
|
|
impl<T: D3D12ShaderVisibleHeapType> From<&D3D12DescriptorHeap<T>> for ID3D12DescriptorHeap {
|
|
|
|
fn from(value: &D3D12DescriptorHeap<T>) -> Self {
|
|
|
|
value.0.borrow().heap.clone()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-01 10:59:49 +11:00
|
|
|
#[derive(Debug)]
|
2023-01-24 18:02:27 +11:00
|
|
|
struct D3D12DescriptorHeapInner {
|
2023-01-27 09:57:54 +11:00
|
|
|
device: ID3D12Device,
|
2023-01-24 18:02:27 +11:00
|
|
|
heap: ID3D12DescriptorHeap,
|
2023-02-01 10:59:49 +11:00
|
|
|
ty: D3D12_DESCRIPTOR_HEAP_TYPE,
|
2023-01-24 18:02:27 +11:00
|
|
|
cpu_start: D3D12_CPU_DESCRIPTOR_HANDLE,
|
2023-01-25 15:15:43 +11:00
|
|
|
gpu_start: Option<D3D12_GPU_DESCRIPTOR_HANDLE>,
|
2023-01-24 18:02:27 +11:00
|
|
|
handle_size: usize,
|
|
|
|
start: usize,
|
2023-02-01 10:59:49 +11:00
|
|
|
num_descriptors: usize,
|
2023-02-06 13:49:55 +11:00
|
|
|
map: BitBox,
|
2023-01-24 18:02:27 +11:00
|
|
|
}
|
|
|
|
|
2023-02-02 10:09:34 +11:00
|
|
|
pub struct D3D12DescriptorHeap<T>(Rc<RefCell<D3D12DescriptorHeapInner>>, PhantomData<T>);
|
2023-01-24 18:02:27 +11:00
|
|
|
|
2023-01-25 15:15:43 +11:00
|
|
|
impl<T: D3D12HeapType> D3D12DescriptorHeap<T> {
|
2023-01-24 18:02:27 +11:00
|
|
|
pub fn new(device: &ID3D12Device, size: usize) -> error::Result<D3D12DescriptorHeap<T>> {
|
|
|
|
let desc = T::get_desc(size);
|
2023-01-25 15:15:43 +11:00
|
|
|
unsafe { D3D12DescriptorHeap::new_with_desc(device, desc) }
|
2023-01-24 18:02:27 +11:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl<T> D3D12DescriptorHeap<T> {
|
2023-01-25 15:15:43 +11:00
|
|
|
pub unsafe fn new_with_desc(
|
|
|
|
device: &ID3D12Device,
|
|
|
|
desc: D3D12_DESCRIPTOR_HEAP_DESC,
|
|
|
|
) -> error::Result<D3D12DescriptorHeap<T>> {
|
2023-01-24 18:02:27 +11:00
|
|
|
unsafe {
|
|
|
|
let heap: ID3D12DescriptorHeap = device.CreateDescriptorHeap(&desc)?;
|
|
|
|
let cpu_start = heap.GetCPUDescriptorHandleForHeapStart();
|
2023-01-25 15:15:43 +11:00
|
|
|
|
|
|
|
let gpu_start = if (desc.Flags & D3D12_DESCRIPTOR_HEAP_FLAG_SHADER_VISIBLE).0 != 0 {
|
|
|
|
Some(heap.GetGPUDescriptorHandleForHeapStart())
|
|
|
|
} else {
|
|
|
|
None
|
|
|
|
};
|
|
|
|
|
|
|
|
Ok(D3D12DescriptorHeap(
|
2023-02-02 10:09:34 +11:00
|
|
|
Rc::new(RefCell::new(D3D12DescriptorHeapInner {
|
2023-01-27 09:57:54 +11:00
|
|
|
device: device.clone(),
|
2023-01-25 15:15:43 +11:00
|
|
|
heap,
|
2023-02-01 10:59:49 +11:00
|
|
|
ty: desc.Type,
|
2023-01-25 15:15:43 +11:00
|
|
|
cpu_start,
|
|
|
|
gpu_start,
|
|
|
|
handle_size: device.GetDescriptorHandleIncrementSize(desc.Type) as usize,
|
|
|
|
start: 0,
|
2023-02-01 10:59:49 +11:00
|
|
|
num_descriptors: desc.NumDescriptors as usize,
|
2023-02-06 13:49:55 +11:00
|
|
|
map: bitvec![0; desc.NumDescriptors as usize].into_boxed_bitslice()
|
2023-01-25 15:15:43 +11:00
|
|
|
})),
|
|
|
|
PhantomData::default(),
|
|
|
|
))
|
2023-01-24 18:02:27 +11:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-01 10:59:49 +11:00
|
|
|
/// suballocates this heap into equally sized chunks.
|
|
|
|
/// if there aren't enough descriptors, throws an error.
|
|
|
|
///
|
|
|
|
/// it is UB (programmer error) to call this if the descriptor heap already has
|
|
|
|
/// descriptors allocated for it.
|
|
|
|
///
|
|
|
|
/// size must also divide equally into the size of the heap.
|
2023-02-06 08:17:23 +11:00
|
|
|
pub unsafe fn suballocate(
|
|
|
|
self,
|
|
|
|
size: usize,
|
|
|
|
) -> (Vec<D3D12DescriptorHeap<T>>, ID3D12DescriptorHeap) {
|
2023-02-01 10:59:49 +11:00
|
|
|
// has to be called right after creation.
|
2023-02-06 08:17:23 +11:00
|
|
|
assert_eq!(
|
|
|
|
Rc::strong_count(&self.0),
|
|
|
|
1,
|
|
|
|
"D3D12DescriptorHeap::suballocate can only be callled immediately after creation."
|
|
|
|
);
|
2023-02-01 10:59:49 +11:00
|
|
|
|
2023-02-02 10:09:34 +11:00
|
|
|
let inner = Rc::try_unwrap(self.0)
|
2023-02-01 10:59:49 +11:00
|
|
|
.expect("[d3d12] undefined behaviour to suballocate a descriptor heap with live descriptors.")
|
|
|
|
.into_inner();
|
|
|
|
|
|
|
|
// number of suballocated heaps
|
|
|
|
let num_heaps = inner.num_descriptors / size;
|
|
|
|
let remainder = inner.num_descriptors % size;
|
|
|
|
|
2023-02-06 08:17:23 +11:00
|
|
|
assert_eq!(
|
|
|
|
remainder, 0,
|
|
|
|
"D3D12DescriptorHeap::suballocate \
|
|
|
|
must be called with a size that equally divides the number of descriptors"
|
|
|
|
);
|
2023-02-01 10:59:49 +11:00
|
|
|
|
|
|
|
let mut heaps = Vec::new();
|
|
|
|
|
|
|
|
let mut start = 0;
|
|
|
|
let root_cpu_ptr = inner.cpu_start.ptr;
|
|
|
|
let root_gpu_ptr = inner.gpu_start.map(|p| p.ptr);
|
|
|
|
|
|
|
|
for _ in 0..num_heaps {
|
2023-02-06 08:17:23 +11:00
|
|
|
let new_cpu_start = root_cpu_ptr + (start * inner.handle_size);
|
|
|
|
let new_gpu_start = root_gpu_ptr.map(|r| D3D12_GPU_DESCRIPTOR_HANDLE {
|
|
|
|
ptr: r + (start as u64 * inner.handle_size as u64),
|
|
|
|
});
|
2023-02-01 10:59:49 +11:00
|
|
|
|
|
|
|
heaps.push(D3D12DescriptorHeapInner {
|
|
|
|
device: inner.device.clone(),
|
|
|
|
heap: inner.heap.clone(),
|
|
|
|
ty: inner.ty,
|
2023-02-06 08:17:23 +11:00
|
|
|
cpu_start: D3D12_CPU_DESCRIPTOR_HANDLE { ptr: new_cpu_start },
|
2023-02-01 10:59:49 +11:00
|
|
|
gpu_start: new_gpu_start,
|
|
|
|
handle_size: inner.handle_size,
|
|
|
|
start: 0,
|
|
|
|
num_descriptors: size,
|
2023-02-06 13:49:55 +11:00
|
|
|
map: bitvec![0; size].into_boxed_bitslice()
|
2023-02-01 10:59:49 +11:00
|
|
|
});
|
|
|
|
|
|
|
|
start += size;
|
|
|
|
}
|
|
|
|
|
2023-02-06 08:17:23 +11:00
|
|
|
(
|
|
|
|
heaps
|
|
|
|
.into_iter()
|
|
|
|
.map(|inner| {
|
|
|
|
D3D12DescriptorHeap(Rc::new(RefCell::new(inner)), PhantomData::default())
|
|
|
|
})
|
|
|
|
.collect(),
|
|
|
|
inner.heap,
|
|
|
|
)
|
2023-02-01 10:59:49 +11:00
|
|
|
}
|
|
|
|
|
2023-01-24 18:02:27 +11:00
|
|
|
pub fn alloc_slot(&mut self) -> error::Result<D3D12DescriptorHeapSlot<T>> {
|
|
|
|
let mut handle = D3D12_CPU_DESCRIPTOR_HANDLE { ptr: 0 };
|
|
|
|
|
|
|
|
let mut inner = self.0.borrow_mut();
|
2023-02-01 10:59:49 +11:00
|
|
|
for i in inner.start..inner.num_descriptors {
|
2023-01-24 18:02:27 +11:00
|
|
|
if !inner.map[i] {
|
2023-02-06 13:49:55 +11:00
|
|
|
inner.map.set(i, true);
|
2023-01-24 18:02:27 +11:00
|
|
|
handle.ptr = inner.cpu_start.ptr + (i * inner.handle_size);
|
|
|
|
inner.start = i + 1;
|
|
|
|
|
2023-01-25 15:15:43 +11:00
|
|
|
let gpu_handle = inner
|
|
|
|
.gpu_start
|
|
|
|
.map(|gpu_start| D3D12_GPU_DESCRIPTOR_HANDLE {
|
2023-01-28 17:38:55 +11:00
|
|
|
ptr: (handle.ptr as u64 - inner.cpu_start.ptr as u64) + gpu_start.ptr,
|
2023-01-25 15:15:43 +11:00
|
|
|
});
|
|
|
|
|
2023-02-02 10:09:34 +11:00
|
|
|
return Ok(Rc::new(D3D12DescriptorHeapSlotInner {
|
2023-01-24 18:02:27 +11:00
|
|
|
cpu_handle: handle,
|
|
|
|
slot: i,
|
2023-02-02 10:09:34 +11:00
|
|
|
heap: Rc::clone(&self.0),
|
2023-01-24 18:02:27 +11:00
|
|
|
gpu_handle,
|
|
|
|
_pd: Default::default(),
|
2023-02-02 10:09:34 +11:00
|
|
|
}));
|
2023-01-24 18:02:27 +11:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-06 11:58:51 +11:00
|
|
|
Err(FilterChainError::DescriptorHeapOverflow)
|
2023-01-24 18:02:27 +11:00
|
|
|
}
|
2023-01-27 09:57:54 +11:00
|
|
|
|
2023-02-06 08:17:23 +11:00
|
|
|
pub fn alloc_range<const NUM_DESC: usize>(
|
|
|
|
&mut self,
|
|
|
|
) -> error::Result<[D3D12DescriptorHeapSlot<T>; NUM_DESC]> {
|
2023-01-27 09:57:54 +11:00
|
|
|
let dest = array_init::try_array_init(|_| self.alloc_slot())?;
|
|
|
|
Ok(dest)
|
|
|
|
}
|
2023-01-24 18:02:27 +11:00
|
|
|
}
|
|
|
|
|
2023-02-02 10:09:34 +11:00
|
|
|
impl<T> Drop for D3D12DescriptorHeapSlotInner<T> {
|
2023-01-24 18:02:27 +11:00
|
|
|
fn drop(&mut self) {
|
|
|
|
let mut inner = self.heap.borrow_mut();
|
2023-02-06 13:49:55 +11:00
|
|
|
inner.map.set(self.slot, false);
|
2023-01-24 18:02:27 +11:00
|
|
|
if inner.start > self.slot {
|
|
|
|
inner.start = self.slot
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|