Update Vulkan headers to 1.3.299
This commit is contained in:
parent
514d3e5393
commit
be03bf524f
15 changed files with 7924 additions and 1552 deletions
621
external/vulkan/vulkan_raii.hpp
vendored
621
external/vulkan/vulkan_raii.hpp
vendored
|
|
@ -8,9 +8,11 @@
|
|||
#ifndef VULKAN_RAII_HPP
|
||||
#define VULKAN_RAII_HPP
|
||||
|
||||
#include <memory> // std::unique_ptr
|
||||
#include <utility> // std::forward
|
||||
#include <vulkan/vulkan.hpp>
|
||||
#if !( defined( VULKAN_HPP_ENABLE_STD_MODULE ) && defined( VULKAN_HPP_STD_MODULE ) )
|
||||
# include <memory> // std::unique_ptr
|
||||
# include <utility> // std::forward
|
||||
#endif
|
||||
|
||||
#if !defined( VULKAN_HPP_DISABLE_ENHANCED_MODE )
|
||||
namespace VULKAN_HPP_NAMESPACE
|
||||
|
|
@ -1203,7 +1205,8 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
vkGetMemoryHostPointerPropertiesEXT = PFN_vkGetMemoryHostPointerPropertiesEXT( vkGetDeviceProcAddr( device, "vkGetMemoryHostPointerPropertiesEXT" ) );
|
||||
|
||||
//=== VK_AMD_buffer_marker ===
|
||||
vkCmdWriteBufferMarkerAMD = PFN_vkCmdWriteBufferMarkerAMD( vkGetDeviceProcAddr( device, "vkCmdWriteBufferMarkerAMD" ) );
|
||||
vkCmdWriteBufferMarkerAMD = PFN_vkCmdWriteBufferMarkerAMD( vkGetDeviceProcAddr( device, "vkCmdWriteBufferMarkerAMD" ) );
|
||||
vkCmdWriteBufferMarker2AMD = PFN_vkCmdWriteBufferMarker2AMD( vkGetDeviceProcAddr( device, "vkCmdWriteBufferMarker2AMD" ) );
|
||||
|
||||
//=== VK_EXT_calibrated_timestamps ===
|
||||
vkGetCalibratedTimestampsEXT = PFN_vkGetCalibratedTimestampsEXT( vkGetDeviceProcAddr( device, "vkGetCalibratedTimestampsEXT" ) );
|
||||
|
|
@ -1220,8 +1223,9 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
vkCmdSetExclusiveScissorNV = PFN_vkCmdSetExclusiveScissorNV( vkGetDeviceProcAddr( device, "vkCmdSetExclusiveScissorNV" ) );
|
||||
|
||||
//=== VK_NV_device_diagnostic_checkpoints ===
|
||||
vkCmdSetCheckpointNV = PFN_vkCmdSetCheckpointNV( vkGetDeviceProcAddr( device, "vkCmdSetCheckpointNV" ) );
|
||||
vkGetQueueCheckpointDataNV = PFN_vkGetQueueCheckpointDataNV( vkGetDeviceProcAddr( device, "vkGetQueueCheckpointDataNV" ) );
|
||||
vkCmdSetCheckpointNV = PFN_vkCmdSetCheckpointNV( vkGetDeviceProcAddr( device, "vkCmdSetCheckpointNV" ) );
|
||||
vkGetQueueCheckpointDataNV = PFN_vkGetQueueCheckpointDataNV( vkGetDeviceProcAddr( device, "vkGetQueueCheckpointDataNV" ) );
|
||||
vkGetQueueCheckpointData2NV = PFN_vkGetQueueCheckpointData2NV( vkGetDeviceProcAddr( device, "vkGetQueueCheckpointData2NV" ) );
|
||||
|
||||
//=== VK_KHR_timeline_semaphore ===
|
||||
vkGetSemaphoreCounterValueKHR = PFN_vkGetSemaphoreCounterValueKHR( vkGetDeviceProcAddr( device, "vkGetSemaphoreCounterValueKHR" ) );
|
||||
|
|
@ -1434,8 +1438,6 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
vkQueueSubmit2KHR = PFN_vkQueueSubmit2KHR( vkGetDeviceProcAddr( device, "vkQueueSubmit2KHR" ) );
|
||||
if ( !vkQueueSubmit2 )
|
||||
vkQueueSubmit2 = vkQueueSubmit2KHR;
|
||||
vkCmdWriteBufferMarker2AMD = PFN_vkCmdWriteBufferMarker2AMD( vkGetDeviceProcAddr( device, "vkCmdWriteBufferMarker2AMD" ) );
|
||||
vkGetQueueCheckpointData2NV = PFN_vkGetQueueCheckpointData2NV( vkGetDeviceProcAddr( device, "vkGetQueueCheckpointData2NV" ) );
|
||||
|
||||
//=== VK_EXT_descriptor_buffer ===
|
||||
vkGetDescriptorSetLayoutSizeEXT = PFN_vkGetDescriptorSetLayoutSizeEXT( vkGetDeviceProcAddr( device, "vkGetDescriptorSetLayoutSizeEXT" ) );
|
||||
|
|
@ -1672,10 +1674,11 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
vkAntiLagUpdateAMD = PFN_vkAntiLagUpdateAMD( vkGetDeviceProcAddr( device, "vkAntiLagUpdateAMD" ) );
|
||||
|
||||
//=== VK_EXT_shader_object ===
|
||||
vkCreateShadersEXT = PFN_vkCreateShadersEXT( vkGetDeviceProcAddr( device, "vkCreateShadersEXT" ) );
|
||||
vkDestroyShaderEXT = PFN_vkDestroyShaderEXT( vkGetDeviceProcAddr( device, "vkDestroyShaderEXT" ) );
|
||||
vkGetShaderBinaryDataEXT = PFN_vkGetShaderBinaryDataEXT( vkGetDeviceProcAddr( device, "vkGetShaderBinaryDataEXT" ) );
|
||||
vkCmdBindShadersEXT = PFN_vkCmdBindShadersEXT( vkGetDeviceProcAddr( device, "vkCmdBindShadersEXT" ) );
|
||||
vkCreateShadersEXT = PFN_vkCreateShadersEXT( vkGetDeviceProcAddr( device, "vkCreateShadersEXT" ) );
|
||||
vkDestroyShaderEXT = PFN_vkDestroyShaderEXT( vkGetDeviceProcAddr( device, "vkDestroyShaderEXT" ) );
|
||||
vkGetShaderBinaryDataEXT = PFN_vkGetShaderBinaryDataEXT( vkGetDeviceProcAddr( device, "vkGetShaderBinaryDataEXT" ) );
|
||||
vkCmdBindShadersEXT = PFN_vkCmdBindShadersEXT( vkGetDeviceProcAddr( device, "vkCmdBindShadersEXT" ) );
|
||||
vkCmdSetDepthClampRangeEXT = PFN_vkCmdSetDepthClampRangeEXT( vkGetDeviceProcAddr( device, "vkCmdSetDepthClampRangeEXT" ) );
|
||||
|
||||
//=== VK_KHR_pipeline_binary ===
|
||||
vkCreatePipelineBinariesKHR = PFN_vkCreatePipelineBinariesKHR( vkGetDeviceProcAddr( device, "vkCreatePipelineBinariesKHR" ) );
|
||||
|
|
@ -1720,6 +1723,20 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
vkCmdSetDescriptorBufferOffsets2EXT = PFN_vkCmdSetDescriptorBufferOffsets2EXT( vkGetDeviceProcAddr( device, "vkCmdSetDescriptorBufferOffsets2EXT" ) );
|
||||
vkCmdBindDescriptorBufferEmbeddedSamplers2EXT =
|
||||
PFN_vkCmdBindDescriptorBufferEmbeddedSamplers2EXT( vkGetDeviceProcAddr( device, "vkCmdBindDescriptorBufferEmbeddedSamplers2EXT" ) );
|
||||
|
||||
//=== VK_EXT_device_generated_commands ===
|
||||
vkGetGeneratedCommandsMemoryRequirementsEXT =
|
||||
PFN_vkGetGeneratedCommandsMemoryRequirementsEXT( vkGetDeviceProcAddr( device, "vkGetGeneratedCommandsMemoryRequirementsEXT" ) );
|
||||
vkCmdPreprocessGeneratedCommandsEXT = PFN_vkCmdPreprocessGeneratedCommandsEXT( vkGetDeviceProcAddr( device, "vkCmdPreprocessGeneratedCommandsEXT" ) );
|
||||
vkCmdExecuteGeneratedCommandsEXT = PFN_vkCmdExecuteGeneratedCommandsEXT( vkGetDeviceProcAddr( device, "vkCmdExecuteGeneratedCommandsEXT" ) );
|
||||
vkCreateIndirectCommandsLayoutEXT = PFN_vkCreateIndirectCommandsLayoutEXT( vkGetDeviceProcAddr( device, "vkCreateIndirectCommandsLayoutEXT" ) );
|
||||
vkDestroyIndirectCommandsLayoutEXT = PFN_vkDestroyIndirectCommandsLayoutEXT( vkGetDeviceProcAddr( device, "vkDestroyIndirectCommandsLayoutEXT" ) );
|
||||
vkCreateIndirectExecutionSetEXT = PFN_vkCreateIndirectExecutionSetEXT( vkGetDeviceProcAddr( device, "vkCreateIndirectExecutionSetEXT" ) );
|
||||
vkDestroyIndirectExecutionSetEXT = PFN_vkDestroyIndirectExecutionSetEXT( vkGetDeviceProcAddr( device, "vkDestroyIndirectExecutionSetEXT" ) );
|
||||
vkUpdateIndirectExecutionSetPipelineEXT =
|
||||
PFN_vkUpdateIndirectExecutionSetPipelineEXT( vkGetDeviceProcAddr( device, "vkUpdateIndirectExecutionSetPipelineEXT" ) );
|
||||
vkUpdateIndirectExecutionSetShaderEXT =
|
||||
PFN_vkUpdateIndirectExecutionSetShaderEXT( vkGetDeviceProcAddr( device, "vkUpdateIndirectExecutionSetShaderEXT" ) );
|
||||
}
|
||||
|
||||
public:
|
||||
|
|
@ -2203,7 +2220,8 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
PFN_vkGetMemoryHostPointerPropertiesEXT vkGetMemoryHostPointerPropertiesEXT = 0;
|
||||
|
||||
//=== VK_AMD_buffer_marker ===
|
||||
PFN_vkCmdWriteBufferMarkerAMD vkCmdWriteBufferMarkerAMD = 0;
|
||||
PFN_vkCmdWriteBufferMarkerAMD vkCmdWriteBufferMarkerAMD = 0;
|
||||
PFN_vkCmdWriteBufferMarker2AMD vkCmdWriteBufferMarker2AMD = 0;
|
||||
|
||||
//=== VK_EXT_calibrated_timestamps ===
|
||||
PFN_vkGetCalibratedTimestampsEXT vkGetCalibratedTimestampsEXT = 0;
|
||||
|
|
@ -2218,8 +2236,9 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
PFN_vkCmdSetExclusiveScissorNV vkCmdSetExclusiveScissorNV = 0;
|
||||
|
||||
//=== VK_NV_device_diagnostic_checkpoints ===
|
||||
PFN_vkCmdSetCheckpointNV vkCmdSetCheckpointNV = 0;
|
||||
PFN_vkGetQueueCheckpointDataNV vkGetQueueCheckpointDataNV = 0;
|
||||
PFN_vkCmdSetCheckpointNV vkCmdSetCheckpointNV = 0;
|
||||
PFN_vkGetQueueCheckpointDataNV vkGetQueueCheckpointDataNV = 0;
|
||||
PFN_vkGetQueueCheckpointData2NV vkGetQueueCheckpointData2NV = 0;
|
||||
|
||||
//=== VK_KHR_timeline_semaphore ===
|
||||
PFN_vkGetSemaphoreCounterValueKHR vkGetSemaphoreCounterValueKHR = 0;
|
||||
|
|
@ -2361,14 +2380,12 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
# endif /*VK_USE_PLATFORM_METAL_EXT*/
|
||||
|
||||
//=== VK_KHR_synchronization2 ===
|
||||
PFN_vkCmdSetEvent2KHR vkCmdSetEvent2KHR = 0;
|
||||
PFN_vkCmdResetEvent2KHR vkCmdResetEvent2KHR = 0;
|
||||
PFN_vkCmdWaitEvents2KHR vkCmdWaitEvents2KHR = 0;
|
||||
PFN_vkCmdPipelineBarrier2KHR vkCmdPipelineBarrier2KHR = 0;
|
||||
PFN_vkCmdWriteTimestamp2KHR vkCmdWriteTimestamp2KHR = 0;
|
||||
PFN_vkQueueSubmit2KHR vkQueueSubmit2KHR = 0;
|
||||
PFN_vkCmdWriteBufferMarker2AMD vkCmdWriteBufferMarker2AMD = 0;
|
||||
PFN_vkGetQueueCheckpointData2NV vkGetQueueCheckpointData2NV = 0;
|
||||
PFN_vkCmdSetEvent2KHR vkCmdSetEvent2KHR = 0;
|
||||
PFN_vkCmdResetEvent2KHR vkCmdResetEvent2KHR = 0;
|
||||
PFN_vkCmdWaitEvents2KHR vkCmdWaitEvents2KHR = 0;
|
||||
PFN_vkCmdPipelineBarrier2KHR vkCmdPipelineBarrier2KHR = 0;
|
||||
PFN_vkCmdWriteTimestamp2KHR vkCmdWriteTimestamp2KHR = 0;
|
||||
PFN_vkQueueSubmit2KHR vkQueueSubmit2KHR = 0;
|
||||
|
||||
//=== VK_EXT_descriptor_buffer ===
|
||||
PFN_vkGetDescriptorSetLayoutSizeEXT vkGetDescriptorSetLayoutSizeEXT = 0;
|
||||
|
|
@ -2566,10 +2583,11 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
PFN_vkAntiLagUpdateAMD vkAntiLagUpdateAMD = 0;
|
||||
|
||||
//=== VK_EXT_shader_object ===
|
||||
PFN_vkCreateShadersEXT vkCreateShadersEXT = 0;
|
||||
PFN_vkDestroyShaderEXT vkDestroyShaderEXT = 0;
|
||||
PFN_vkGetShaderBinaryDataEXT vkGetShaderBinaryDataEXT = 0;
|
||||
PFN_vkCmdBindShadersEXT vkCmdBindShadersEXT = 0;
|
||||
PFN_vkCreateShadersEXT vkCreateShadersEXT = 0;
|
||||
PFN_vkDestroyShaderEXT vkDestroyShaderEXT = 0;
|
||||
PFN_vkGetShaderBinaryDataEXT vkGetShaderBinaryDataEXT = 0;
|
||||
PFN_vkCmdBindShadersEXT vkCmdBindShadersEXT = 0;
|
||||
PFN_vkCmdSetDepthClampRangeEXT vkCmdSetDepthClampRangeEXT = 0;
|
||||
|
||||
//=== VK_KHR_pipeline_binary ===
|
||||
PFN_vkCreatePipelineBinariesKHR vkCreatePipelineBinariesKHR = 0;
|
||||
|
|
@ -2612,6 +2630,17 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
PFN_vkCmdPushDescriptorSetWithTemplate2KHR vkCmdPushDescriptorSetWithTemplate2KHR = 0;
|
||||
PFN_vkCmdSetDescriptorBufferOffsets2EXT vkCmdSetDescriptorBufferOffsets2EXT = 0;
|
||||
PFN_vkCmdBindDescriptorBufferEmbeddedSamplers2EXT vkCmdBindDescriptorBufferEmbeddedSamplers2EXT = 0;
|
||||
|
||||
//=== VK_EXT_device_generated_commands ===
|
||||
PFN_vkGetGeneratedCommandsMemoryRequirementsEXT vkGetGeneratedCommandsMemoryRequirementsEXT = 0;
|
||||
PFN_vkCmdPreprocessGeneratedCommandsEXT vkCmdPreprocessGeneratedCommandsEXT = 0;
|
||||
PFN_vkCmdExecuteGeneratedCommandsEXT vkCmdExecuteGeneratedCommandsEXT = 0;
|
||||
PFN_vkCreateIndirectCommandsLayoutEXT vkCreateIndirectCommandsLayoutEXT = 0;
|
||||
PFN_vkDestroyIndirectCommandsLayoutEXT vkDestroyIndirectCommandsLayoutEXT = 0;
|
||||
PFN_vkCreateIndirectExecutionSetEXT vkCreateIndirectExecutionSetEXT = 0;
|
||||
PFN_vkDestroyIndirectExecutionSetEXT vkDestroyIndirectExecutionSetEXT = 0;
|
||||
PFN_vkUpdateIndirectExecutionSetPipelineEXT vkUpdateIndirectExecutionSetPipelineEXT = 0;
|
||||
PFN_vkUpdateIndirectExecutionSetShaderEXT vkUpdateIndirectExecutionSetShaderEXT = 0;
|
||||
};
|
||||
|
||||
//========================================
|
||||
|
|
@ -2717,6 +2746,10 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
//=== VK_KHR_pipeline_binary ===
|
||||
class PipelineBinaryKHR;
|
||||
|
||||
//=== VK_EXT_device_generated_commands ===
|
||||
class IndirectCommandsLayoutEXT;
|
||||
class IndirectExecutionSetEXT;
|
||||
|
||||
//====================
|
||||
//=== RAII HANDLES ===
|
||||
//====================
|
||||
|
|
@ -4619,6 +4652,27 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
VULKAN_HPP_NODISCARD std::pair<uint64_t, uint64_t>
|
||||
getCalibratedTimestampKHR( const VULKAN_HPP_NAMESPACE::CalibratedTimestampInfoKHR & timestampInfo ) const;
|
||||
|
||||
//=== VK_EXT_device_generated_commands ===
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_NAMESPACE::MemoryRequirements2
|
||||
getGeneratedCommandsMemoryRequirementsEXT( const VULKAN_HPP_NAMESPACE::GeneratedCommandsMemoryRequirementsInfoEXT & info ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
template <typename X, typename Y, typename... Z>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...>
|
||||
getGeneratedCommandsMemoryRequirementsEXT( const VULKAN_HPP_NAMESPACE::GeneratedCommandsMemoryRequirementsInfoEXT & info ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
VULKAN_HPP_NODISCARD
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::CreateReturnType<VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::IndirectCommandsLayoutEXT>::Type
|
||||
createIndirectCommandsLayoutEXT( VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutCreateInfoEXT const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr ) const
|
||||
VULKAN_HPP_RAII_CREATE_NOEXCEPT;
|
||||
|
||||
VULKAN_HPP_NODISCARD
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::CreateReturnType<VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::IndirectExecutionSetEXT>::Type
|
||||
createIndirectExecutionSetEXT( VULKAN_HPP_NAMESPACE::IndirectExecutionSetCreateInfoEXT const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr ) const
|
||||
VULKAN_HPP_RAII_CREATE_NOEXCEPT;
|
||||
|
||||
private:
|
||||
VULKAN_HPP_NAMESPACE::Device m_device = {};
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * m_allocator = {};
|
||||
|
|
@ -5893,15 +5947,20 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
# if defined( VK_ENABLE_BETA_EXTENSIONS )
|
||||
//=== VK_AMDX_shader_enqueue ===
|
||||
|
||||
void initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch ) const VULKAN_HPP_NOEXCEPT;
|
||||
void initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
void dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
void dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
void dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
|
|
@ -6022,6 +6081,11 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
VULKAN_HPP_NAMESPACE::DeviceSize dstOffset,
|
||||
uint32_t marker ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
void writeBufferMarker2AMD( VULKAN_HPP_NAMESPACE::PipelineStageFlags2 stage,
|
||||
VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize dstOffset,
|
||||
uint32_t marker ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
//=== VK_NV_mesh_shader ===
|
||||
|
||||
void drawMeshTasksNV( uint32_t taskCount, uint32_t firstTask ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
|
@ -6152,11 +6216,6 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
VULKAN_HPP_NAMESPACE::QueryPool queryPool,
|
||||
uint32_t query ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
void writeBufferMarker2AMD( VULKAN_HPP_NAMESPACE::PipelineStageFlags2 stage,
|
||||
VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize dstOffset,
|
||||
uint32_t marker ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
//=== VK_EXT_descriptor_buffer ===
|
||||
|
||||
void bindDescriptorBuffersEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::DescriptorBufferBindingInfoEXT> const & bindingInfos ) const
|
||||
|
|
@ -6390,6 +6449,10 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
void bindShadersEXT( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ShaderStageFlagBits> const & stages,
|
||||
VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::ShaderEXT> const & shaders ) const;
|
||||
|
||||
void setDepthClampRangeEXT( VULKAN_HPP_NAMESPACE::DepthClampModeEXT depthClampMode,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::DepthClampRangeEXT> depthClampRange
|
||||
VULKAN_HPP_DEFAULT_ARGUMENT_NULLPTR_ASSIGNMENT ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
//=== VK_EXT_attachment_feedback_loop_dynamic_state ===
|
||||
|
||||
void setAttachmentFeedbackLoopEnableEXT( VULKAN_HPP_NAMESPACE::ImageAspectFlags aspectMask VULKAN_HPP_DEFAULT_ARGUMENT_ASSIGNMENT ) const
|
||||
|
|
@ -6416,6 +6479,14 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
void bindDescriptorBufferEmbeddedSamplers2EXT(
|
||||
const VULKAN_HPP_NAMESPACE::BindDescriptorBufferEmbeddedSamplersInfoEXT & bindDescriptorBufferEmbeddedSamplersInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
//=== VK_EXT_device_generated_commands ===
|
||||
|
||||
void preprocessGeneratedCommandsEXT( const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoEXT & generatedCommandsInfo,
|
||||
VULKAN_HPP_NAMESPACE::CommandBuffer stateCommandBuffer ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
void executeGeneratedCommandsEXT( VULKAN_HPP_NAMESPACE::Bool32 isPreprocessed,
|
||||
const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoEXT & generatedCommandsInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
private:
|
||||
VULKAN_HPP_NAMESPACE::Device m_device = {};
|
||||
VULKAN_HPP_NAMESPACE::CommandPool m_commandPool = {};
|
||||
|
|
@ -8898,6 +8969,126 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * m_dispatcher = nullptr;
|
||||
};
|
||||
|
||||
class IndirectCommandsLayoutEXT
|
||||
{
|
||||
public:
|
||||
using CType = VkIndirectCommandsLayoutEXT;
|
||||
using CppType = VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT;
|
||||
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::ObjectType objectType = VULKAN_HPP_NAMESPACE::ObjectType::eIndirectCommandsLayoutEXT;
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT debugReportObjectType =
|
||||
VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT::eUnknown;
|
||||
|
||||
public:
|
||||
# if !defined( VULKAN_HPP_RAII_NO_EXCEPTIONS )
|
||||
IndirectCommandsLayoutEXT( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::Device const & device,
|
||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutCreateInfoEXT const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr )
|
||||
{
|
||||
*this = device.createIndirectCommandsLayoutEXT( createInfo, allocator );
|
||||
}
|
||||
# endif
|
||||
|
||||
IndirectCommandsLayoutEXT( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::Device const & device,
|
||||
VkIndirectCommandsLayoutEXT indirectCommandsLayout,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr )
|
||||
: m_device( device )
|
||||
, m_indirectCommandsLayout( indirectCommandsLayout )
|
||||
, m_allocator( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) )
|
||||
, m_dispatcher( device.getDispatcher() )
|
||||
{
|
||||
}
|
||||
|
||||
IndirectCommandsLayoutEXT( std::nullptr_t ) {}
|
||||
|
||||
~IndirectCommandsLayoutEXT()
|
||||
{
|
||||
clear();
|
||||
}
|
||||
|
||||
IndirectCommandsLayoutEXT() = delete;
|
||||
IndirectCommandsLayoutEXT( IndirectCommandsLayoutEXT const & ) = delete;
|
||||
|
||||
IndirectCommandsLayoutEXT( IndirectCommandsLayoutEXT && rhs ) VULKAN_HPP_NOEXCEPT
|
||||
: m_device( VULKAN_HPP_NAMESPACE::exchange( rhs.m_device, {} ) )
|
||||
, m_indirectCommandsLayout( VULKAN_HPP_NAMESPACE::exchange( rhs.m_indirectCommandsLayout, {} ) )
|
||||
, m_allocator( VULKAN_HPP_NAMESPACE::exchange( rhs.m_allocator, {} ) )
|
||||
, m_dispatcher( VULKAN_HPP_NAMESPACE::exchange( rhs.m_dispatcher, nullptr ) )
|
||||
{
|
||||
}
|
||||
|
||||
IndirectCommandsLayoutEXT & operator=( IndirectCommandsLayoutEXT const & ) = delete;
|
||||
|
||||
IndirectCommandsLayoutEXT & operator=( IndirectCommandsLayoutEXT && rhs ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
if ( this != &rhs )
|
||||
{
|
||||
std::swap( m_device, rhs.m_device );
|
||||
std::swap( m_indirectCommandsLayout, rhs.m_indirectCommandsLayout );
|
||||
std::swap( m_allocator, rhs.m_allocator );
|
||||
std::swap( m_dispatcher, rhs.m_dispatcher );
|
||||
}
|
||||
return *this;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT const & operator*() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_indirectCommandsLayout;
|
||||
}
|
||||
|
||||
operator VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_indirectCommandsLayout;
|
||||
}
|
||||
|
||||
void clear() VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
if ( m_indirectCommandsLayout )
|
||||
{
|
||||
getDispatcher()->vkDestroyIndirectCommandsLayoutEXT( static_cast<VkDevice>( m_device ),
|
||||
static_cast<VkIndirectCommandsLayoutEXT>( m_indirectCommandsLayout ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( m_allocator ) );
|
||||
}
|
||||
m_device = nullptr;
|
||||
m_indirectCommandsLayout = nullptr;
|
||||
m_allocator = nullptr;
|
||||
m_dispatcher = nullptr;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT release()
|
||||
{
|
||||
m_device = nullptr;
|
||||
m_allocator = nullptr;
|
||||
m_dispatcher = nullptr;
|
||||
return VULKAN_HPP_NAMESPACE::exchange( m_indirectCommandsLayout, nullptr );
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::Device getDevice() const
|
||||
{
|
||||
return m_device;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * getDispatcher() const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( m_dispatcher->getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
return m_dispatcher;
|
||||
}
|
||||
|
||||
void swap( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::IndirectCommandsLayoutEXT & rhs ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::swap( m_device, rhs.m_device );
|
||||
std::swap( m_indirectCommandsLayout, rhs.m_indirectCommandsLayout );
|
||||
std::swap( m_allocator, rhs.m_allocator );
|
||||
std::swap( m_dispatcher, rhs.m_dispatcher );
|
||||
}
|
||||
|
||||
private:
|
||||
VULKAN_HPP_NAMESPACE::Device m_device = {};
|
||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT m_indirectCommandsLayout = {};
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * m_allocator = {};
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * m_dispatcher = nullptr;
|
||||
};
|
||||
|
||||
class IndirectCommandsLayoutNV
|
||||
{
|
||||
public:
|
||||
|
|
@ -9018,6 +9209,134 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * m_dispatcher = nullptr;
|
||||
};
|
||||
|
||||
class IndirectExecutionSetEXT
|
||||
{
|
||||
public:
|
||||
using CType = VkIndirectExecutionSetEXT;
|
||||
using CppType = VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT;
|
||||
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::ObjectType objectType = VULKAN_HPP_NAMESPACE::ObjectType::eIndirectExecutionSetEXT;
|
||||
static VULKAN_HPP_CONST_OR_CONSTEXPR VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT debugReportObjectType =
|
||||
VULKAN_HPP_NAMESPACE::DebugReportObjectTypeEXT::eUnknown;
|
||||
|
||||
public:
|
||||
# if !defined( VULKAN_HPP_RAII_NO_EXCEPTIONS )
|
||||
IndirectExecutionSetEXT( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::Device const & device,
|
||||
VULKAN_HPP_NAMESPACE::IndirectExecutionSetCreateInfoEXT const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr )
|
||||
{
|
||||
*this = device.createIndirectExecutionSetEXT( createInfo, allocator );
|
||||
}
|
||||
# endif
|
||||
|
||||
IndirectExecutionSetEXT( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::Device const & device,
|
||||
VkIndirectExecutionSetEXT indirectExecutionSet,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator = nullptr )
|
||||
: m_device( device )
|
||||
, m_indirectExecutionSet( indirectExecutionSet )
|
||||
, m_allocator( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) )
|
||||
, m_dispatcher( device.getDispatcher() )
|
||||
{
|
||||
}
|
||||
|
||||
IndirectExecutionSetEXT( std::nullptr_t ) {}
|
||||
|
||||
~IndirectExecutionSetEXT()
|
||||
{
|
||||
clear();
|
||||
}
|
||||
|
||||
IndirectExecutionSetEXT() = delete;
|
||||
IndirectExecutionSetEXT( IndirectExecutionSetEXT const & ) = delete;
|
||||
|
||||
IndirectExecutionSetEXT( IndirectExecutionSetEXT && rhs ) VULKAN_HPP_NOEXCEPT
|
||||
: m_device( VULKAN_HPP_NAMESPACE::exchange( rhs.m_device, {} ) )
|
||||
, m_indirectExecutionSet( VULKAN_HPP_NAMESPACE::exchange( rhs.m_indirectExecutionSet, {} ) )
|
||||
, m_allocator( VULKAN_HPP_NAMESPACE::exchange( rhs.m_allocator, {} ) )
|
||||
, m_dispatcher( VULKAN_HPP_NAMESPACE::exchange( rhs.m_dispatcher, nullptr ) )
|
||||
{
|
||||
}
|
||||
|
||||
IndirectExecutionSetEXT & operator=( IndirectExecutionSetEXT const & ) = delete;
|
||||
|
||||
IndirectExecutionSetEXT & operator=( IndirectExecutionSetEXT && rhs ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
if ( this != &rhs )
|
||||
{
|
||||
std::swap( m_device, rhs.m_device );
|
||||
std::swap( m_indirectExecutionSet, rhs.m_indirectExecutionSet );
|
||||
std::swap( m_allocator, rhs.m_allocator );
|
||||
std::swap( m_dispatcher, rhs.m_dispatcher );
|
||||
}
|
||||
return *this;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT const & operator*() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_indirectExecutionSet;
|
||||
}
|
||||
|
||||
operator VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT() const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
return m_indirectExecutionSet;
|
||||
}
|
||||
|
||||
void clear() VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
if ( m_indirectExecutionSet )
|
||||
{
|
||||
getDispatcher()->vkDestroyIndirectExecutionSetEXT( static_cast<VkDevice>( m_device ),
|
||||
static_cast<VkIndirectExecutionSetEXT>( m_indirectExecutionSet ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( m_allocator ) );
|
||||
}
|
||||
m_device = nullptr;
|
||||
m_indirectExecutionSet = nullptr;
|
||||
m_allocator = nullptr;
|
||||
m_dispatcher = nullptr;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT release()
|
||||
{
|
||||
m_device = nullptr;
|
||||
m_allocator = nullptr;
|
||||
m_dispatcher = nullptr;
|
||||
return VULKAN_HPP_NAMESPACE::exchange( m_indirectExecutionSet, nullptr );
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::Device getDevice() const
|
||||
{
|
||||
return m_device;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * getDispatcher() const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( m_dispatcher->getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
return m_dispatcher;
|
||||
}
|
||||
|
||||
void swap( VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::IndirectExecutionSetEXT & rhs ) VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
std::swap( m_device, rhs.m_device );
|
||||
std::swap( m_indirectExecutionSet, rhs.m_indirectExecutionSet );
|
||||
std::swap( m_allocator, rhs.m_allocator );
|
||||
std::swap( m_dispatcher, rhs.m_dispatcher );
|
||||
}
|
||||
|
||||
//=== VK_EXT_device_generated_commands ===
|
||||
|
||||
void updatePipeline( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::WriteIndirectExecutionSetPipelineEXT> const & executionSetWrites ) const
|
||||
VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
void updateShader( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::WriteIndirectExecutionSetShaderEXT> const & executionSetWrites ) const
|
||||
VULKAN_HPP_NOEXCEPT;
|
||||
|
||||
private:
|
||||
VULKAN_HPP_NAMESPACE::Device m_device = {};
|
||||
VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT m_indirectExecutionSet = {};
|
||||
const VULKAN_HPP_NAMESPACE::AllocationCallbacks * m_allocator = {};
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::DeviceDispatcher const * m_dispatcher = nullptr;
|
||||
};
|
||||
|
||||
class MicromapEXT
|
||||
{
|
||||
public:
|
||||
|
|
@ -10447,6 +10766,8 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
|
||||
VULKAN_HPP_NODISCARD std::vector<VULKAN_HPP_NAMESPACE::CheckpointDataNV> getCheckpointDataNV() const;
|
||||
|
||||
VULKAN_HPP_NODISCARD std::vector<VULKAN_HPP_NAMESPACE::CheckpointData2NV> getCheckpointData2NV() const;
|
||||
|
||||
//=== VK_INTEL_performance_query ===
|
||||
|
||||
void setPerformanceConfigurationINTEL( VULKAN_HPP_NAMESPACE::PerformanceConfigurationINTEL configuration ) const;
|
||||
|
|
@ -10456,8 +10777,6 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
void submit2KHR( VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::SubmitInfo2> const & submits,
|
||||
VULKAN_HPP_NAMESPACE::Fence fence VULKAN_HPP_DEFAULT_ARGUMENT_ASSIGNMENT ) const;
|
||||
|
||||
VULKAN_HPP_NODISCARD std::vector<VULKAN_HPP_NAMESPACE::CheckpointData2NV> getCheckpointData2NV() const;
|
||||
|
||||
//=== VK_NV_low_latency2 ===
|
||||
|
||||
void notifyOutOfBandNV( const VULKAN_HPP_NAMESPACE::OutOfBandQueueTypeInfoNV & queueTypeInfo ) const VULKAN_HPP_NOEXCEPT;
|
||||
|
|
@ -18062,43 +18381,55 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
return nodeIndex;
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void CommandBuffer::initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch ) const VULKAN_HPP_NOEXCEPT
|
||||
VULKAN_HPP_INLINE void CommandBuffer::initializeGraphScratchMemoryAMDX( VULKAN_HPP_NAMESPACE::Pipeline executionGraph,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdInitializeGraphScratchMemoryAMDX &&
|
||||
"Function <vkCmdInitializeGraphScratchMemoryAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||
|
||||
getDispatcher()->vkCmdInitializeGraphScratchMemoryAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ), static_cast<VkDeviceAddress>( scratch ) );
|
||||
getDispatcher()->vkCmdInitializeGraphScratchMemoryAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkPipeline>( executionGraph ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ) );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphAMDX && "Function <vkCmdDispatchGraphAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||
|
||||
getDispatcher()->vkCmdDispatchGraphAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void
|
||||
CommandBuffer::dispatchGraphIndirectAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
const VULKAN_HPP_NAMESPACE::DispatchGraphCountInfoAMDX & countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphIndirectAMDX && "Function <vkCmdDispatchGraphIndirectAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||
|
||||
getDispatcher()->vkCmdDispatchGraphIndirectAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
reinterpret_cast<const VkDispatchGraphCountInfoAMDX *>( &countInfo ) );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void CommandBuffer::dispatchGraphIndirectCountAMDX( VULKAN_HPP_NAMESPACE::DeviceAddress scratch,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize scratchSize,
|
||||
VULKAN_HPP_NAMESPACE::DeviceAddress countInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdDispatchGraphIndirectCountAMDX &&
|
||||
"Function <vkCmdDispatchGraphIndirectCountAMDX> requires <VK_AMDX_shader_enqueue>" );
|
||||
|
||||
getDispatcher()->vkCmdDispatchGraphIndirectCountAMDX(
|
||||
static_cast<VkCommandBuffer>( m_commandBuffer ), static_cast<VkDeviceAddress>( scratch ), static_cast<VkDeviceAddress>( countInfo ) );
|
||||
getDispatcher()->vkCmdDispatchGraphIndirectCountAMDX( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkDeviceAddress>( scratch ),
|
||||
static_cast<VkDeviceSize>( scratchSize ),
|
||||
static_cast<VkDeviceAddress>( countInfo ) );
|
||||
}
|
||||
# endif /*VK_ENABLE_BETA_EXTENSIONS*/
|
||||
|
||||
|
|
@ -19322,6 +19653,20 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
marker );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void CommandBuffer::writeBufferMarker2AMD( VULKAN_HPP_NAMESPACE::PipelineStageFlags2 stage,
|
||||
VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize dstOffset,
|
||||
uint32_t marker ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdWriteBufferMarker2AMD && "Function <vkCmdWriteBufferMarker2AMD> requires <VK_AMD_buffer_marker>" );
|
||||
|
||||
getDispatcher()->vkCmdWriteBufferMarker2AMD( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkPipelineStageFlags2>( stage ),
|
||||
static_cast<VkBuffer>( dstBuffer ),
|
||||
static_cast<VkDeviceSize>( dstOffset ),
|
||||
marker );
|
||||
}
|
||||
|
||||
//=== VK_EXT_calibrated_timestamps ===
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<VULKAN_HPP_NAMESPACE::TimeDomainKHR> PhysicalDevice::getCalibrateableTimeDomainsEXT() const
|
||||
|
|
@ -19483,6 +19828,26 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
return checkpointData;
|
||||
}
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<VULKAN_HPP_NAMESPACE::CheckpointData2NV> Queue::getCheckpointData2NV() const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkGetQueueCheckpointData2NV &&
|
||||
"Function <vkGetQueueCheckpointData2NV> requires <VK_NV_device_diagnostic_checkpoints>" );
|
||||
|
||||
std::vector<VULKAN_HPP_NAMESPACE::CheckpointData2NV> checkpointData;
|
||||
uint32_t checkpointDataCount;
|
||||
getDispatcher()->vkGetQueueCheckpointData2NV( static_cast<VkQueue>( m_queue ), &checkpointDataCount, nullptr );
|
||||
checkpointData.resize( checkpointDataCount );
|
||||
getDispatcher()->vkGetQueueCheckpointData2NV(
|
||||
static_cast<VkQueue>( m_queue ), &checkpointDataCount, reinterpret_cast<VkCheckpointData2NV *>( checkpointData.data() ) );
|
||||
|
||||
VULKAN_HPP_ASSERT( checkpointDataCount <= checkpointData.size() );
|
||||
if ( checkpointDataCount < checkpointData.size() )
|
||||
{
|
||||
checkpointData.resize( checkpointDataCount );
|
||||
}
|
||||
return checkpointData;
|
||||
}
|
||||
|
||||
//=== VK_KHR_timeline_semaphore ===
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE uint64_t Semaphore::getCounterValueKHR() const
|
||||
|
|
@ -20933,39 +21298,6 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
VULKAN_HPP_NAMESPACE::detail::resultCheck( result, VULKAN_HPP_NAMESPACE_STRING "::Queue::submit2KHR" );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void CommandBuffer::writeBufferMarker2AMD( VULKAN_HPP_NAMESPACE::PipelineStageFlags2 stage,
|
||||
VULKAN_HPP_NAMESPACE::Buffer dstBuffer,
|
||||
VULKAN_HPP_NAMESPACE::DeviceSize dstOffset,
|
||||
uint32_t marker ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdWriteBufferMarker2AMD && "Function <vkCmdWriteBufferMarker2AMD> requires <VK_KHR_synchronization2>" );
|
||||
|
||||
getDispatcher()->vkCmdWriteBufferMarker2AMD( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkPipelineStageFlags2>( stage ),
|
||||
static_cast<VkBuffer>( dstBuffer ),
|
||||
static_cast<VkDeviceSize>( dstOffset ),
|
||||
marker );
|
||||
}
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE std::vector<VULKAN_HPP_NAMESPACE::CheckpointData2NV> Queue::getCheckpointData2NV() const
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkGetQueueCheckpointData2NV && "Function <vkGetQueueCheckpointData2NV> requires <VK_KHR_synchronization2>" );
|
||||
|
||||
std::vector<VULKAN_HPP_NAMESPACE::CheckpointData2NV> checkpointData;
|
||||
uint32_t checkpointDataCount;
|
||||
getDispatcher()->vkGetQueueCheckpointData2NV( static_cast<VkQueue>( m_queue ), &checkpointDataCount, nullptr );
|
||||
checkpointData.resize( checkpointDataCount );
|
||||
getDispatcher()->vkGetQueueCheckpointData2NV(
|
||||
static_cast<VkQueue>( m_queue ), &checkpointDataCount, reinterpret_cast<VkCheckpointData2NV *>( checkpointData.data() ) );
|
||||
|
||||
VULKAN_HPP_ASSERT( checkpointDataCount <= checkpointData.size() );
|
||||
if ( checkpointDataCount < checkpointData.size() )
|
||||
{
|
||||
checkpointData.resize( checkpointDataCount );
|
||||
}
|
||||
return checkpointData;
|
||||
}
|
||||
|
||||
//=== VK_EXT_descriptor_buffer ===
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::DeviceSize DescriptorSetLayout::getSizeEXT() const VULKAN_HPP_NOEXCEPT
|
||||
|
|
@ -21252,8 +21584,9 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
Dispatch const & d ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( d.getVkHeaderVersion() == VK_HEADER_VERSION );
|
||||
return static_cast<Result>( d.vkGetDeviceFaultInfoEXT(
|
||||
m_device, reinterpret_cast<VkDeviceFaultCountsEXT *>( pFaultCounts ), reinterpret_cast<VkDeviceFaultInfoEXT *>( pFaultInfo ) ) );
|
||||
return static_cast<Result>( d.vkGetDeviceFaultInfoEXT( static_cast<VkDevice>( m_device ),
|
||||
reinterpret_cast<VkDeviceFaultCountsEXT *>( pFaultCounts ),
|
||||
reinterpret_cast<VkDeviceFaultInfoEXT *>( pFaultInfo ) ) );
|
||||
}
|
||||
# if defined( VK_USE_PLATFORM_WIN32_KHR )
|
||||
//=== VK_NV_acquire_winrt_display ===
|
||||
|
|
@ -22782,6 +23115,19 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
reinterpret_cast<const VkShaderEXT *>( shaders.data() ) );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void
|
||||
CommandBuffer::setDepthClampRangeEXT( VULKAN_HPP_NAMESPACE::DepthClampModeEXT depthClampMode,
|
||||
Optional<const VULKAN_HPP_NAMESPACE::DepthClampRangeEXT> depthClampRange ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdSetDepthClampRangeEXT &&
|
||||
"Function <vkCmdSetDepthClampRangeEXT> requires <VK_EXT_depth_clamp_control> or <VK_EXT_shader_object>" );
|
||||
|
||||
getDispatcher()->vkCmdSetDepthClampRangeEXT(
|
||||
static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkDepthClampModeEXT>( depthClampMode ),
|
||||
reinterpret_cast<const VkDepthClampRangeEXT *>( static_cast<const VULKAN_HPP_NAMESPACE::DepthClampRangeEXT *>( depthClampRange ) ) );
|
||||
}
|
||||
|
||||
//=== VK_KHR_pipeline_binary ===
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE
|
||||
|
|
@ -23212,6 +23558,135 @@ namespace VULKAN_HPP_NAMESPACE
|
|||
reinterpret_cast<const VkBindDescriptorBufferEmbeddedSamplersInfoEXT *>( &bindDescriptorBufferEmbeddedSamplersInfo ) );
|
||||
}
|
||||
|
||||
//=== VK_EXT_device_generated_commands ===
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::MemoryRequirements2 Device::getGeneratedCommandsMemoryRequirementsEXT(
|
||||
const VULKAN_HPP_NAMESPACE::GeneratedCommandsMemoryRequirementsInfoEXT & info ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkGetGeneratedCommandsMemoryRequirementsEXT &&
|
||||
"Function <vkGetGeneratedCommandsMemoryRequirementsEXT> requires <VK_EXT_device_generated_commands>" );
|
||||
|
||||
VULKAN_HPP_NAMESPACE::MemoryRequirements2 memoryRequirements;
|
||||
getDispatcher()->vkGetGeneratedCommandsMemoryRequirementsEXT( static_cast<VkDevice>( m_device ),
|
||||
reinterpret_cast<const VkGeneratedCommandsMemoryRequirementsInfoEXT *>( &info ),
|
||||
reinterpret_cast<VkMemoryRequirements2 *>( &memoryRequirements ) );
|
||||
|
||||
return memoryRequirements;
|
||||
}
|
||||
|
||||
template <typename X, typename Y, typename... Z>
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> Device::getGeneratedCommandsMemoryRequirementsEXT(
|
||||
const VULKAN_HPP_NAMESPACE::GeneratedCommandsMemoryRequirementsInfoEXT & info ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkGetGeneratedCommandsMemoryRequirementsEXT &&
|
||||
"Function <vkGetGeneratedCommandsMemoryRequirementsEXT> requires <VK_EXT_device_generated_commands>" );
|
||||
|
||||
VULKAN_HPP_NAMESPACE::StructureChain<X, Y, Z...> structureChain;
|
||||
VULKAN_HPP_NAMESPACE::MemoryRequirements2 & memoryRequirements = structureChain.template get<VULKAN_HPP_NAMESPACE::MemoryRequirements2>();
|
||||
getDispatcher()->vkGetGeneratedCommandsMemoryRequirementsEXT( static_cast<VkDevice>( m_device ),
|
||||
reinterpret_cast<const VkGeneratedCommandsMemoryRequirementsInfoEXT *>( &info ),
|
||||
reinterpret_cast<VkMemoryRequirements2 *>( &memoryRequirements ) );
|
||||
|
||||
return structureChain;
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void CommandBuffer::preprocessGeneratedCommandsEXT( const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoEXT & generatedCommandsInfo,
|
||||
VULKAN_HPP_NAMESPACE::CommandBuffer stateCommandBuffer ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdPreprocessGeneratedCommandsEXT &&
|
||||
"Function <vkCmdPreprocessGeneratedCommandsEXT> requires <VK_EXT_device_generated_commands>" );
|
||||
|
||||
getDispatcher()->vkCmdPreprocessGeneratedCommandsEXT( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
reinterpret_cast<const VkGeneratedCommandsInfoEXT *>( &generatedCommandsInfo ),
|
||||
static_cast<VkCommandBuffer>( stateCommandBuffer ) );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void
|
||||
CommandBuffer::executeGeneratedCommandsEXT( VULKAN_HPP_NAMESPACE::Bool32 isPreprocessed,
|
||||
const VULKAN_HPP_NAMESPACE::GeneratedCommandsInfoEXT & generatedCommandsInfo ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkCmdExecuteGeneratedCommandsEXT &&
|
||||
"Function <vkCmdExecuteGeneratedCommandsEXT> requires <VK_EXT_device_generated_commands>" );
|
||||
|
||||
getDispatcher()->vkCmdExecuteGeneratedCommandsEXT( static_cast<VkCommandBuffer>( m_commandBuffer ),
|
||||
static_cast<VkBool32>( isPreprocessed ),
|
||||
reinterpret_cast<const VkGeneratedCommandsInfoEXT *>( &generatedCommandsInfo ) );
|
||||
}
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::CreateReturnType<VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::IndirectCommandsLayoutEXT>::Type
|
||||
Device::createIndirectCommandsLayoutEXT( VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutCreateInfoEXT const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator ) const
|
||||
VULKAN_HPP_RAII_CREATE_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_NAMESPACE::IndirectCommandsLayoutEXT indirectCommandsLayout;
|
||||
VULKAN_HPP_NAMESPACE::Result result = static_cast<VULKAN_HPP_NAMESPACE::Result>( getDispatcher()->vkCreateIndirectCommandsLayoutEXT(
|
||||
static_cast<VkDevice>( m_device ),
|
||||
reinterpret_cast<const VkIndirectCommandsLayoutCreateInfoEXT *>( &createInfo ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) ),
|
||||
reinterpret_cast<VkIndirectCommandsLayoutEXT *>( &indirectCommandsLayout ) ) );
|
||||
if ( result != VULKAN_HPP_NAMESPACE::Result::eSuccess )
|
||||
{
|
||||
# if defined( VULKAN_HPP_RAII_NO_EXCEPTIONS )
|
||||
return VULKAN_HPP_UNEXPECTED( result );
|
||||
# else
|
||||
VULKAN_HPP_NAMESPACE::detail::throwResultException( result, "Device::createIndirectCommandsLayoutEXT" );
|
||||
# endif
|
||||
}
|
||||
|
||||
return VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::IndirectCommandsLayoutEXT(
|
||||
*this, *reinterpret_cast<VkIndirectCommandsLayoutEXT *>( &indirectCommandsLayout ), allocator );
|
||||
}
|
||||
|
||||
VULKAN_HPP_NODISCARD VULKAN_HPP_INLINE
|
||||
VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::CreateReturnType<VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::IndirectExecutionSetEXT>::Type
|
||||
Device::createIndirectExecutionSetEXT( VULKAN_HPP_NAMESPACE::IndirectExecutionSetCreateInfoEXT const & createInfo,
|
||||
VULKAN_HPP_NAMESPACE::Optional<const VULKAN_HPP_NAMESPACE::AllocationCallbacks> allocator ) const
|
||||
VULKAN_HPP_RAII_CREATE_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_NAMESPACE::IndirectExecutionSetEXT indirectExecutionSet;
|
||||
VULKAN_HPP_NAMESPACE::Result result = static_cast<VULKAN_HPP_NAMESPACE::Result>( getDispatcher()->vkCreateIndirectExecutionSetEXT(
|
||||
static_cast<VkDevice>( m_device ),
|
||||
reinterpret_cast<const VkIndirectExecutionSetCreateInfoEXT *>( &createInfo ),
|
||||
reinterpret_cast<const VkAllocationCallbacks *>( static_cast<const VULKAN_HPP_NAMESPACE::AllocationCallbacks *>( allocator ) ),
|
||||
reinterpret_cast<VkIndirectExecutionSetEXT *>( &indirectExecutionSet ) ) );
|
||||
if ( result != VULKAN_HPP_NAMESPACE::Result::eSuccess )
|
||||
{
|
||||
# if defined( VULKAN_HPP_RAII_NO_EXCEPTIONS )
|
||||
return VULKAN_HPP_UNEXPECTED( result );
|
||||
# else
|
||||
VULKAN_HPP_NAMESPACE::detail::throwResultException( result, "Device::createIndirectExecutionSetEXT" );
|
||||
# endif
|
||||
}
|
||||
|
||||
return VULKAN_HPP_NAMESPACE::VULKAN_HPP_RAII_NAMESPACE::IndirectExecutionSetEXT(
|
||||
*this, *reinterpret_cast<VkIndirectExecutionSetEXT *>( &indirectExecutionSet ), allocator );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void IndirectExecutionSetEXT::updatePipeline(
|
||||
VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::WriteIndirectExecutionSetPipelineEXT> const & executionSetWrites ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkUpdateIndirectExecutionSetPipelineEXT &&
|
||||
"Function <vkUpdateIndirectExecutionSetPipelineEXT> requires <VK_EXT_device_generated_commands>" );
|
||||
|
||||
getDispatcher()->vkUpdateIndirectExecutionSetPipelineEXT( static_cast<VkDevice>( m_device ),
|
||||
static_cast<VkIndirectExecutionSetEXT>( m_indirectExecutionSet ),
|
||||
executionSetWrites.size(),
|
||||
reinterpret_cast<const VkWriteIndirectExecutionSetPipelineEXT *>( executionSetWrites.data() ) );
|
||||
}
|
||||
|
||||
VULKAN_HPP_INLINE void IndirectExecutionSetEXT::updateShader(
|
||||
VULKAN_HPP_NAMESPACE::ArrayProxy<const VULKAN_HPP_NAMESPACE::WriteIndirectExecutionSetShaderEXT> const & executionSetWrites ) const VULKAN_HPP_NOEXCEPT
|
||||
{
|
||||
VULKAN_HPP_ASSERT( getDispatcher()->vkUpdateIndirectExecutionSetShaderEXT &&
|
||||
"Function <vkUpdateIndirectExecutionSetShaderEXT> requires <VK_EXT_device_generated_commands>" );
|
||||
|
||||
getDispatcher()->vkUpdateIndirectExecutionSetShaderEXT( static_cast<VkDevice>( m_device ),
|
||||
static_cast<VkIndirectExecutionSetEXT>( m_indirectExecutionSet ),
|
||||
executionSetWrites.size(),
|
||||
reinterpret_cast<const VkWriteIndirectExecutionSetShaderEXT *>( executionSetWrites.data() ) );
|
||||
}
|
||||
|
||||
//====================
|
||||
//=== RAII Helpers ===
|
||||
//====================
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue