├── .gitmodules ├── CMakeLists.txt ├── LICENSE ├── NOTICE ├── README.md ├── deps ├── D3DX12 │ ├── LICENSE │ └── d3dx12.h ├── stdx │ ├── CMakeLists.txt │ ├── README.md │ └── include │ │ └── stdx │ │ ├── hash.hpp │ │ ├── lib.hpp │ │ ├── match.hpp │ │ ├── range.hpp │ │ └── types.hpp └── vulkan │ ├── vk_icd.h │ ├── vk_layer.h │ ├── vk_platform.h │ ├── vk_sdk_platform.h │ ├── vulkan.h │ └── vulkan.hpp ├── manifest └── rostkatze.json └── src ├── command_list.cpp ├── command_list.hpp ├── command_recorder.hpp ├── descriptors_cpu.hpp ├── descriptors_gpu.hpp ├── descriptors_virtual.hpp ├── device.hpp ├── formats.cpp ├── icd.hpp ├── impl.cpp ├── impl.hpp ├── lib.cpp └── wsi.cpp /.gitmodules: -------------------------------------------------------------------------------- 1 | [submodule "deps/spirv-cross"] 2 | path = deps/spirv-cross 3 | url = https://github.com/msiglreith/SPIRV-Cross.git 4 | [submodule "deps/gsl"] 5 | path = deps/gsl 6 | url = https://github.com/Microsoft/GSL.git 7 | [submodule "deps/spdlog"] 8 | path = deps/spdlog 9 | url = https://github.com/gabime/spdlog.git 10 | -------------------------------------------------------------------------------- /CMakeLists.txt: -------------------------------------------------------------------------------- 1 | cmake_minimum_required(VERSION 2.8) 2 | 3 | project(rostkatze) 4 | 5 | add_compile_options("/std:c++latest") 6 | 7 | include_directories(include) 8 | include_directories(deps) 9 | include_directories(deps/stdx/external/fmt) 10 | include_directories(deps/stdx/include) 11 | 12 | ## dependencies 13 | # stdx 14 | add_subdirectory(deps/stdx) 15 | # gsl 16 | include_directories(deps/gsl/include) 17 | # vulkan 18 | include_directories(deps/vulkan) 19 | # spdlog 20 | include_directories(deps/spdlog/include) 21 | # SPIRV-Cross 22 | add_subdirectory(deps/spirv-cross) 23 | # gsl 24 | include_directories(deps/D3DX12) 25 | 26 | add_library(rostkatze SHARED 27 | src/command_list.cpp 28 | src/formats.cpp 29 | src/lib.cpp 30 | src/impl.cpp 31 | src/wsi.cpp 32 | ) 33 | 34 | target_link_libraries(rostkatze dxgi d3d12 d3dcompiler spirv-cross-hlsl) 35 | -------------------------------------------------------------------------------- /LICENSE: -------------------------------------------------------------------------------- 1 | Apache License 2 | Version 2.0, January 2004 3 | http://www.apache.org/licenses/ 4 | 5 | TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION 6 | 7 | 1. Definitions. 8 | 9 | "License" shall mean the terms and conditions for use, reproduction, 10 | and distribution as defined by Sections 1 through 9 of this document. 11 | 12 | "Licensor" shall mean the copyright owner or entity authorized by 13 | the copyright owner that is granting the License. 14 | 15 | "Legal Entity" shall mean the union of the acting entity and all 16 | other entities that control, are controlled by, or are under common 17 | control with that entity. For the purposes of this definition, 18 | "control" means (i) the power, direct or indirect, to cause the 19 | direction or management of such entity, whether by contract or 20 | otherwise, or (ii) ownership of fifty percent (50%) or more of the 21 | outstanding shares, or (iii) beneficial ownership of such entity. 22 | 23 | "You" (or "Your") shall mean an individual or Legal Entity 24 | exercising permissions granted by this License. 25 | 26 | "Source" form shall mean the preferred form for making modifications, 27 | including but not limited to software source code, documentation 28 | source, and configuration files. 29 | 30 | "Object" form shall mean any form resulting from mechanical 31 | transformation or translation of a Source form, including but 32 | not limited to compiled object code, generated documentation, 33 | and conversions to other media types. 34 | 35 | "Work" shall mean the work of authorship, whether in Source or 36 | Object form, made available under the License, as indicated by a 37 | copyright notice that is included in or attached to the work 38 | (an example is provided in the Appendix below). 39 | 40 | "Derivative Works" shall mean any work, whether in Source or Object 41 | form, that is based on (or derived from) the Work and for which the 42 | editorial revisions, annotations, elaborations, or other modifications 43 | represent, as a whole, an original work of authorship. For the purposes 44 | of this License, Derivative Works shall not include works that remain 45 | separable from, or merely link (or bind by name) to the interfaces of, 46 | the Work and Derivative Works thereof. 47 | 48 | "Contribution" shall mean any work of authorship, including 49 | the original version of the Work and any modifications or additions 50 | to that Work or Derivative Works thereof, that is intentionally 51 | submitted to Licensor for inclusion in the Work by the copyright owner 52 | or by an individual or Legal Entity authorized to submit on behalf of 53 | the copyright owner. For the purposes of this definition, "submitted" 54 | means any form of electronic, verbal, or written communication sent 55 | to the Licensor or its representatives, including but not limited to 56 | communication on electronic mailing lists, source code control systems, 57 | and issue tracking systems that are managed by, or on behalf of, the 58 | Licensor for the purpose of discussing and improving the Work, but 59 | excluding communication that is conspicuously marked or otherwise 60 | designated in writing by the copyright owner as "Not a Contribution." 61 | 62 | "Contributor" shall mean Licensor and any individual or Legal Entity 63 | on behalf of whom a Contribution has been received by Licensor and 64 | subsequently incorporated within the Work. 65 | 66 | 2. Grant of Copyright License. Subject to the terms and conditions of 67 | this License, each Contributor hereby grants to You a perpetual, 68 | worldwide, non-exclusive, no-charge, royalty-free, irrevocable 69 | copyright license to reproduce, prepare Derivative Works of, 70 | publicly display, publicly perform, sublicense, and distribute the 71 | Work and such Derivative Works in Source or Object form. 72 | 73 | 3. Grant of Patent License. Subject to the terms and conditions of 74 | this License, each Contributor hereby grants to You a perpetual, 75 | worldwide, non-exclusive, no-charge, royalty-free, irrevocable 76 | (except as stated in this section) patent license to make, have made, 77 | use, offer to sell, sell, import, and otherwise transfer the Work, 78 | where such license applies only to those patent claims licensable 79 | by such Contributor that are necessarily infringed by their 80 | Contribution(s) alone or by combination of their Contribution(s) 81 | with the Work to which such Contribution(s) was submitted. If You 82 | institute patent litigation against any entity (including a 83 | cross-claim or counterclaim in a lawsuit) alleging that the Work 84 | or a Contribution incorporated within the Work constitutes direct 85 | or contributory patent infringement, then any patent licenses 86 | granted to You under this License for that Work shall terminate 87 | as of the date such litigation is filed. 88 | 89 | 4. Redistribution. You may reproduce and distribute copies of the 90 | Work or Derivative Works thereof in any medium, with or without 91 | modifications, and in Source or Object form, provided that You 92 | meet the following conditions: 93 | 94 | (a) You must give any other recipients of the Work or 95 | Derivative Works a copy of this License; and 96 | 97 | (b) You must cause any modified files to carry prominent notices 98 | stating that You changed the files; and 99 | 100 | (c) You must retain, in the Source form of any Derivative Works 101 | that You distribute, all copyright, patent, trademark, and 102 | attribution notices from the Source form of the Work, 103 | excluding those notices that do not pertain to any part of 104 | the Derivative Works; and 105 | 106 | (d) If the Work includes a "NOTICE" text file as part of its 107 | distribution, then any Derivative Works that You distribute must 108 | include a readable copy of the attribution notices contained 109 | within such NOTICE file, excluding those notices that do not 110 | pertain to any part of the Derivative Works, in at least one 111 | of the following places: within a NOTICE text file distributed 112 | as part of the Derivative Works; within the Source form or 113 | documentation, if provided along with the Derivative Works; or, 114 | within a display generated by the Derivative Works, if and 115 | wherever such third-party notices normally appear. The contents 116 | of the NOTICE file are for informational purposes only and 117 | do not modify the License. You may add Your own attribution 118 | notices within Derivative Works that You distribute, alongside 119 | or as an addendum to the NOTICE text from the Work, provided 120 | that such additional attribution notices cannot be construed 121 | as modifying the License. 122 | 123 | You may add Your own copyright statement to Your modifications and 124 | may provide additional or different license terms and conditions 125 | for use, reproduction, or distribution of Your modifications, or 126 | for any such Derivative Works as a whole, provided Your use, 127 | reproduction, and distribution of the Work otherwise complies with 128 | the conditions stated in this License. 129 | 130 | 5. Submission of Contributions. Unless You explicitly state otherwise, 131 | any Contribution intentionally submitted for inclusion in the Work 132 | by You to the Licensor shall be under the terms and conditions of 133 | this License, without any additional terms or conditions. 134 | Notwithstanding the above, nothing herein shall supersede or modify 135 | the terms of any separate license agreement you may have executed 136 | with Licensor regarding such Contributions. 137 | 138 | 6. Trademarks. This License does not grant permission to use the trade 139 | names, trademarks, service marks, or product names of the Licensor, 140 | except as required for reasonable and customary use in describing the 141 | origin of the Work and reproducing the content of the NOTICE file. 142 | 143 | 7. Disclaimer of Warranty. Unless required by applicable law or 144 | agreed to in writing, Licensor provides the Work (and each 145 | Contributor provides its Contributions) on an "AS IS" BASIS, 146 | WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or 147 | implied, including, without limitation, any warranties or conditions 148 | of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A 149 | PARTICULAR PURPOSE. You are solely responsible for determining the 150 | appropriateness of using or redistributing the Work and assume any 151 | risks associated with Your exercise of permissions under this License. 152 | 153 | 8. Limitation of Liability. In no event and under no legal theory, 154 | whether in tort (including negligence), contract, or otherwise, 155 | unless required by applicable law (such as deliberate and grossly 156 | negligent acts) or agreed to in writing, shall any Contributor be 157 | liable to You for damages, including any direct, indirect, special, 158 | incidental, or consequential damages of any character arising as a 159 | result of this License or out of the use or inability to use the 160 | Work (including but not limited to damages for loss of goodwill, 161 | work stoppage, computer failure or malfunction, or any and all 162 | other commercial damages or losses), even if such Contributor 163 | has been advised of the possibility of such damages. 164 | 165 | 9. Accepting Warranty or Additional Liability. While redistributing 166 | the Work or Derivative Works thereof, You may choose to offer, 167 | and charge a fee for, acceptance of support, warranty, indemnity, 168 | or other liability obligations and/or rights consistent with this 169 | License. However, in accepting such obligations, You may act only 170 | on Your own behalf and on Your sole responsibility, not on behalf 171 | of any other Contributor, and only if You agree to indemnify, 172 | defend, and hold each Contributor harmless for any liability 173 | incurred by, or claims asserted against, such Contributor by reason 174 | of your accepting any such warranty or additional liability. 175 | 176 | END OF TERMS AND CONDITIONS 177 | 178 | APPENDIX: How to apply the Apache License to your work. 179 | 180 | To apply the Apache License to your work, attach the following 181 | boilerplate notice, with the fields enclosed by brackets "{}" 182 | replaced with your own identifying information. (Don't include 183 | the brackets!) The text should be enclosed in the appropriate 184 | comment syntax for the file format. We also recommend that a 185 | file or class name and description of purpose be included on the 186 | same "printed page" as the copyright notice for easier 187 | identification within third-party archives. 188 | 189 | Copyright {yyyy} {name of copyright owner} 190 | 191 | Licensed under the Apache License, Version 2.0 (the "License"); 192 | you may not use this file except in compliance with the License. 193 | You may obtain a copy of the License at 194 | 195 | http://www.apache.org/licenses/LICENSE-2.0 196 | 197 | Unless required by applicable law or agreed to in writing, software 198 | distributed under the License is distributed on an "AS IS" BASIS, 199 | WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 200 | See the License for the specific language governing permissions and 201 | limitations under the License. 202 | -------------------------------------------------------------------------------- /NOTICE: -------------------------------------------------------------------------------- 1 | - libraries in deps might come with a different license 2 | - Using the texture copy splitting algorithm by NXT, see comments in code 3 | https://github.com/google/nxt-standalone/blob/master/src/backend/d3d12/TextureCopySplitter.cpp 4 | 5 | // Copyright 2017 The NXT Authors 6 | // 7 | // Licensed under the Apache License, Version 2.0 (the "License"); 8 | // you may not use this file except in compliance with the License. 9 | // You may obtain a copy of the License at 10 | // 11 | // http://www.apache.org/licenses/LICENSE-2.0 12 | // 13 | // Unless required by applicable law or agreed to in writing, software 14 | // distributed under the License is distributed on an "AS IS" BASIS, 15 | // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 16 | // See the License for the specific language governing permissions and 17 | // limitations under the License. 18 | -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- 1 | 2 | # Rostkatze 3 | 4 | An offspring of the gfx-rs project, implementing a C++ Vulkan driver on top of D3D12. 5 | Currently, lacks a lot functionality but can run quite a few Vulkan samples already. 6 | 7 | ## Missing pieces 8 | 9 | - Pipeline barriers 10 | - RenderPasses 11 | - Secondary command buffers 12 | - Simultaneous command buffer submission 13 | - Queries 14 | - Validation of the current implementation(!) 15 | - Resource Tier 1 support 16 | 17 | ... (100+ bullet points) 18 | 19 | ## Running 20 | 21 | To able to use it latest Windows 10 Update is required (atm) and a GPU with Resource Tier 2 hardware (no NVIDIA!). 22 | 23 | The library can be built with VS 2017 (C++17 support). In order to use the `rostkatze` ICD set the `VK_ICD_FILENAMES` variable to the path of `manifest/rostkatze.json`. 24 | Adjust the library path field in the manifest to the absolute path of the rostkatze shared library. 25 | 26 | Good luck! 27 | 28 | 29 | ## "Working" samples (SaschaWillems) 30 | 31 | - glTF-PBR 32 | - triangle 33 | - pipelines 34 | - texture 35 | - texturecubmap 36 | - texturearray 37 | - texturemipmapgen 38 | - mesh 39 | - specialization constants 40 | - offscreen 41 | - radialblur 42 | - textoverlay 43 | - particlefire (small particles only) 44 | - scenerendering 45 | - HDR 46 | - instancing 47 | - indirect drawing 48 | - shadowmapping 49 | - shadowmappingcascade 50 | - shadowmappingomni 51 | - skeletalanimation 52 | - bloom 53 | - deferred 54 | - pbrbasic 55 | - pbribl 56 | - pbrtexture 57 | - computeshaderparticles 58 | - computeshader 59 | - sphericalenvmapping 60 | - gears 61 | - distancefieldrendering 62 | - vulkanscene 63 | - imgui 64 | - multisampling 65 | - ssao 66 | - parallaxmapping 67 | - conservative rasterization 68 | 69 | ## Partially working 70 | 71 | - computecloth (bugs) 72 | - pushconstants (push constant array members) 73 | 74 | # Almost/Not working 75 | 76 | - multithreading (seconday buffers not implemented) 77 | - dynamic uniform buffers (not implemented) 78 | - occlusion queries (not implemented) 79 | - deferred shadows (geometry shader support) 80 | - n-body (bug in samples/UB/portability) 81 | - raytracing (reading structs from byteaddressbuffer not supported) 82 | - culllod (indirect drawing, num_workgroups builtin, ..) 83 | 84 | - All tessellation and geometry samples 85 | -------------------------------------------------------------------------------- /deps/D3DX12/LICENSE: -------------------------------------------------------------------------------- 1 | The MIT License (MIT) 2 | 3 | Copyright (c) 2015 Microsoft 4 | 5 | Permission is hereby granted, free of charge, to any person obtaining a copy 6 | of this software and associated documentation files (the "Software"), to deal 7 | in the Software without restriction, including without limitation the rights 8 | to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 9 | copies of the Software, and to permit persons to whom the Software is 10 | furnished to do so, subject to the following conditions: 11 | 12 | The above copyright notice and this permission notice shall be included in all 13 | copies or substantial portions of the Software. 14 | 15 | THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 | IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 | FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 18 | AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 19 | LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 20 | OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 21 | SOFTWARE. 22 | 23 | -------------------------------------------------------------------------------- /deps/stdx/CMakeLists.txt: -------------------------------------------------------------------------------- 1 | cmake_minimum_required(VERSION 2.8) 2 | 3 | project(stdx) 4 | 5 | include_directories(include) 6 | -------------------------------------------------------------------------------- /deps/stdx/README.md: -------------------------------------------------------------------------------- 1 | 2 | ## Documentation 3 | 4 | ``` 5 | doxygen && moxygen --anchors --output=doc/api.md doc/xml 6 | ``` 7 | -------------------------------------------------------------------------------- /deps/stdx/include/stdx/hash.hpp: -------------------------------------------------------------------------------- 1 | #pragma once 2 | 3 | #include "types.hpp" 4 | 5 | namespace stdx { 6 | 7 | static auto hash_combine(usize& seed) -> void { } 8 | 9 | template 10 | static auto hash_combine(usize& hash, T0 const& t0, Ts... ts) -> void { 11 | auto h0 { std::hash{}(t0) }; 12 | hash = hash * 37u + h0; 13 | hash_combine(hash, ts...); 14 | } 15 | 16 | } -------------------------------------------------------------------------------- /deps/stdx/include/stdx/lib.hpp: -------------------------------------------------------------------------------- 1 | 2 | #pragma once 3 | 4 | // Library prelude, including fundamental components 5 | #include "types.hpp" 6 | -------------------------------------------------------------------------------- /deps/stdx/include/stdx/match.hpp: -------------------------------------------------------------------------------- 1 | 2 | #pragma once 3 | 4 | namespace stdx { 5 | 6 | template 7 | struct match_t; 8 | 9 | template<> 10 | struct match_t<> {}; 11 | 12 | template 13 | struct match_t : T { 14 | match_t(T t) : 15 | T { std::move(t) } 16 | {} 17 | 18 | using T::operator(); 19 | }; 20 | 21 | template 22 | struct match_t : 23 | T0, 24 | match_t 25 | { 26 | match_t(T0 t0, Ts... ts) : 27 | T0 { std::move(t0) }, 28 | match_t { std::move(ts)... } 29 | {} 30 | 31 | using T0::operator(); 32 | using match_t::operator(); 33 | }; 34 | 35 | template 36 | auto match(Ts... ts) { 37 | return match_t(std::move(ts)...); 38 | } 39 | 40 | } // namespace stdx 41 | -------------------------------------------------------------------------------- /deps/stdx/include/stdx/range.hpp: -------------------------------------------------------------------------------- 1 | #pragma once 2 | 3 | namespace stdx { 4 | 5 | /// 6 | template 7 | class range_iterator { 8 | public: 9 | using value_type = T; 10 | using reference = T&; 11 | 12 | public: 13 | explicit range_iterator(T cur) : 14 | cur(cur) 15 | { } 16 | 17 | auto operator++ () -> range_iterator& { 18 | cur += 1; 19 | return *this; 20 | } 21 | 22 | auto operator++ (int) -> range_iterator { 23 | return range_iterator(this->cur + 1) 24 | } 25 | 26 | auto operator== (range_iterator const& rhs) const { 27 | return this->cur >= rhs.cur; 28 | } 29 | 30 | auto operator!= (range_iterator const& rhs) const { 31 | return this->cur < rhs.cur; 32 | } 33 | 34 | auto operator*() -> reference { return cur; } 35 | 36 | private: 37 | T cur; 38 | }; 39 | 40 | /// 41 | template 42 | class range_t { 43 | public: 44 | /// 45 | T _start; 46 | /// 47 | T _end; 48 | 49 | public: 50 | using iterator = range_iterator; 51 | 52 | public: 53 | /// 54 | range_t(T start, T end) : 55 | _start(start), 56 | _end(end) 57 | { } 58 | 59 | /// 60 | auto begin() const -> iterator { 61 | return range_iterator(this->_start); 62 | } 63 | 64 | /// 65 | auto end() const -> iterator { 66 | return range_iterator(this->_end); 67 | } 68 | }; 69 | 70 | template 71 | auto range(T start, T end) -> range_t { 72 | return range_t(start, end); 73 | } 74 | 75 | template 76 | auto range(T end) -> range_t { 77 | return range_t(T(0), end); 78 | } 79 | 80 | } // namespace stdx 81 | -------------------------------------------------------------------------------- /deps/stdx/include/stdx/types.hpp: -------------------------------------------------------------------------------- 1 | #pragma once 2 | 3 | #include 4 | 5 | namespace stdx { 6 | 7 | /// 8-bit signed integer. 8 | using i8 = int8_t; 9 | /// 16-bit signed integer. 10 | using i16 = int16_t; 11 | /// 32-bit signed integer. 12 | using i32 = int32_t; 13 | /// 64-bit signed integer. 14 | using i64 = int64_t; 15 | 16 | /// 8-bit unsigned integer. 17 | using u8 = uint8_t; 18 | /// 16-bit unsigned integer. 19 | using u16 = uint16_t; 20 | /// 32-bit unsigned integer. 21 | using u32 = uint32_t; 22 | /// 64-bit unsigned integer. 23 | using u64 = uint64_t; 24 | 25 | /// Unsigned integer which can hold an array index. 26 | using usize = size_t; 27 | 28 | constexpr auto operator "" _uz (unsigned long long int x) -> usize { 29 | return x; 30 | } 31 | 32 | } // namespace stdx 33 | -------------------------------------------------------------------------------- /deps/vulkan/vk_icd.h: -------------------------------------------------------------------------------- 1 | // 2 | // File: vk_icd.h 3 | // 4 | /* 5 | * Copyright (c) 2015-2016 The Khronos Group Inc. 6 | * Copyright (c) 2015-2016 Valve Corporation 7 | * Copyright (c) 2015-2016 LunarG, Inc. 8 | * 9 | * Licensed under the Apache License, Version 2.0 (the "License"); 10 | * you may not use this file except in compliance with the License. 11 | * You may obtain a copy of the License at 12 | * 13 | * http://www.apache.org/licenses/LICENSE-2.0 14 | * 15 | * Unless required by applicable law or agreed to in writing, software 16 | * distributed under the License is distributed on an "AS IS" BASIS, 17 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 18 | * See the License for the specific language governing permissions and 19 | * limitations under the License. 20 | * 21 | */ 22 | 23 | #ifndef VKICD_H 24 | #define VKICD_H 25 | 26 | #include "vulkan.h" 27 | #include 28 | 29 | // Loader-ICD version negotiation API. Versions add the following features: 30 | // Version 0 - Initial. Doesn't support vk_icdGetInstanceProcAddr 31 | // or vk_icdNegotiateLoaderICDInterfaceVersion. 32 | // Version 1 - Add support for vk_icdGetInstanceProcAddr. 33 | // Version 2 - Add Loader/ICD Interface version negotiation 34 | // via vk_icdNegotiateLoaderICDInterfaceVersion. 35 | // Version 3 - Add ICD creation/destruction of KHR_surface objects. 36 | // Version 4 - Add unknown physical device extension qyering via 37 | // vk_icdGetPhysicalDeviceProcAddr. 38 | // Version 5 - Tells ICDs that the loader is now paying attention to the 39 | // application version of Vulkan passed into the ApplicationInfo 40 | // structure during vkCreateInstance. This will tell the ICD 41 | // that if the loader is older, it should automatically fail a 42 | // call for any API version > 1.0. Otherwise, the loader will 43 | // manually determine if it can support the expected version. 44 | #define CURRENT_LOADER_ICD_INTERFACE_VERSION 5 45 | #define MIN_SUPPORTED_LOADER_ICD_INTERFACE_VERSION 0 46 | #define MIN_PHYS_DEV_EXTENSION_ICD_INTERFACE_VERSION 4 47 | typedef VkResult (VKAPI_PTR *PFN_vkNegotiateLoaderICDInterfaceVersion)(uint32_t *pVersion); 48 | 49 | // This is defined in vk_layer.h which will be found by the loader, but if an ICD is building against this 50 | // file directly, it won't be found. 51 | #ifndef PFN_GetPhysicalDeviceProcAddr 52 | typedef PFN_vkVoidFunction (VKAPI_PTR *PFN_GetPhysicalDeviceProcAddr)(VkInstance instance, const char* pName); 53 | #endif 54 | 55 | /* 56 | * The ICD must reserve space for a pointer for the loader's dispatch 57 | * table, at the start of . 58 | * The ICD must initialize this variable using the SET_LOADER_MAGIC_VALUE macro. 59 | */ 60 | 61 | #define ICD_LOADER_MAGIC 0x01CDC0DE 62 | 63 | typedef union { 64 | uintptr_t loaderMagic; 65 | void *loaderData; 66 | } VK_LOADER_DATA; 67 | 68 | static inline void set_loader_magic_value(void *pNewObject) { 69 | VK_LOADER_DATA *loader_info = (VK_LOADER_DATA *)pNewObject; 70 | loader_info->loaderMagic = ICD_LOADER_MAGIC; 71 | } 72 | 73 | static inline bool valid_loader_magic_value(void *pNewObject) { 74 | const VK_LOADER_DATA *loader_info = (VK_LOADER_DATA *)pNewObject; 75 | return (loader_info->loaderMagic & 0xffffffff) == ICD_LOADER_MAGIC; 76 | } 77 | 78 | /* 79 | * Windows and Linux ICDs will treat VkSurfaceKHR as a pointer to a struct that 80 | * contains the platform-specific connection and surface information. 81 | */ 82 | typedef enum { 83 | VK_ICD_WSI_PLATFORM_MIR, 84 | VK_ICD_WSI_PLATFORM_WAYLAND, 85 | VK_ICD_WSI_PLATFORM_WIN32, 86 | VK_ICD_WSI_PLATFORM_XCB, 87 | VK_ICD_WSI_PLATFORM_XLIB, 88 | VK_ICD_WSI_PLATFORM_DISPLAY 89 | } VkIcdWsiPlatform; 90 | 91 | typedef struct { 92 | VkIcdWsiPlatform platform; 93 | } VkIcdSurfaceBase; 94 | 95 | #ifdef VK_USE_PLATFORM_MIR_KHR 96 | typedef struct { 97 | VkIcdSurfaceBase base; 98 | MirConnection *connection; 99 | MirSurface *mirSurface; 100 | } VkIcdSurfaceMir; 101 | #endif // VK_USE_PLATFORM_MIR_KHR 102 | 103 | #ifdef VK_USE_PLATFORM_WAYLAND_KHR 104 | typedef struct { 105 | VkIcdSurfaceBase base; 106 | struct wl_display *display; 107 | struct wl_surface *surface; 108 | } VkIcdSurfaceWayland; 109 | #endif // VK_USE_PLATFORM_WAYLAND_KHR 110 | 111 | #ifdef VK_USE_PLATFORM_WIN32_KHR 112 | typedef struct { 113 | VkIcdSurfaceBase base; 114 | HINSTANCE hinstance; 115 | HWND hwnd; 116 | } VkIcdSurfaceWin32; 117 | #endif // VK_USE_PLATFORM_WIN32_KHR 118 | 119 | #ifdef VK_USE_PLATFORM_XCB_KHR 120 | typedef struct { 121 | VkIcdSurfaceBase base; 122 | xcb_connection_t *connection; 123 | xcb_window_t window; 124 | } VkIcdSurfaceXcb; 125 | #endif // VK_USE_PLATFORM_XCB_KHR 126 | 127 | #ifdef VK_USE_PLATFORM_XLIB_KHR 128 | typedef struct { 129 | VkIcdSurfaceBase base; 130 | Display *dpy; 131 | Window window; 132 | } VkIcdSurfaceXlib; 133 | #endif // VK_USE_PLATFORM_XLIB_KHR 134 | 135 | #ifdef VK_USE_PLATFORM_ANDROID_KHR 136 | typedef struct { 137 | ANativeWindow* window; 138 | } VkIcdSurfaceAndroid; 139 | #endif //VK_USE_PLATFORM_ANDROID_KHR 140 | 141 | typedef struct { 142 | VkIcdSurfaceBase base; 143 | VkDisplayModeKHR displayMode; 144 | uint32_t planeIndex; 145 | uint32_t planeStackIndex; 146 | VkSurfaceTransformFlagBitsKHR transform; 147 | float globalAlpha; 148 | VkDisplayPlaneAlphaFlagBitsKHR alphaMode; 149 | VkExtent2D imageExtent; 150 | } VkIcdSurfaceDisplay; 151 | 152 | #endif // VKICD_H 153 | -------------------------------------------------------------------------------- /deps/vulkan/vk_layer.h: -------------------------------------------------------------------------------- 1 | // 2 | // File: vk_layer.h 3 | // 4 | /* 5 | * Copyright (c) 2015-2017 The Khronos Group Inc. 6 | * Copyright (c) 2015-2017 Valve Corporation 7 | * Copyright (c) 2015-2017 LunarG, Inc. 8 | * 9 | * Licensed under the Apache License, Version 2.0 (the "License"); 10 | * you may not use this file except in compliance with the License. 11 | * You may obtain a copy of the License at 12 | * 13 | * http://www.apache.org/licenses/LICENSE-2.0 14 | * 15 | * Unless required by applicable law or agreed to in writing, software 16 | * distributed under the License is distributed on an "AS IS" BASIS, 17 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 18 | * See the License for the specific language governing permissions and 19 | * limitations under the License. 20 | * 21 | */ 22 | 23 | /* Need to define dispatch table 24 | * Core struct can then have ptr to dispatch table at the top 25 | * Along with object ptrs for current and next OBJ 26 | */ 27 | #pragma once 28 | 29 | #include "vulkan.h" 30 | #if defined(__GNUC__) && __GNUC__ >= 4 31 | #define VK_LAYER_EXPORT __attribute__((visibility("default"))) 32 | #elif defined(__SUNPRO_C) && (__SUNPRO_C >= 0x590) 33 | #define VK_LAYER_EXPORT __attribute__((visibility("default"))) 34 | #else 35 | #define VK_LAYER_EXPORT 36 | #endif 37 | 38 | // Definition for VkLayerDispatchTable and VkLayerInstanceDispatchTable now appear in externally generated header 39 | #include "vk_layer_dispatch_table.h" 40 | 41 | #define MAX_NUM_UNKNOWN_EXTS 250 42 | 43 | // Loader-Layer version negotiation API. Versions add the following features: 44 | // Versions 0/1 - Initial. Doesn't support vk_layerGetPhysicalDeviceProcAddr 45 | // or vk_icdNegotiateLoaderLayerInterfaceVersion. 46 | // Version 2 - Add support for vk_layerGetPhysicalDeviceProcAddr and 47 | // vk_icdNegotiateLoaderLayerInterfaceVersion. 48 | #define CURRENT_LOADER_LAYER_INTERFACE_VERSION 2 49 | #define MIN_SUPPORTED_LOADER_LAYER_INTERFACE_VERSION 1 50 | 51 | #define VK_CURRENT_CHAIN_VERSION 1 52 | 53 | // Version negotiation values 54 | typedef enum VkNegotiateLayerStructType { 55 | LAYER_NEGOTIATE_UNINTIALIZED = 0, 56 | LAYER_NEGOTIATE_INTERFACE_STRUCT = 1, 57 | } VkNegotiateLayerStructType; 58 | 59 | // Version negotiation structures 60 | typedef struct VkNegotiateLayerInterface { 61 | VkNegotiateLayerStructType sType; 62 | void *pNext; 63 | uint32_t loaderLayerInterfaceVersion; 64 | PFN_vkGetInstanceProcAddr pfnGetInstanceProcAddr; 65 | PFN_vkGetDeviceProcAddr pfnGetDeviceProcAddr; 66 | PFN_GetPhysicalDeviceProcAddr pfnGetPhysicalDeviceProcAddr; 67 | } VkNegotiateLayerInterface; 68 | 69 | // Version negotiation functions 70 | typedef VkResult (VKAPI_PTR *PFN_vkNegotiateLoaderLayerInterfaceVersion)(VkNegotiateLayerInterface *pVersionStruct); 71 | 72 | // Function prototype for unknown physical device extension command 73 | typedef VkResult(VKAPI_PTR *PFN_PhysDevExt)(VkPhysicalDevice phys_device); 74 | 75 | // ------------------------------------------------------------------------------------------------ 76 | // CreateInstance and CreateDevice support structures 77 | 78 | /* Sub type of structure for instance and device loader ext of CreateInfo. 79 | * When sType == VK_STRUCTURE_TYPE_LOADER_INSTANCE_CREATE_INFO 80 | * or sType == VK_STRUCTURE_TYPE_LOADER_DEVICE_CREATE_INFO 81 | * then VkLayerFunction indicates struct type pointed to by pNext 82 | */ 83 | typedef enum VkLayerFunction_ { 84 | VK_LAYER_LINK_INFO = 0, 85 | VK_LOADER_DATA_CALLBACK = 1 86 | } VkLayerFunction; 87 | 88 | typedef struct VkLayerInstanceLink_ { 89 | struct VkLayerInstanceLink_ *pNext; 90 | PFN_vkGetInstanceProcAddr pfnNextGetInstanceProcAddr; 91 | PFN_GetPhysicalDeviceProcAddr pfnNextGetPhysicalDeviceProcAddr; 92 | } VkLayerInstanceLink; 93 | 94 | /* 95 | * When creating the device chain the loader needs to pass 96 | * down information about it's device structure needed at 97 | * the end of the chain. Passing the data via the 98 | * VkLayerDeviceInfo avoids issues with finding the 99 | * exact instance being used. 100 | */ 101 | typedef struct VkLayerDeviceInfo_ { 102 | void *device_info; 103 | PFN_vkGetInstanceProcAddr pfnNextGetInstanceProcAddr; 104 | } VkLayerDeviceInfo; 105 | 106 | typedef VkResult (VKAPI_PTR *PFN_vkSetInstanceLoaderData)(VkInstance instance, 107 | void *object); 108 | typedef VkResult (VKAPI_PTR *PFN_vkSetDeviceLoaderData)(VkDevice device, 109 | void *object); 110 | 111 | typedef struct { 112 | VkStructureType sType; // VK_STRUCTURE_TYPE_LOADER_INSTANCE_CREATE_INFO 113 | const void *pNext; 114 | VkLayerFunction function; 115 | union { 116 | VkLayerInstanceLink *pLayerInfo; 117 | PFN_vkSetInstanceLoaderData pfnSetInstanceLoaderData; 118 | } u; 119 | } VkLayerInstanceCreateInfo; 120 | 121 | typedef struct VkLayerDeviceLink_ { 122 | struct VkLayerDeviceLink_ *pNext; 123 | PFN_vkGetInstanceProcAddr pfnNextGetInstanceProcAddr; 124 | PFN_vkGetDeviceProcAddr pfnNextGetDeviceProcAddr; 125 | } VkLayerDeviceLink; 126 | 127 | typedef struct { 128 | VkStructureType sType; // VK_STRUCTURE_TYPE_LOADER_DEVICE_CREATE_INFO 129 | const void *pNext; 130 | VkLayerFunction function; 131 | union { 132 | VkLayerDeviceLink *pLayerInfo; 133 | PFN_vkSetDeviceLoaderData pfnSetDeviceLoaderData; 134 | } u; 135 | } VkLayerDeviceCreateInfo; 136 | 137 | #ifdef __cplusplus 138 | extern "C" { 139 | #endif 140 | 141 | VKAPI_ATTR VkResult VKAPI_CALL vkNegotiateLoaderLayerInterfaceVersion(VkNegotiateLayerInterface *pVersionStruct); 142 | 143 | typedef enum VkChainType { 144 | VK_CHAIN_TYPE_UNKNOWN = 0, 145 | VK_CHAIN_TYPE_ENUMERATE_INSTANCE_EXTENSION_PROPERTIES = 1, 146 | VK_CHAIN_TYPE_ENUMERATE_INSTANCE_LAYER_PROPERTIES = 2, 147 | } VkChainType; 148 | 149 | typedef struct VkChainHeader { 150 | VkChainType type; 151 | uint32_t version; 152 | uint32_t size; 153 | } VkChainHeader; 154 | 155 | typedef struct VkEnumerateInstanceExtensionPropertiesChain { 156 | VkChainHeader header; 157 | VkResult(VKAPI_PTR *pfnNextLayer)(const struct VkEnumerateInstanceExtensionPropertiesChain *, const char *, uint32_t *, 158 | VkExtensionProperties *); 159 | const struct VkEnumerateInstanceExtensionPropertiesChain *pNextLink; 160 | 161 | #if defined(__cplusplus) 162 | inline VkResult CallDown(const char *pLayerName, uint32_t *pPropertyCount, VkExtensionProperties *pProperties) const { 163 | return pfnNextLayer(pNextLink, pLayerName, pPropertyCount, pProperties); 164 | } 165 | #endif 166 | } VkEnumerateInstanceExtensionPropertiesChain; 167 | 168 | typedef struct VkEnumerateInstanceLayerPropertiesChain { 169 | VkChainHeader header; 170 | VkResult(VKAPI_PTR *pfnNextLayer)(const struct VkEnumerateInstanceLayerPropertiesChain *, uint32_t *, VkLayerProperties *); 171 | const struct VkEnumerateInstanceLayerPropertiesChain *pNextLink; 172 | 173 | #if defined(__cplusplus) 174 | inline VkResult CallDown(uint32_t *pPropertyCount, VkLayerProperties *pProperties) const { 175 | return pfnNextLayer(pNextLink, pPropertyCount, pProperties); 176 | } 177 | #endif 178 | } VkEnumerateInstanceLayerPropertiesChain; 179 | 180 | #ifdef __cplusplus 181 | } 182 | #endif 183 | -------------------------------------------------------------------------------- /deps/vulkan/vk_platform.h: -------------------------------------------------------------------------------- 1 | // 2 | // File: vk_platform.h 3 | // 4 | /* 5 | ** Copyright (c) 2014-2017 The Khronos Group Inc. 6 | ** 7 | ** Licensed under the Apache License, Version 2.0 (the "License"); 8 | ** you may not use this file except in compliance with the License. 9 | ** You may obtain a copy of the License at 10 | ** 11 | ** http://www.apache.org/licenses/LICENSE-2.0 12 | ** 13 | ** Unless required by applicable law or agreed to in writing, software 14 | ** distributed under the License is distributed on an "AS IS" BASIS, 15 | ** WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 16 | ** See the License for the specific language governing permissions and 17 | ** limitations under the License. 18 | */ 19 | 20 | 21 | #ifndef VK_PLATFORM_H_ 22 | #define VK_PLATFORM_H_ 23 | 24 | #ifdef __cplusplus 25 | extern "C" 26 | { 27 | #endif // __cplusplus 28 | 29 | /* 30 | *************************************************************************************************** 31 | * Platform-specific directives and type declarations 32 | *************************************************************************************************** 33 | */ 34 | 35 | /* Platform-specific calling convention macros. 36 | * 37 | * Platforms should define these so that Vulkan clients call Vulkan commands 38 | * with the same calling conventions that the Vulkan implementation expects. 39 | * 40 | * VKAPI_ATTR - Placed before the return type in function declarations. 41 | * Useful for C++11 and GCC/Clang-style function attribute syntax. 42 | * VKAPI_CALL - Placed after the return type in function declarations. 43 | * Useful for MSVC-style calling convention syntax. 44 | * VKAPI_PTR - Placed between the '(' and '*' in function pointer types. 45 | * 46 | * Function declaration: VKAPI_ATTR void VKAPI_CALL vkCommand(void); 47 | * Function pointer type: typedef void (VKAPI_PTR *PFN_vkCommand)(void); 48 | */ 49 | #if defined(_WIN32) 50 | // On Windows, Vulkan commands use the stdcall convention 51 | #define VKAPI_ATTR // __declspec(dllexport) 52 | #define VKAPI_CALL __stdcall 53 | #define VKAPI_PTR VKAPI_CALL 54 | #elif defined(__ANDROID__) && defined(__ARM_ARCH) && __ARM_ARCH < 7 55 | #error "Vulkan isn't supported for the 'armeabi' NDK ABI" 56 | #elif defined(__ANDROID__) && defined(__ARM_ARCH) && __ARM_ARCH >= 7 && defined(__ARM_32BIT_STATE) 57 | // On Android 32-bit ARM targets, Vulkan functions use the "hardfloat" 58 | // calling convention, i.e. float parameters are passed in registers. This 59 | // is true even if the rest of the application passes floats on the stack, 60 | // as it does by default when compiling for the armeabi-v7a NDK ABI. 61 | #define VKAPI_ATTR __attribute__((pcs("aapcs-vfp"))) 62 | #define VKAPI_CALL 63 | #define VKAPI_PTR VKAPI_ATTR 64 | #else 65 | // On other platforms, use the default calling convention 66 | #define VKAPI_ATTR 67 | #define VKAPI_CALL 68 | #define VKAPI_PTR 69 | #endif 70 | 71 | #include 72 | 73 | #if !defined(VK_NO_STDINT_H) 74 | #if defined(_MSC_VER) && (_MSC_VER < 1600) 75 | typedef signed __int8 int8_t; 76 | typedef unsigned __int8 uint8_t; 77 | typedef signed __int16 int16_t; 78 | typedef unsigned __int16 uint16_t; 79 | typedef signed __int32 int32_t; 80 | typedef unsigned __int32 uint32_t; 81 | typedef signed __int64 int64_t; 82 | typedef unsigned __int64 uint64_t; 83 | #else 84 | #include 85 | #endif 86 | #endif // !defined(VK_NO_STDINT_H) 87 | 88 | #ifdef __cplusplus 89 | } // extern "C" 90 | #endif // __cplusplus 91 | 92 | // Platform-specific headers required by platform window system extensions. 93 | // These are enabled prior to #including "vulkan.h". The same enable then 94 | // controls inclusion of the extension interfaces in vulkan.h. 95 | 96 | #ifdef VK_USE_PLATFORM_ANDROID_KHR 97 | #include 98 | #endif 99 | 100 | #ifdef VK_USE_PLATFORM_MIR_KHR 101 | #include 102 | #endif 103 | 104 | #ifdef VK_USE_PLATFORM_WAYLAND_KHR 105 | #include 106 | #endif 107 | 108 | #ifdef VK_USE_PLATFORM_WIN32_KHR 109 | #include 110 | #endif 111 | 112 | #ifdef VK_USE_PLATFORM_XLIB_KHR 113 | #include 114 | #endif 115 | 116 | #ifdef VK_USE_PLATFORM_XCB_KHR 117 | #include 118 | #endif 119 | 120 | #endif 121 | -------------------------------------------------------------------------------- /deps/vulkan/vk_sdk_platform.h: -------------------------------------------------------------------------------- 1 | // 2 | // File: vk_sdk_platform.h 3 | // 4 | /* 5 | * Copyright (c) 2015-2016 The Khronos Group Inc. 6 | * Copyright (c) 2015-2016 Valve Corporation 7 | * Copyright (c) 2015-2016 LunarG, Inc. 8 | * 9 | * Licensed under the Apache License, Version 2.0 (the "License"); 10 | * you may not use this file except in compliance with the License. 11 | * You may obtain a copy of the License at 12 | * 13 | * http://www.apache.org/licenses/LICENSE-2.0 14 | * 15 | * Unless required by applicable law or agreed to in writing, software 16 | * distributed under the License is distributed on an "AS IS" BASIS, 17 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 18 | * See the License for the specific language governing permissions and 19 | * limitations under the License. 20 | */ 21 | 22 | #ifndef VK_SDK_PLATFORM_H 23 | #define VK_SDK_PLATFORM_H 24 | 25 | #if defined(_WIN32) 26 | #define NOMINMAX 27 | #ifndef __cplusplus 28 | #undef inline 29 | #define inline __inline 30 | #endif // __cplusplus 31 | 32 | #if (defined(_MSC_VER) && _MSC_VER < 1900 /*vs2015*/) 33 | // C99: 34 | // Microsoft didn't implement C99 in Visual Studio; but started adding it with 35 | // VS2013. However, VS2013 still didn't have snprintf(). The following is a 36 | // work-around (Note: The _CRT_SECURE_NO_WARNINGS macro must be set in the 37 | // "CMakeLists.txt" file). 38 | // NOTE: This is fixed in Visual Studio 2015. 39 | #define snprintf _snprintf 40 | #endif 41 | 42 | #define strdup _strdup 43 | 44 | #endif // _WIN32 45 | 46 | // Check for noexcept support using clang, with fallback to Windows or GCC version numbers 47 | #ifndef NOEXCEPT 48 | #if defined(__clang__) 49 | #if __has_feature(cxx_noexcept) 50 | #define HAS_NOEXCEPT 51 | #endif 52 | #else 53 | #if defined(__GXX_EXPERIMENTAL_CXX0X__) && __GNUC__ * 10 + __GNUC_MINOR__ >= 46 54 | #define HAS_NOEXCEPT 55 | #else 56 | #if defined(_MSC_FULL_VER) && _MSC_FULL_VER >= 190023026 && defined(_HAS_EXCEPTIONS) && _HAS_EXCEPTIONS 57 | #define HAS_NOEXCEPT 58 | #endif 59 | #endif 60 | #endif 61 | 62 | #ifdef HAS_NOEXCEPT 63 | #define NOEXCEPT noexcept 64 | #else 65 | #define NOEXCEPT 66 | #endif 67 | #endif 68 | 69 | #endif // VK_SDK_PLATFORM_H 70 | -------------------------------------------------------------------------------- /manifest/rostkatze.json: -------------------------------------------------------------------------------- 1 | { 2 | "file_format_version" : "1.0.1", 3 | "ICD": { 4 | "library_path": " 9 | #include 10 | 11 | enum class draw_type { 12 | DRAW, 13 | DRAW_INDEXED, 14 | }; 15 | 16 | class command_buffer_t { 17 | public: 18 | struct pass_cache_t { 19 | // Current subpass 20 | size_t subpass; 21 | // 22 | render_pass_t* render_pass; 23 | framebuffer_t* framebuffer; 24 | // 25 | std::vector clear_values; 26 | D3D12_RECT render_area; 27 | }; 28 | 29 | enum pipeline_slot_type { 30 | SLOT_GRAPHICS, 31 | SLOT_COMPUTE, 32 | }; 33 | 34 | struct pipeline_slot_t { 35 | public: 36 | enum class data_entry_type { 37 | UNDEFINED, // TODO: Not needed probably 38 | SAMPLER, 39 | CBV_SRV_UAV, 40 | CONSTANT, 41 | }; 42 | 43 | // Virtual root user data 44 | struct user_data_t { 45 | public: 46 | user_data_t() : 47 | dirty { 0 }, 48 | data { 0 }, 49 | type { data_entry_type::UNDEFINED } 50 | { } 51 | 52 | auto set_constant(size_t slot, uint32_t data) -> void { 53 | this->dirty.set(slot); 54 | this->data[slot] = data; 55 | this->type[slot] = data_entry_type::CONSTANT; 56 | } 57 | 58 | auto set_cbv_srv_uav(size_t slot, uint32_t data) -> void { 59 | this->dirty.set(slot); 60 | this->data[slot] = data; 61 | this->type[slot] = data_entry_type::CBV_SRV_UAV; 62 | } 63 | 64 | auto set_sampler(size_t slot, uint32_t data) -> void { 65 | this->dirty.set(slot); 66 | this->data[slot] = data; 67 | this->type[slot] = data_entry_type::SAMPLER; 68 | } 69 | 70 | public: 71 | std::bitset dirty; 72 | std::array data; 73 | std::array type; 74 | }; 75 | 76 | public: 77 | pipeline_slot_t() : 78 | pipeline { nullptr }, 79 | signature { nullptr }, 80 | root_data { } 81 | {} 82 | 83 | public: 84 | struct pipeline_t* pipeline; 85 | ID3D12RootSignature* signature; 86 | 87 | user_data_t root_data; 88 | }; 89 | 90 | public: 91 | command_buffer_t( 92 | ID3D12CommandAllocator* allocator, 93 | ComPtr cmd_list, 94 | device_t* device, 95 | ID3D12DescriptorHeap* heap_gpu_cbv_srv_uav, 96 | ID3D12DescriptorHeap* heap_cpu_sampler 97 | ) : 98 | loader_magic { ICD_LOADER_MAGIC }, 99 | allocator { allocator }, 100 | command_list { cmd_list }, 101 | command_recorder { cmd_list.Get() }, 102 | heap_gpu_cbv_srv_uav { heap_gpu_cbv_srv_uav }, 103 | heap_cpu_sampler { heap_cpu_sampler }, 104 | heap_gpu_sampler { nullptr }, 105 | pass_cache { std::nullopt }, 106 | active_slot { std::nullopt }, 107 | active_pipeline { nullptr }, 108 | dynamic_state_dirty { false }, 109 | viewports_dirty { false }, 110 | scissors_dirty { false }, 111 | num_viewports_scissors { 0 }, 112 | vertex_buffer_views_dirty { 0 }, 113 | _device { device }, 114 | dynamic_state { }, 115 | index_type { VK_INDEX_TYPE_UINT16 } 116 | { } 117 | 118 | ~command_buffer_t() {} 119 | 120 | public: 121 | auto reset() -> void { 122 | this->command_list->Reset(this->allocator, nullptr); 123 | 124 | this->pass_cache = std::nullopt; 125 | this->active_slot = std::nullopt; 126 | this->active_pipeline = nullptr; 127 | this->graphics_slot = pipeline_slot_t(); 128 | this->compute_slot = pipeline_slot_t(); 129 | this->num_viewports_scissors = 0; 130 | this->viewports_dirty = false; 131 | this->scissors_dirty = false; 132 | this->vertex_buffer_views_dirty = 0; 133 | this->dynamic_state = dynamic_state_t(); 134 | this->dynamic_state_dirty = false; 135 | this->index_type = VK_INDEX_TYPE_UINT16; 136 | } 137 | 138 | auto end() -> void { 139 | this->command_list->Close(); 140 | // TODO: error handling 141 | } 142 | 143 | auto device() -> ID3D12Device3* { 144 | return this->_device->device.Get(); 145 | } 146 | 147 | auto bind_descriptor_heaps() { 148 | std::array heaps { 149 | this->heap_gpu_cbv_srv_uav, 150 | this->heap_gpu_sampler 151 | }; 152 | this->command_recorder.cmd_set_descriptor_heaps( 153 | this->heap_gpu_sampler ? 2 : 1, 154 | &heaps[0] 155 | ); 156 | } 157 | 158 | auto update_user_data( 159 | pipeline_slot_t& slot, 160 | std::function set_constant, 161 | std::function set_table 162 | ) { 163 | auto user_data { slot.root_data }; 164 | if (user_data.dirty.none()) { 165 | return; 166 | } 167 | 168 | const auto start_cbv_srv_uav { this->heap_gpu_cbv_srv_uav->GetGPUDescriptorHandleForHeapStart() }; 169 | const auto start_cpu_sampler { this->heap_cpu_sampler->GetCPUDescriptorHandleForHeapStart() }; 170 | auto start_sampler { 171 | this->heap_gpu_sampler ? 172 | this->heap_gpu_sampler->GetGPUDescriptorHandleForHeapStart() : 173 | D3D12_GPU_DESCRIPTOR_HANDLE { 0 } 174 | }; 175 | 176 | const auto num_root_constant_entries { slot.pipeline->root_constants.size() }; 177 | const auto num_dynamic_offsets { slot.pipeline->num_dynamic_offsets }; 178 | const auto num_table_entries { slot.pipeline->num_signature_entries - num_root_constant_entries - num_dynamic_offsets }; 179 | const auto descriptor_table_start_slot { slot.pipeline->num_root_constants + num_dynamic_offsets }; 180 | 181 | auto start_constant { 0u }; 182 | for (auto i : range(num_root_constant_entries)) { 183 | auto const& root_constant { slot.pipeline->root_constants[i] }; 184 | 185 | for (auto c : range(start_constant, start_constant + root_constant.size / 4)) { 186 | if (user_data.dirty[c]) { 187 | set_constant( 188 | static_cast(i), 189 | user_data.data[c], 190 | c - start_constant 191 | ); 192 | user_data.dirty.reset(c); 193 | } 194 | } 195 | 196 | start_constant += root_constant.size / 4; 197 | } 198 | 199 | std::bitset active_sampler_sets { 0 }; 200 | for (auto i : range(num_table_entries)) { 201 | const auto data_slot { descriptor_table_start_slot + i }; 202 | if (!user_data.dirty[data_slot]) { 203 | continue; 204 | } 205 | 206 | if (user_data.type[data_slot] == pipeline_slot_t::data_entry_type::SAMPLER) { 207 | // Only cache samplers as they are currently living inside a CPU heap only. 208 | active_sampler_sets.set(i); 209 | } 210 | } 211 | 212 | if (active_sampler_sets.any()) { 213 | // TODO: thread safe access to descriptors 214 | auto num_sampler_sets { 0 }; 215 | 216 | // Find a suitable gpu heap to bind. 217 | std::map potential_heaps; 218 | for (auto i : range(num_table_entries)) { 219 | if (!active_sampler_sets[i]) { 220 | continue; 221 | } 222 | 223 | auto const data_slot { descriptor_table_start_slot + i }; 224 | auto const offset { user_data.data[data_slot] }; 225 | auto const handle { D3D12_CPU_DESCRIPTOR_HANDLE { start_cpu_sampler.ptr + offset } }; 226 | 227 | auto const sampler_set { this->_device->descriptors_gpu_sampler.get_set(offset) }; 228 | assert(sampler_set->set_sampler); 229 | num_sampler_sets += 1; 230 | 231 | auto const& placed_heaps { sampler_set->set_sampler->heaps_placed }; 232 | for (auto const& heap : placed_heaps) { 233 | auto ret { potential_heaps.insert(std::make_pair(heap, 1)) }; 234 | if (!ret.second) { 235 | ret.first->second += 1; 236 | } 237 | } 238 | } 239 | 240 | // Select heap from potential heaps 241 | auto heap { 0 }; 242 | auto heap_it { std::max_element(potential_heaps.begin(), potential_heaps.end()) }; 243 | if (heap_it == potential_heaps.end()) { 244 | // Haven't found any heap to use, so create a new one. 245 | heap = this->_device->descriptors_gpu_sampler.add_gpu_heap(); 246 | } else if (num_sampler_sets == heap_it->second) { 247 | heap = heap_it->first; 248 | } else { 249 | assert(!"unimplemented"); 250 | } 251 | auto& descriptor_heap { this->_device->descriptors_gpu_sampler.sampler_heap(heap) }; 252 | if (this->heap_gpu_sampler != descriptor_heap.heap.heap()) { 253 | this->heap_gpu_sampler = descriptor_heap.heap.heap(); 254 | start_sampler = this->heap_gpu_sampler->GetGPUDescriptorHandleForHeapStart(); 255 | bind_descriptor_heaps(); 256 | } 257 | 258 | // Upload descriptor sets into selected heap and set descriptor tables 259 | for (auto i : range(num_table_entries)) { 260 | if (!active_sampler_sets[i]) { 261 | continue; 262 | } 263 | 264 | auto const data_slot { descriptor_table_start_slot + i }; 265 | auto const offset { user_data.data[data_slot] }; 266 | auto const cpu_handle { D3D12_CPU_DESCRIPTOR_HANDLE { start_cpu_sampler.ptr + offset } }; 267 | 268 | auto sampler_set { this->_device->descriptors_gpu_sampler.get_set(offset) }; 269 | assert(sampler_set->set_sampler); 270 | 271 | auto& placed_heaps { sampler_set->set_sampler->heaps_placed }; 272 | D3D12_GPU_DESCRIPTOR_HANDLE gpu_handle { 0 }; 273 | 274 | auto placed_heap { placed_heaps.begin() }; 275 | if (placed_heap == placed_heaps.end()) { 276 | // Not placed inside the heap so far 277 | sampler_set->set_sampler->heaps_placed.emplace(heap); 278 | 279 | // Allocate a new slice in the gpu heap 280 | auto const num_descriptors { sampler_set->set_sampler->num_descriptors }; 281 | auto const sampler_handle { descriptor_heap.heap.alloc(num_descriptors) }; 282 | descriptor_heap.placed_sets.emplace(sampler_set, placed_descriptor_set_t { 1, sampler_handle, num_descriptors }); 283 | auto const [cpu_sampler, gpu_sampler] { sampler_handle }; 284 | gpu_handle = gpu_sampler; 285 | 286 | // Upload cpu samplers into gpu heap 287 | device()->CopyDescriptors( 288 | 1, 289 | &cpu_sampler, 290 | &num_descriptors, 291 | 1, 292 | &cpu_handle, 293 | &num_descriptors, 294 | D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER 295 | ); 296 | } else { 297 | auto& placed_set { descriptor_heap.placed_sets[sampler_set] }; 298 | placed_set.ref_count += 1; // TODO: sync 299 | gpu_handle = std::get<1>(placed_set.start); 300 | } 301 | 302 | user_data.data[data_slot] = gpu_handle.ptr - start_sampler.ptr; 303 | } 304 | } 305 | 306 | for (auto i : range(num_table_entries)) { 307 | const auto data_slot { descriptor_table_start_slot + i }; 308 | if (!user_data.dirty[data_slot]) { 309 | continue; 310 | } 311 | 312 | const auto offset { user_data.data[data_slot] }; 313 | SIZE_T descriptor_start { 0u }; 314 | switch (user_data.type[data_slot]) { 315 | case pipeline_slot_t::data_entry_type::CBV_SRV_UAV: descriptor_start = start_cbv_srv_uav.ptr; break; 316 | case pipeline_slot_t::data_entry_type::SAMPLER: descriptor_start = start_sampler.ptr; break; 317 | default: WARN("Unexpected user data entry: {}", static_cast(user_data.type[i])); 318 | } 319 | 320 | const auto handle { D3D12_GPU_DESCRIPTOR_HANDLE { descriptor_start + offset } }; 321 | set_table( 322 | static_cast(num_root_constant_entries + i), 323 | handle 324 | ); 325 | user_data.dirty.reset(data_slot); 326 | } 327 | } 328 | 329 | auto bind_graphics_slot(draw_type draw_type) -> void { 330 | if (!this->active_slot) { 331 | bind_descriptor_heaps(); 332 | } 333 | 334 | if ( 335 | (this->dynamic_state_dirty && this->graphics_slot.pipeline->dynamic_states) || 336 | this->active_slot != SLOT_GRAPHICS // Check if we are switching to Graphics 337 | ) { 338 | this->active_pipeline = nullptr; 339 | } 340 | 341 | if (!this->active_pipeline) { 342 | this->command_recorder.cmd_set_pipeline_state( 343 | std::visit( 344 | stdx::match( 345 | [] (pipeline_t::unique_pso_t& pso) { 346 | return pso.pipeline.Get(); 347 | }, 348 | [&] (pipeline_t::dynamic_pso_t& pso) { 349 | // Check if we have one available already 350 | { 351 | auto dynamic_state { this->dynamic_state }; 352 | this->graphics_slot.pipeline->pso_access.lock_shared(); 353 | switch (draw_type) { 354 | case draw_type::DRAW: { 355 | // Check all three strip cut variants 356 | dynamic_state.strip_cut = D3D12_INDEX_BUFFER_STRIP_CUT_VALUE_DISABLED; 357 | auto pipeline { pso.pipelines.find(dynamic_state) }; 358 | if (pipeline != pso.pipelines.end()) { 359 | return pipeline->second.Get(); 360 | } 361 | dynamic_state.strip_cut = D3D12_INDEX_BUFFER_STRIP_CUT_VALUE_0xFFFF; 362 | pipeline = pso.pipelines.find(dynamic_state); 363 | if (pipeline != pso.pipelines.end()) { 364 | return pipeline->second.Get(); 365 | } 366 | dynamic_state.strip_cut = D3D12_INDEX_BUFFER_STRIP_CUT_VALUE_0xFFFFFFFF; 367 | pipeline = pso.pipelines.find(dynamic_state); 368 | if (pipeline != pso.pipelines.end()) { 369 | return pipeline->second.Get(); 370 | } 371 | } break; 372 | case draw_type::DRAW_INDEXED: { 373 | if (graphics_slot.pipeline->dynamic_states & DYNAMIC_STATE_PRIMITIVE_RESTART) { 374 | switch (index_type) { 375 | case VK_INDEX_TYPE_UINT16: 376 | dynamic_state.strip_cut = D3D12_INDEX_BUFFER_STRIP_CUT_VALUE_0xFFFF; 377 | break; 378 | case VK_INDEX_TYPE_UINT32: 379 | dynamic_state.strip_cut = D3D12_INDEX_BUFFER_STRIP_CUT_VALUE_0xFFFFFFFF; 380 | break; 381 | } 382 | } else { 383 | dynamic_state.strip_cut = D3D12_INDEX_BUFFER_STRIP_CUT_VALUE_DISABLED; 384 | } 385 | 386 | auto pipeline { pso.pipelines.find(dynamic_state) }; 387 | if (pipeline != pso.pipelines.end()) { 388 | return pipeline->second.Get(); 389 | } 390 | } break; 391 | } 392 | this->graphics_slot.pipeline->pso_access.unlock_shared(); 393 | } 394 | 395 | // Generate a new one 396 | auto shader_bc = [] (ComPtr shader) { 397 | return shader ? 398 | D3D12_SHADER_BYTECODE { shader->GetBufferPointer(), shader->GetBufferSize() } : 399 | D3D12_SHADER_BYTECODE { 0, 0 }; 400 | }; 401 | 402 | // Apply currently set dynamic state values 403 | const auto dynamics { this->graphics_slot.pipeline->dynamic_states }; 404 | auto const& dynamic_state { this->dynamic_state }; 405 | 406 | auto rasterizer_state { pso.rasterizer_state }; 407 | if (dynamics & DYNAMIC_STATE_DEPTH_BIAS) { 408 | rasterizer_state.DepthBias = dynamic_state.depth_bias; 409 | rasterizer_state.DepthBiasClamp = dynamic_state.depth_bias_clamp; 410 | rasterizer_state.SlopeScaledDepthBias = dynamic_state.depth_bias_slope; 411 | } 412 | 413 | auto depth_stencil_state { pso.depth_stencil_state }; 414 | if (dynamics & DYNAMIC_STATE_STENCIL_COMPARE_MASK) { 415 | depth_stencil_state.StencilReadMask = dynamic_state.stencil_read_mask; 416 | } 417 | if (dynamics & DYNAMIC_STATE_STENCIL_WRITE_MASK) { 418 | depth_stencil_state.StencilWriteMask = dynamic_state.stencil_write_mask; 419 | } 420 | 421 | D3D12_GRAPHICS_PIPELINE_STATE_DESC desc { 422 | this->graphics_slot.pipeline->signature, 423 | shader_bc(pso.vertex_shader), 424 | shader_bc(pso.pixel_shader), 425 | shader_bc(pso.domain_shader), 426 | shader_bc(pso.hull_shader), 427 | shader_bc(pso.geometry_shader), 428 | D3D12_STREAM_OUTPUT_DESC { }, // not used 429 | pso.blend_state, 430 | pso.sample_mask, 431 | rasterizer_state, 432 | depth_stencil_state, 433 | D3D12_INPUT_LAYOUT_DESC { 434 | pso.input_elements.data(), 435 | static_cast(pso.input_elements.size()) 436 | }, 437 | dynamic_state.strip_cut, 438 | pso.topology_type, 439 | pso.num_render_targets, 440 | pso.rtv_formats[0], pso.rtv_formats[1], pso.rtv_formats[2], pso.rtv_formats[3], 441 | pso.rtv_formats[4], pso.rtv_formats[5], pso.rtv_formats[6], pso.rtv_formats[7], 442 | pso.dsv_format, 443 | pso.sample_desc, 444 | 0, // NodeMask 445 | D3D12_CACHED_PIPELINE_STATE { }, // TODO 446 | D3D12_PIPELINE_STATE_FLAG_NONE, // TODO 447 | }; 448 | 449 | ComPtr pipeline; 450 | auto const hr { this->device()->CreateGraphicsPipelineState(&desc, IID_PPV_ARGS(&pipeline)) }; 451 | // TODO: error handling 452 | 453 | this->graphics_slot.pipeline->pso_access.lock(); 454 | pso.pipelines.emplace(dynamic_state, pipeline); 455 | this->graphics_slot.pipeline->pso_access.unlock(); 456 | 457 | return pipeline.Get(); 458 | } 459 | ), 460 | this->graphics_slot.pipeline->pso 461 | ) 462 | ); 463 | } 464 | 465 | this->active_slot = SLOT_GRAPHICS; 466 | 467 | if (this->viewports_dirty) { 468 | this->command_recorder.cmd_set_viewports( 469 | this->num_viewports_scissors, 470 | this->viewports 471 | ); 472 | this->viewports_dirty = false; 473 | } 474 | 475 | if (this->scissors_dirty) { 476 | this->command_recorder.cmd_set_scissors( 477 | this->num_viewports_scissors, 478 | this->scissors 479 | ); 480 | this->scissors_dirty = false; 481 | } 482 | 483 | if (this->vertex_buffer_views_dirty.any()) { 484 | std::optional in_range { std::nullopt }; 485 | for (auto i : range(MAX_VERTEX_BUFFER_SLOTS)) { 486 | const auto has_flag { this->vertex_buffer_views_dirty[i] }; 487 | 488 | if (has_flag) { 489 | this->vertex_buffer_views_dirty.reset(i); 490 | if (!in_range) { 491 | in_range = i; 492 | } 493 | } else if (in_range) { 494 | this->command_recorder.cmd_set_vertex_buffers( 495 | static_cast(*in_range), 496 | static_cast(i - *in_range), 497 | this->vertex_buffer_views 498 | ); 499 | in_range = std::nullopt; 500 | } 501 | } 502 | 503 | if (in_range) { 504 | this->command_recorder.cmd_set_vertex_buffers( 505 | static_cast(*in_range), 506 | static_cast(MAX_VERTEX_BUFFER_SLOTS - *in_range), 507 | this->vertex_buffer_views 508 | ); 509 | } 510 | } 511 | 512 | update_user_data( 513 | this->graphics_slot, 514 | [&] (UINT slot, uint32_t data, UINT offset) { 515 | this->command_recorder.cmd_set_graphics_root_constant( 516 | slot, 517 | data, 518 | offset 519 | ); 520 | }, 521 | [&] (UINT slot, D3D12_GPU_DESCRIPTOR_HANDLE handle) { 522 | this->command_recorder.cmd_set_graphics_root_descriptor_table( 523 | slot, 524 | handle 525 | ); 526 | } 527 | ); 528 | } 529 | 530 | auto bind_compute_slot() -> void { 531 | if (!this->active_slot) { 532 | bind_descriptor_heaps(); 533 | } 534 | 535 | if (this->active_slot != SLOT_COMPUTE) { 536 | this->active_pipeline = nullptr; 537 | this->active_slot = SLOT_COMPUTE; 538 | } 539 | 540 | if (!this->active_pipeline) { 541 | this->command_recorder.cmd_set_pipeline_state( 542 | std::get(this->compute_slot.pipeline->pso).pipeline.Get() 543 | ); 544 | } 545 | 546 | update_user_data( 547 | this->compute_slot, 548 | [&] (UINT slot, uint32_t data, UINT offset) { 549 | this->command_recorder.cmd_set_compute_root_constant( 550 | slot, 551 | data, 552 | offset 553 | ); 554 | }, 555 | [&] (UINT slot, D3D12_GPU_DESCRIPTOR_HANDLE handle) { 556 | this->command_recorder.cmd_set_compute_root_descriptor_table( 557 | slot, 558 | handle 559 | ); 560 | } 561 | ); 562 | } 563 | 564 | auto begin_subpass(VkSubpassContents contents) -> void { 565 | // TODO: contents 566 | 567 | auto const& pass_cache { *this->pass_cache }; 568 | auto clear_values { span(pass_cache.clear_values) }; 569 | auto framebuffer { pass_cache.framebuffer }; 570 | auto render_pass { pass_cache.render_pass }; 571 | 572 | const auto subpass_id { pass_cache.subpass }; 573 | auto const& subpass { render_pass->subpasses[subpass_id] }; 574 | 575 | // Clear attachments on first use 576 | for (auto const& color_attachment : subpass.color_attachments) { 577 | if (color_attachment.attachment == VK_ATTACHMENT_UNUSED) { 578 | continue; 579 | } 580 | 581 | auto const& attachment { render_pass->attachments[color_attachment.attachment] }; 582 | if (attachment.first_use == subpass_id) { 583 | auto view { framebuffer->attachments[color_attachment.attachment] }; 584 | if (attachment.desc.loadOp == VK_ATTACHMENT_LOAD_OP_CLEAR) { 585 | // TODO: temp barriers... 586 | this->command_recorder.cmd_resource_barrier(1, 587 | &CD3DX12_RESOURCE_BARRIER::Transition(view->image, D3D12_RESOURCE_STATE_COMMON, D3D12_RESOURCE_STATE_RENDER_TARGET) 588 | ); 589 | this->command_recorder.cmd_clear_render_target_view( 590 | std::get<0>(*view->rtv), 591 | clear_values[color_attachment.attachment].color.float32, 592 | 1, 593 | &pass_cache.render_area 594 | ); 595 | } 596 | } 597 | } 598 | 599 | const auto depth_attachment { subpass.depth_attachment.attachment }; 600 | if (depth_attachment != VK_ATTACHMENT_UNUSED) { 601 | auto const& attachment { 602 | render_pass->attachments[depth_attachment] 603 | }; 604 | 605 | if (attachment.first_use == subpass_id) { 606 | auto view { framebuffer->attachments[depth_attachment] }; 607 | 608 | D3D12_CLEAR_FLAGS clear_flags { static_cast(0) }; 609 | float depth { 0 }; 610 | UINT8 stencil { 0 }; 611 | 612 | if (attachment.desc.loadOp == VK_ATTACHMENT_LOAD_OP_CLEAR) { 613 | clear_flags |= D3D12_CLEAR_FLAG_DEPTH; 614 | depth = clear_values[depth_attachment].depthStencil.depth; 615 | } 616 | 617 | // TODO: stencil 618 | 619 | if (clear_flags) { 620 | // TODO: temp barriers... 621 | this->command_recorder.cmd_resource_barrier(1, 622 | &CD3DX12_RESOURCE_BARRIER::Transition(view->image, D3D12_RESOURCE_STATE_COMMON, D3D12_RESOURCE_STATE_DEPTH_WRITE) 623 | ); 624 | this->command_recorder.cmd_clear_depth_stencil_view( 625 | std::get<0>(*view->dsv), 626 | clear_flags, 627 | depth, 628 | stencil, 629 | 1, 630 | &pass_cache.render_area 631 | ); 632 | } 633 | } 634 | } 635 | 636 | // Bind render targets 637 | auto color_attachments { span(subpass.color_attachments) }; 638 | const auto num_rtvs { color_attachments.size() }; 639 | D3D12_CPU_DESCRIPTOR_HANDLE render_targets[D3D12_SIMULTANEOUS_RENDER_TARGET_COUNT]; 640 | for (auto i : range(num_rtvs)) { 641 | // TODO: support VK_ATTACHMENT_UNUSED 642 | const auto color_attachment { color_attachments[i].attachment }; 643 | assert(color_attachment != VK_ATTACHMENT_UNUSED); 644 | render_targets[i] = std::get<0>(*framebuffer->attachments[color_attachment]->rtv); 645 | } 646 | 647 | D3D12_CPU_DESCRIPTOR_HANDLE* dsv { nullptr }; 648 | if (depth_attachment != VK_ATTACHMENT_UNUSED) { 649 | dsv = &std::get<0>(*framebuffer->attachments[depth_attachment]->dsv); 650 | } 651 | 652 | this->command_recorder.cmd_set_render_targets( 653 | static_cast(num_rtvs), 654 | render_targets, 655 | !!dsv, 656 | dsv 657 | ); 658 | } 659 | 660 | auto end_subpass() -> void { 661 | auto const& pass_cache { *this->pass_cache }; 662 | auto framebuffer { pass_cache.framebuffer }; 663 | auto render_pass { pass_cache.render_pass }; 664 | 665 | const auto subpass_id { pass_cache.subpass }; 666 | auto const& subpass { render_pass->subpasses[subpass_id] }; 667 | 668 | // Clear attachments on first use 669 | // TODO 670 | for (auto const& color_attachment : subpass.color_attachments) { 671 | if (color_attachment.attachment == VK_ATTACHMENT_UNUSED) { 672 | continue; 673 | } 674 | 675 | auto const& attachment { render_pass->attachments[color_attachment.attachment] }; 676 | if (attachment.first_use == subpass_id) { 677 | auto view { framebuffer->attachments[color_attachment.attachment] }; 678 | if (attachment.desc.loadOp == VK_ATTACHMENT_LOAD_OP_CLEAR) { 679 | // TODO: temp barriers... 680 | this->command_recorder.cmd_resource_barrier(1, 681 | &CD3DX12_RESOURCE_BARRIER::Transition(view->image, D3D12_RESOURCE_STATE_RENDER_TARGET, D3D12_RESOURCE_STATE_COMMON) 682 | ); 683 | } 684 | } 685 | } 686 | 687 | const auto depth_attachment { subpass.depth_attachment.attachment }; 688 | if (depth_attachment != VK_ATTACHMENT_UNUSED) { 689 | auto const& attachment { 690 | render_pass->attachments[depth_attachment] 691 | }; 692 | 693 | if (attachment.first_use == subpass_id) { 694 | auto view { framebuffer->attachments[depth_attachment] }; 695 | 696 | if (attachment.desc.loadOp == VK_ATTACHMENT_LOAD_OP_CLEAR) { 697 | // TODO: temp barriers... 698 | this->command_recorder.cmd_resource_barrier(1, 699 | &CD3DX12_RESOURCE_BARRIER::Transition(view->image, D3D12_RESOURCE_STATE_DEPTH_WRITE, D3D12_RESOURCE_STATE_COMMON) 700 | ); 701 | } 702 | } 703 | } 704 | 705 | for (auto i : range(subpass.resolve_attachments.size())) { 706 | auto const& color_attachment { subpass.color_attachments[i] }; 707 | auto const& resolve_attachment { subpass.resolve_attachments[i] }; 708 | 709 | if (resolve_attachment.attachment == VK_ATTACHMENT_UNUSED) { 710 | continue; 711 | } 712 | 713 | auto color_view { framebuffer->attachments[color_attachment.attachment] }; 714 | auto resolve_view { framebuffer->attachments[resolve_attachment.attachment] }; 715 | 716 | this->command_recorder.resolve_subresource( 717 | resolve_view->image, 718 | 0, // TODO: D3D12CalcSubresource(resolve_view->) 719 | color_view->image, 720 | 0, // TODO 721 | formats[render_pass->attachments[color_attachment.attachment].desc.format] 722 | ); 723 | } 724 | } 725 | 726 | auto raw_command_list() const -> ID3D12CommandList* { 727 | return this->command_list.Get(); 728 | } 729 | 730 | private: 731 | /// Dispatchable object 732 | uintptr_t loader_magic; 733 | 734 | ComPtr command_list; 735 | 736 | public: 737 | command_buffer_recorder_native_t command_recorder; 738 | 739 | std::optional active_slot; 740 | ID3D12PipelineState* active_pipeline; 741 | 742 | ID3D12DescriptorHeap* heap_gpu_cbv_srv_uav; // gpu 743 | ID3D12DescriptorHeap* heap_cpu_sampler; // cpu 744 | ID3D12DescriptorHeap* heap_gpu_sampler; // current gpu heap 745 | 746 | bool dynamic_state_dirty; 747 | bool viewports_dirty; 748 | bool scissors_dirty; 749 | std::bitset vertex_buffer_views_dirty; 750 | 751 | // Currently set dynamic state 752 | dynamic_state_t dynamic_state; 753 | VkIndexType index_type; 754 | 755 | device_t* _device; 756 | 757 | private: 758 | // Owning command allocator, required for reset 759 | ID3D12CommandAllocator* allocator; 760 | 761 | public: 762 | std::optional pass_cache; 763 | 764 | pipeline_slot_t graphics_slot; 765 | pipeline_slot_t compute_slot; 766 | 767 | D3D12_VIEWPORT viewports[D3D12_VIEWPORT_AND_SCISSORRECT_OBJECT_COUNT_PER_PIPELINE]; 768 | D3D12_RECT scissors[D3D12_VIEWPORT_AND_SCISSORRECT_OBJECT_COUNT_PER_PIPELINE]; 769 | UINT num_viewports_scissors; 770 | 771 | // Stride values are not known when calling `vkCmdBindVertexBuffers` 772 | D3D12_VERTEX_BUFFER_VIEW vertex_buffer_views[MAX_VERTEX_BUFFER_SLOTS]; 773 | 774 | std::vector> temp_heaps; 775 | }; 776 | -------------------------------------------------------------------------------- /src/command_recorder.hpp: -------------------------------------------------------------------------------- 1 | #pragma once 2 | 3 | #include "impl.hpp" 4 | 5 | class command_buffer_recorder_t { 6 | public: 7 | virtual ~command_buffer_recorder_t() { } 8 | 9 | virtual auto resolve_subresource( 10 | ID3D12Resource *pDstResource, 11 | UINT DstSubresource, 12 | ID3D12Resource *pSrcResource, 13 | UINT SrcSubresource, 14 | DXGI_FORMAT Format 15 | ) -> void = 0; 16 | 17 | virtual auto cmd_set_descriptor_heaps( 18 | UINT NumDescriptorHeaps, 19 | ID3D12DescriptorHeap *const *ppDescriptorHeaps 20 | ) -> void = 0; 21 | 22 | virtual auto cmd_set_compute_root_signature(ID3D12RootSignature *pRootSignature) -> void = 0; 23 | virtual auto cmd_set_graphics_root_signature(ID3D12RootSignature *pRootSignature) -> void = 0; 24 | 25 | virtual auto cmd_set_compute_root_descriptor_table( 26 | UINT RootParameterIndex, 27 | D3D12_GPU_DESCRIPTOR_HANDLE BaseDescriptor 28 | ) -> void = 0; 29 | 30 | virtual auto cmd_set_compute_root_constant( 31 | UINT RootParameterIndex, 32 | UINT SrcData, 33 | UINT DestOffsetIn32BitValues 34 | ) -> void = 0; 35 | 36 | virtual auto cmd_set_compute_root_constants( 37 | UINT RootParameterIndex, 38 | UINT Num32BitValuesToSet, 39 | const void *pSrcData, 40 | UINT DestOffsetIn32BitValues 41 | ) -> void = 0; 42 | 43 | virtual auto cmd_set_compute_root_shader_resource_view( 44 | UINT RootParameterIndex, 45 | D3D12_GPU_VIRTUAL_ADDRESS BufferLocation 46 | ) -> void = 0; 47 | 48 | virtual auto cmd_set_graphics_root_constant( 49 | UINT RootParameterIndex, 50 | UINT SrcData, 51 | UINT DestOffsetIn32BitValues 52 | ) -> void = 0; 53 | 54 | virtual auto cmd_set_graphics_root_descriptor_table( 55 | UINT RootParameterIndex, 56 | D3D12_GPU_DESCRIPTOR_HANDLE BaseDescriptor 57 | ) -> void = 0; 58 | virtual auto cmd_clear_render_target_view( 59 | D3D12_CPU_DESCRIPTOR_HANDLE RenderTargetView, 60 | const FLOAT ColorRGBA[4], 61 | UINT NumRects, 62 | const D3D12_RECT *pRects 63 | ) -> void = 0; 64 | 65 | virtual auto cmd_clear_depth_stencil_view( 66 | D3D12_CPU_DESCRIPTOR_HANDLE DepthStencilView, 67 | D3D12_CLEAR_FLAGS ClearFlags, 68 | FLOAT Depth, 69 | UINT8 Stencil, 70 | UINT NumRects, 71 | const D3D12_RECT *pRects 72 | ) -> void = 0; 73 | 74 | virtual auto cmd_set_render_targets( 75 | UINT NumRenderTargetDescriptors, 76 | const D3D12_CPU_DESCRIPTOR_HANDLE *pRenderTargetDescriptors, 77 | BOOL RTsSingleHandleToDescriptorRange, 78 | const D3D12_CPU_DESCRIPTOR_HANDLE *pDepthStencilDescriptor 79 | ) -> void = 0; 80 | 81 | virtual auto cmd_set_primitive_topolgy(D3D12_PRIMITIVE_TOPOLOGY PrimitiveTopology) -> void = 0; 82 | virtual auto cmd_set_scissors(UINT NumRects, const D3D12_RECT *pRects) -> void = 0; 83 | virtual auto cmd_set_viewports(UINT NumViewports, const D3D12_VIEWPORT *pViewports) -> void = 0; 84 | virtual auto cmd_set_blend_factor(const FLOAT BlendFactor[4]) -> void = 0; 85 | virtual auto cmd_set_stencil_ref(UINT StencilRef) -> void = 0; 86 | virtual auto cmd_set_depth_bounds(FLOAT Min, FLOAT Max) -> void = 0; 87 | virtual auto cmd_set_index_buffer(const D3D12_INDEX_BUFFER_VIEW *pView) -> void = 0; 88 | 89 | virtual auto cmd_copy_buffer_region( 90 | ID3D12Resource *pDstBuffer, 91 | UINT64 DstOffset, 92 | ID3D12Resource *pSrcBuffer, 93 | UINT64 SrcOffset, 94 | UINT64 NumBytes 95 | ) -> void = 0; 96 | 97 | virtual auto cmd_copy_texture_region( 98 | const D3D12_TEXTURE_COPY_LOCATION *pDst, 99 | UINT DstX, 100 | UINT DstY, 101 | UINT DstZ, 102 | const D3D12_TEXTURE_COPY_LOCATION *pSrc, 103 | const D3D12_BOX *pSrcBox 104 | ) -> void = 0; 105 | 106 | virtual auto cmd_set_pipeline_state(ID3D12PipelineState *pPipelineState) -> void = 0; 107 | 108 | virtual auto cmd_dispatch( 109 | UINT ThreadGroupCountX, 110 | UINT ThreadGroupCountY, 111 | UINT ThreadGroupCountZ 112 | ) -> void = 0; 113 | 114 | virtual auto cmd_draw_instanced( 115 | UINT VertexCountPerInstance, 116 | UINT InstanceCount, 117 | UINT StartVertexLocation, 118 | UINT StartInstanceLocation 119 | ) -> void = 0; 120 | 121 | virtual auto cmd_draw_indexed_instanced( 122 | UINT IndexCountPerInstance, 123 | UINT InstanceCount, 124 | UINT StartIndexLocation, 125 | INT BaseVertexLocation, 126 | UINT StartInstanceLocation 127 | ) -> void = 0; 128 | 129 | virtual auto cmd_execute_indirect( 130 | ID3D12CommandSignature *pCommandSignature, 131 | UINT MaxCommandCount, 132 | ID3D12Resource *pArgumentBuffer, 133 | UINT64 ArgumentBufferOffset, 134 | ID3D12Resource *pCountBuffer, 135 | UINT64 CountBufferOffset 136 | ) -> void = 0; 137 | 138 | virtual auto cmd_resource_barrier( 139 | UINT NumBarriers, 140 | const D3D12_RESOURCE_BARRIER *pBarriers 141 | ) -> void = 0; 142 | 143 | virtual auto cmd_set_vertex_buffers( 144 | UINT StartSlot, 145 | UINT NumViews, 146 | const D3D12_VERTEX_BUFFER_VIEW *pViews 147 | ) -> void = 0; 148 | }; 149 | 150 | class command_buffer_recorder_native_t : public command_buffer_recorder_t { 151 | public: 152 | explicit command_buffer_recorder_native_t(ID3D12GraphicsCommandList2* cmd_list) : 153 | command_list { cmd_list } 154 | {} 155 | 156 | public: 157 | virtual auto resolve_subresource( 158 | ID3D12Resource *pDstResource, 159 | UINT DstSubresource, 160 | ID3D12Resource *pSrcResource, 161 | UINT SrcSubresource, 162 | DXGI_FORMAT Format 163 | ) -> void override { 164 | this->command_list->ResolveSubresource( 165 | pDstResource, 166 | DstSubresource, 167 | pSrcResource, 168 | SrcSubresource, 169 | Format 170 | ); 171 | } 172 | 173 | virtual auto cmd_set_descriptor_heaps( 174 | UINT NumDescriptorHeaps, 175 | ID3D12DescriptorHeap *const *ppDescriptorHeaps 176 | ) -> void override { 177 | this->command_list->SetDescriptorHeaps( 178 | NumDescriptorHeaps, 179 | ppDescriptorHeaps 180 | ); 181 | } 182 | 183 | virtual auto cmd_set_compute_root_signature(ID3D12RootSignature *pRootSignature) -> void override { 184 | this->command_list->SetComputeRootSignature(pRootSignature); 185 | } 186 | 187 | virtual auto cmd_set_graphics_root_signature(ID3D12RootSignature *pRootSignature) -> void override { 188 | this->command_list->SetGraphicsRootSignature(pRootSignature); 189 | } 190 | 191 | virtual auto cmd_set_compute_root_descriptor_table( 192 | UINT RootParameterIndex, 193 | D3D12_GPU_DESCRIPTOR_HANDLE BaseDescriptor 194 | ) -> void override { 195 | this->command_list->SetComputeRootDescriptorTable( 196 | RootParameterIndex, 197 | BaseDescriptor 198 | ); 199 | } 200 | 201 | virtual auto cmd_set_compute_root_constant( 202 | UINT RootParameterIndex, 203 | UINT SrcData, 204 | UINT DestOffsetIn32BitValues 205 | ) -> void override { 206 | this->command_list->SetComputeRoot32BitConstant( 207 | RootParameterIndex, 208 | SrcData, 209 | DestOffsetIn32BitValues 210 | ); 211 | } 212 | 213 | virtual auto cmd_set_compute_root_constants( 214 | UINT RootParameterIndex, 215 | UINT Num32BitValuesToSet, 216 | const void *pSrcData, 217 | UINT DestOffsetIn32BitValues 218 | ) -> void override { 219 | this->command_list->SetComputeRoot32BitConstants( 220 | RootParameterIndex, 221 | Num32BitValuesToSet, 222 | pSrcData, 223 | DestOffsetIn32BitValues 224 | ); 225 | } 226 | 227 | virtual auto cmd_set_compute_root_shader_resource_view( 228 | UINT RootParameterIndex, 229 | D3D12_GPU_VIRTUAL_ADDRESS BufferLocation 230 | ) -> void override { 231 | this->command_list->SetComputeRootShaderResourceView( 232 | RootParameterIndex, 233 | BufferLocation 234 | ); 235 | } 236 | 237 | virtual auto cmd_set_graphics_root_constant( 238 | UINT RootParameterIndex, 239 | UINT SrcData, 240 | UINT DestOffsetIn32BitValues 241 | ) -> void override { 242 | this->command_list->SetGraphicsRoot32BitConstant( 243 | RootParameterIndex, 244 | SrcData, 245 | DestOffsetIn32BitValues 246 | ); 247 | } 248 | 249 | virtual auto cmd_set_graphics_root_descriptor_table( 250 | UINT RootParameterIndex, 251 | D3D12_GPU_DESCRIPTOR_HANDLE BaseDescriptor 252 | ) -> void override { 253 | this->command_list->SetGraphicsRootDescriptorTable( 254 | RootParameterIndex, 255 | BaseDescriptor 256 | ); 257 | } 258 | 259 | virtual auto cmd_clear_render_target_view( 260 | D3D12_CPU_DESCRIPTOR_HANDLE RenderTargetView, 261 | const FLOAT ColorRGBA[4], 262 | UINT NumRects, 263 | const D3D12_RECT *pRects 264 | ) -> void override { 265 | this->command_list->ClearRenderTargetView( 266 | RenderTargetView, 267 | ColorRGBA, 268 | NumRects, 269 | pRects 270 | ); 271 | } 272 | 273 | virtual auto cmd_clear_depth_stencil_view( 274 | D3D12_CPU_DESCRIPTOR_HANDLE DepthStencilView, 275 | D3D12_CLEAR_FLAGS ClearFlags, 276 | FLOAT Depth, 277 | UINT8 Stencil, 278 | UINT NumRects, 279 | const D3D12_RECT *pRects 280 | ) -> void override { 281 | this->command_list->ClearDepthStencilView( 282 | DepthStencilView, 283 | ClearFlags, 284 | Depth, 285 | Stencil, 286 | NumRects, 287 | pRects 288 | ); 289 | } 290 | 291 | virtual auto cmd_set_render_targets( 292 | UINT NumRenderTargetDescriptors, 293 | const D3D12_CPU_DESCRIPTOR_HANDLE *pRenderTargetDescriptors, 294 | BOOL RTsSingleHandleToDescriptorRange, 295 | const D3D12_CPU_DESCRIPTOR_HANDLE *pDepthStencilDescriptor 296 | ) -> void override { 297 | this->command_list->OMSetRenderTargets( 298 | NumRenderTargetDescriptors, 299 | pRenderTargetDescriptors, 300 | RTsSingleHandleToDescriptorRange, 301 | pDepthStencilDescriptor 302 | ); 303 | } 304 | 305 | virtual auto cmd_set_primitive_topolgy(D3D12_PRIMITIVE_TOPOLOGY PrimitiveTopology) -> void override { 306 | this->command_list->IASetPrimitiveTopology(PrimitiveTopology); 307 | } 308 | 309 | virtual auto cmd_set_scissors(UINT NumRects, const D3D12_RECT *pRects) -> void override { 310 | this->command_list->RSSetScissorRects(NumRects, pRects); 311 | } 312 | 313 | virtual auto cmd_set_viewports(UINT NumViewports, const D3D12_VIEWPORT *pViewports) -> void override { 314 | this->command_list->RSSetViewports(NumViewports, pViewports); 315 | } 316 | 317 | virtual auto cmd_set_blend_factor(const FLOAT BlendFactor[4]) -> void override { 318 | this->command_list->OMSetBlendFactor(BlendFactor); 319 | } 320 | 321 | virtual auto cmd_set_stencil_ref(UINT StencilRef) -> void override { 322 | this->command_list->OMSetStencilRef(StencilRef); 323 | } 324 | 325 | virtual auto cmd_set_depth_bounds(FLOAT Min, FLOAT Max) -> void override { 326 | this->command_list->OMSetDepthBounds(Min, Max); 327 | } 328 | 329 | virtual auto cmd_set_index_buffer(const D3D12_INDEX_BUFFER_VIEW *pView) -> void override { 330 | this->command_list->IASetIndexBuffer(pView); 331 | } 332 | 333 | virtual auto cmd_copy_buffer_region( 334 | ID3D12Resource *pDstBuffer, 335 | UINT64 DstOffset, 336 | ID3D12Resource *pSrcBuffer, 337 | UINT64 SrcOffset, 338 | UINT64 NumBytes 339 | ) -> void override { 340 | this->command_list->CopyBufferRegion( 341 | pDstBuffer, 342 | DstOffset, 343 | pSrcBuffer, 344 | SrcOffset, 345 | NumBytes 346 | ); 347 | } 348 | 349 | virtual auto cmd_copy_texture_region( 350 | const D3D12_TEXTURE_COPY_LOCATION *pDst, 351 | UINT DstX, 352 | UINT DstY, 353 | UINT DstZ, 354 | const D3D12_TEXTURE_COPY_LOCATION *pSrc, 355 | const D3D12_BOX *pSrcBox 356 | ) -> void override { 357 | this->command_list->CopyTextureRegion( 358 | pDst, 359 | DstX, 360 | DstY, 361 | DstZ, 362 | pSrc, 363 | pSrcBox 364 | ); 365 | } 366 | 367 | virtual auto cmd_set_pipeline_state(ID3D12PipelineState *pPipelineState) -> void override { 368 | this->command_list->SetPipelineState(pPipelineState); 369 | } 370 | 371 | virtual auto cmd_dispatch( 372 | UINT ThreadGroupCountX, 373 | UINT ThreadGroupCountY, 374 | UINT ThreadGroupCountZ 375 | ) -> void override { 376 | this->command_list->Dispatch( 377 | ThreadGroupCountX, 378 | ThreadGroupCountY, 379 | ThreadGroupCountZ 380 | ); 381 | } 382 | 383 | virtual auto cmd_draw_instanced( 384 | UINT VertexCountPerInstance, 385 | UINT InstanceCount, 386 | UINT StartVertexLocation, 387 | UINT StartInstanceLocation 388 | ) -> void override { 389 | this->command_list->DrawInstanced( 390 | VertexCountPerInstance, 391 | InstanceCount, 392 | StartVertexLocation, 393 | StartInstanceLocation 394 | ); 395 | } 396 | 397 | virtual auto cmd_draw_indexed_instanced( 398 | UINT IndexCountPerInstance, 399 | UINT InstanceCount, 400 | UINT StartIndexLocation, 401 | INT BaseVertexLocation, 402 | UINT StartInstanceLocation 403 | ) -> void override { 404 | this->command_list->DrawIndexedInstanced( 405 | IndexCountPerInstance, 406 | InstanceCount, 407 | StartIndexLocation, 408 | BaseVertexLocation, 409 | StartInstanceLocation 410 | ); 411 | } 412 | 413 | virtual auto cmd_execute_indirect( 414 | ID3D12CommandSignature *pCommandSignature, 415 | UINT MaxCommandCount, 416 | ID3D12Resource *pArgumentBuffer, 417 | UINT64 ArgumentBufferOffset, 418 | ID3D12Resource *pCountBuffer, 419 | UINT64 CountBufferOffset 420 | ) -> void override { 421 | this->command_list->ExecuteIndirect( 422 | pCommandSignature, 423 | MaxCommandCount, 424 | pArgumentBuffer, 425 | ArgumentBufferOffset, 426 | pCountBuffer, 427 | CountBufferOffset 428 | ); 429 | } 430 | 431 | virtual auto cmd_resource_barrier( 432 | UINT NumBarriers, 433 | const D3D12_RESOURCE_BARRIER *pBarriers 434 | ) -> void override { 435 | this->command_list->ResourceBarrier(NumBarriers, pBarriers); 436 | } 437 | 438 | virtual auto cmd_set_vertex_buffers( 439 | UINT StartSlot, 440 | UINT NumViews, 441 | const D3D12_VERTEX_BUFFER_VIEW *pViews 442 | ) -> void override { 443 | this->command_list->IASetVertexBuffers( 444 | StartSlot, 445 | NumViews, 446 | pViews 447 | ); 448 | } 449 | 450 | private: 451 | ID3D12GraphicsCommandList2* command_list; 452 | }; 453 | 454 | class command_buffer_recorder_store_t : public command_buffer_recorder_t { 455 | enum class command_t { 456 | RESOLVE_SUBRESOURCE, 457 | SET_DESCRIPTOR_HEAPS, 458 | SET_COMPUTE_ROOT_SIGNATURE, 459 | SET_COMPUTE_ROOT_DESCRIPTOR_TABLE, 460 | SET_COMPUTE_ROOT_CONSTANT, 461 | SET_COMPUTE_ROOT_CONSTANTS, 462 | SET_COMPUTE_ROOT_SHADER_RESOURCE_VIEW, 463 | SET_GRAPHICS_ROOT_SIGNATURE, 464 | SET_GRAPHICS_ROOT_CONSTANT, 465 | SET_GRAPHICS_ROOT_DESCRIPTOR_TABLE, 466 | CLEAR_RENDER_TARGET_VIEW, 467 | CLEAR_DEPTH_STENCIL_VIEW, 468 | SET_RENDER_TARGET, 469 | SET_PRIMITIVE_TOPOLOGY, 470 | SET_SCISSORS, 471 | SET_VIEWPORTS, 472 | SET_BLEND_FACTOR, 473 | SET_STENCIL_REF, 474 | SET_DEPTH_BOUNDS, 475 | SET_INDEX_BUFFER, 476 | COPY_BUFFER_REGION, 477 | COPY_TEXTURE_REGION, 478 | SET_PIPELINE_STATE, 479 | DISPATCH, 480 | DRAW_INSTANCED, 481 | DRAW_INDEXED_INSTANCED, 482 | EXECUTE_INDIRECT, 483 | RESOURCE_BARRIER, 484 | SET_VERTEX_BUFFERS, 485 | }; 486 | 487 | struct resolve_subresource_t { 488 | ID3D12Resource *pDstResource; 489 | UINT DstSubresource; 490 | ID3D12Resource *pSrcResource; 491 | UINT SrcSubresource; 492 | DXGI_FORMAT Format; 493 | }; 494 | 495 | struct set_descriptor_heaps_t { 496 | UINT NumDescriptorHeaps; 497 | ID3D12DescriptorHeap const* DescriptorHeaps[2]; // Max 2 Heaps atm 498 | }; 499 | 500 | struct set_compute_root_signature_t { 501 | ID3D12RootSignature *pRootSignature; 502 | }; 503 | 504 | struct set_graphics_root_signature_t { 505 | ID3D12RootSignature *pRootSignature; 506 | }; 507 | 508 | struct set_compute_root_descriptor_table_t { 509 | UINT RootParameterIndex; 510 | D3D12_GPU_DESCRIPTOR_HANDLE BaseDescriptor; 511 | }; 512 | 513 | struct set_graphics_root_descriptor_table_t { 514 | UINT RootParameterIndex; 515 | D3D12_GPU_DESCRIPTOR_HANDLE BaseDescriptor; 516 | }; 517 | 518 | struct set_compute_root_constant_t { 519 | UINT RootParameterIndex; 520 | UINT SrcData; 521 | UINT DestOffsetIn32BitValues; 522 | }; 523 | 524 | struct set_graphics_root_constant_t { 525 | UINT RootParameterIndex; 526 | UINT SrcData; 527 | UINT DestOffsetIn32BitValues; 528 | }; 529 | 530 | struct set_compute_root_shader_resource_view_t { 531 | UINT RootParameterIndex; 532 | D3D12_GPU_VIRTUAL_ADDRESS BufferLocation; 533 | }; 534 | 535 | struct set_pipeline_state_t { 536 | ID3D12PipelineState *state; 537 | }; 538 | 539 | struct set_primitive_topology_t { 540 | D3D12_PRIMITIVE_TOPOLOGY PrimitiveTopology; 541 | }; 542 | 543 | struct set_stencil_ref_t { 544 | UINT StencilRef; 545 | }; 546 | 547 | struct set_depth_bounds_t { 548 | FLOAT Min; 549 | FLOAT Max; 550 | }; 551 | 552 | struct dispatch_t { 553 | UINT ThreadGroupCountX; 554 | UINT ThreadGroupCountY; 555 | UINT ThreadGroupCountZ; 556 | }; 557 | 558 | struct draw_instanced_t { 559 | UINT VertexCountPerInstance; 560 | UINT InstanceCount; 561 | UINT StartVertexLocation; 562 | UINT StartInstanceLocation; 563 | }; 564 | 565 | struct draw_indexed_instanced_t { 566 | UINT IndexCountPerInstance; 567 | UINT InstanceCount; 568 | UINT StartIndexLocation; 569 | INT BaseVertexLocation; 570 | UINT StartInstanceLocation; 571 | }; 572 | 573 | struct execute_indirect_t { 574 | ID3D12CommandSignature *pCommandSignature; 575 | UINT MaxCommandCount; 576 | ID3D12Resource *pArgumentBuffer; 577 | UINT64 ArgumentBufferOffset; 578 | ID3D12Resource *pCountBuffer; 579 | UINT64 CountBufferOffset; 580 | }; 581 | 582 | public: 583 | virtual auto resolve_subresource( 584 | ID3D12Resource *pDstResource, 585 | UINT DstSubresource, 586 | ID3D12Resource *pSrcResource, 587 | UINT SrcSubresource, 588 | DXGI_FORMAT Format 589 | ) -> void { 590 | encode( 591 | command_t::RESOLVE_SUBRESOURCE, 592 | resolve_subresource_t { 593 | pDstResource, 594 | DstSubresource, 595 | pSrcResource, 596 | SrcSubresource, 597 | Format, 598 | } 599 | ); 600 | } 601 | 602 | virtual auto cmd_set_descriptor_heaps( 603 | UINT NumDescriptorHeaps, 604 | ID3D12DescriptorHeap *const *ppDescriptorHeaps 605 | ) -> void { 606 | set_descriptor_heaps_t cmd_data { NumDescriptorHeaps }; 607 | for (auto i : range(NumDescriptorHeaps)) { 608 | cmd_data.DescriptorHeaps[i] = ppDescriptorHeaps[i]; 609 | } 610 | encode(command_t::SET_DESCRIPTOR_HEAPS, cmd_data); 611 | } 612 | 613 | virtual auto cmd_set_compute_root_signature(ID3D12RootSignature *pRootSignature) -> void { 614 | encode(command_t::SET_COMPUTE_ROOT_SIGNATURE, set_compute_root_signature_t { pRootSignature }); 615 | } 616 | 617 | virtual auto cmd_set_graphics_root_signature(ID3D12RootSignature *pRootSignature) -> void { 618 | encode(command_t::SET_GRAPHICS_ROOT_SIGNATURE, set_graphics_root_signature_t { pRootSignature }); 619 | } 620 | 621 | virtual auto cmd_set_compute_root_descriptor_table( 622 | UINT RootParameterIndex, 623 | D3D12_GPU_DESCRIPTOR_HANDLE BaseDescriptor 624 | ) -> void { 625 | encode( 626 | command_t::SET_COMPUTE_ROOT_DESCRIPTOR_TABLE, 627 | set_compute_root_descriptor_table_t { 628 | RootParameterIndex, 629 | BaseDescriptor, 630 | } 631 | ); 632 | } 633 | 634 | virtual auto cmd_set_compute_root_constant( 635 | UINT RootParameterIndex, 636 | UINT SrcData, 637 | UINT DestOffsetIn32BitValues 638 | ) -> void { 639 | encode( 640 | command_t::SET_COMPUTE_ROOT_CONSTANT, 641 | set_compute_root_constant_t { 642 | RootParameterIndex, 643 | SrcData, 644 | DestOffsetIn32BitValues, 645 | } 646 | ); 647 | } 648 | 649 | virtual auto cmd_set_compute_root_constants( 650 | UINT RootParameterIndex, 651 | UINT Num32BitValuesToSet, 652 | const void *pSrcData, 653 | UINT DestOffsetIn32BitValues 654 | ) -> void { 655 | this->commands.push_back(command_t::SET_COMPUTE_ROOT_CONSTANTS); 656 | } 657 | 658 | virtual auto cmd_set_compute_root_shader_resource_view( 659 | UINT RootParameterIndex, 660 | D3D12_GPU_VIRTUAL_ADDRESS BufferLocation 661 | ) -> void { 662 | encode( 663 | command_t::SET_COMPUTE_ROOT_SHADER_RESOURCE_VIEW, 664 | set_compute_root_shader_resource_view_t { 665 | RootParameterIndex, 666 | BufferLocation, 667 | } 668 | ); 669 | } 670 | 671 | virtual auto cmd_set_graphics_root_constant( 672 | UINT RootParameterIndex, 673 | UINT SrcData, 674 | UINT DestOffsetIn32BitValues 675 | ) -> void { 676 | encode( 677 | command_t::SET_GRAPHICS_ROOT_CONSTANT, 678 | set_graphics_root_constant_t { 679 | RootParameterIndex, 680 | SrcData, 681 | DestOffsetIn32BitValues, 682 | } 683 | ); 684 | } 685 | 686 | virtual auto cmd_set_graphics_root_descriptor_table( 687 | UINT RootParameterIndex, 688 | D3D12_GPU_DESCRIPTOR_HANDLE BaseDescriptor 689 | ) -> void { 690 | encode( 691 | command_t::SET_GRAPHICS_ROOT_DESCRIPTOR_TABLE, 692 | set_graphics_root_descriptor_table_t { 693 | RootParameterIndex, 694 | BaseDescriptor, 695 | } 696 | ); 697 | } 698 | 699 | virtual auto cmd_clear_render_target_view( 700 | D3D12_CPU_DESCRIPTOR_HANDLE RenderTargetView, 701 | const FLOAT ColorRGBA[4], 702 | UINT NumRects, 703 | const D3D12_RECT *pRects 704 | ) -> void { 705 | this->commands.push_back(command_t::CLEAR_RENDER_TARGET_VIEW); 706 | } 707 | 708 | virtual auto cmd_clear_depth_stencil_view( 709 | D3D12_CPU_DESCRIPTOR_HANDLE DepthStencilView, 710 | D3D12_CLEAR_FLAGS ClearFlags, 711 | FLOAT Depth, 712 | UINT8 Stencil, 713 | UINT NumRects, 714 | const D3D12_RECT *pRects 715 | ) -> void { 716 | this->commands.push_back(command_t::CLEAR_DEPTH_STENCIL_VIEW); 717 | } 718 | 719 | virtual auto cmd_set_render_targets( 720 | UINT NumRenderTargetDescriptors, 721 | const D3D12_CPU_DESCRIPTOR_HANDLE *pRenderTargetDescriptors, 722 | BOOL RTsSingleHandleToDescriptorRange, 723 | const D3D12_CPU_DESCRIPTOR_HANDLE *pDepthStencilDescriptor 724 | ) -> void { 725 | this->commands.push_back(command_t::SET_RENDER_TARGET); 726 | } 727 | 728 | virtual auto cmd_set_primitive_topolgy(D3D12_PRIMITIVE_TOPOLOGY PrimitiveTopology) -> void { 729 | encode( 730 | command_t::SET_PRIMITIVE_TOPOLOGY, 731 | set_primitive_topology_t { PrimitiveTopology } 732 | ); 733 | } 734 | 735 | virtual auto cmd_set_scissors(UINT NumRects, const D3D12_RECT *pRects) -> void { 736 | this->commands.push_back(command_t::SET_SCISSORS); 737 | } 738 | 739 | virtual auto cmd_set_viewports(UINT NumViewports, const D3D12_VIEWPORT *pViewports) -> void { 740 | this->commands.push_back(command_t::SET_VIEWPORTS); 741 | } 742 | 743 | virtual auto cmd_set_blend_factor(const FLOAT BlendFactor[4]) -> void { 744 | this->commands.push_back(command_t::SET_BLEND_FACTOR); 745 | } 746 | 747 | virtual auto cmd_set_stencil_ref(UINT StencilRef) -> void { 748 | encode( 749 | command_t::SET_STENCIL_REF, 750 | set_stencil_ref_t { StencilRef } 751 | ); 752 | } 753 | 754 | virtual auto cmd_set_depth_bounds(FLOAT Min, FLOAT Max) -> void { 755 | encode( 756 | command_t::SET_DEPTH_BOUNDS, 757 | set_depth_bounds_t { Min, Max } 758 | ); 759 | } 760 | 761 | virtual auto cmd_set_index_buffer(const D3D12_INDEX_BUFFER_VIEW *pView) -> void { 762 | this->commands.push_back(command_t::SET_INDEX_BUFFER); 763 | } 764 | 765 | virtual auto cmd_copy_buffer_region( 766 | ID3D12Resource *pDstBuffer, 767 | UINT64 DstOffset, 768 | ID3D12Resource *pSrcBuffer, 769 | UINT64 SrcOffset, 770 | UINT64 NumBytes 771 | ) -> void { 772 | this->commands.push_back(command_t::COPY_BUFFER_REGION); 773 | } 774 | 775 | virtual auto cmd_copy_texture_region( 776 | const D3D12_TEXTURE_COPY_LOCATION *pDst, 777 | UINT DstX, 778 | UINT DstY, 779 | UINT DstZ, 780 | const D3D12_TEXTURE_COPY_LOCATION *pSrc, 781 | const D3D12_BOX *pSrcBox 782 | ) -> void { 783 | this->commands.push_back(command_t::COPY_TEXTURE_REGION); 784 | } 785 | 786 | virtual auto cmd_set_pipeline_state(ID3D12PipelineState *pPipelineState) -> void { 787 | encode( 788 | command_t::SET_PIPELINE_STATE, 789 | set_pipeline_state_t { pPipelineState } 790 | ); 791 | } 792 | 793 | virtual auto cmd_dispatch( 794 | UINT ThreadGroupCountX, 795 | UINT ThreadGroupCountY, 796 | UINT ThreadGroupCountZ 797 | ) -> void { 798 | encode( 799 | command_t::DISPATCH, 800 | dispatch_t { 801 | ThreadGroupCountX, 802 | ThreadGroupCountY, 803 | ThreadGroupCountZ, 804 | } 805 | ); 806 | } 807 | 808 | virtual auto cmd_draw_instanced( 809 | UINT VertexCountPerInstance, 810 | UINT InstanceCount, 811 | UINT StartVertexLocation, 812 | UINT StartInstanceLocation 813 | ) -> void { 814 | encode( 815 | command_t::DRAW_INSTANCED, 816 | draw_instanced_t { 817 | VertexCountPerInstance, 818 | InstanceCount, 819 | StartVertexLocation, 820 | StartInstanceLocation, 821 | } 822 | ); 823 | } 824 | 825 | virtual auto cmd_draw_indexed_instanced( 826 | UINT IndexCountPerInstance, 827 | UINT InstanceCount, 828 | UINT StartIndexLocation, 829 | INT BaseVertexLocation, 830 | UINT StartInstanceLocation 831 | ) -> void { 832 | encode( 833 | command_t::DRAW_INDEXED_INSTANCED, 834 | draw_indexed_instanced_t { 835 | IndexCountPerInstance, 836 | InstanceCount, 837 | StartIndexLocation, 838 | BaseVertexLocation, 839 | StartInstanceLocation, 840 | } 841 | ); 842 | } 843 | 844 | virtual auto cmd_execute_indirect( 845 | ID3D12CommandSignature *pCommandSignature, 846 | UINT MaxCommandCount, 847 | ID3D12Resource *pArgumentBuffer, 848 | UINT64 ArgumentBufferOffset, 849 | ID3D12Resource *pCountBuffer, 850 | UINT64 CountBufferOffset 851 | ) -> void { 852 | encode( 853 | command_t::EXECUTE_INDIRECT, 854 | execute_indirect_t { 855 | pCommandSignature, 856 | MaxCommandCount, 857 | pArgumentBuffer, 858 | ArgumentBufferOffset, 859 | pCountBuffer, 860 | CountBufferOffset, 861 | } 862 | ); 863 | } 864 | 865 | virtual auto cmd_resource_barrier( 866 | UINT NumBarriers, 867 | const D3D12_RESOURCE_BARRIER *pBarriers 868 | ) -> void { 869 | this->commands.push_back(command_t::RESOURCE_BARRIER); 870 | // TODO 871 | } 872 | 873 | virtual auto cmd_set_vertex_buffers( 874 | UINT StartSlot, 875 | UINT NumViews, 876 | const D3D12_VERTEX_BUFFER_VIEW *pViews 877 | ) -> void { 878 | this->commands.push_back(command_t::SET_VERTEX_BUFFERS); 879 | encode_raw(StartSlot); 880 | encode_raw(NumViews); 881 | for (auto i : range(NumViews)) { 882 | encode_raw(pViews[i]); 883 | } 884 | } 885 | 886 | private: 887 | template 888 | auto encode(command_t cmd, T cmd_data) { 889 | this->commands.push_back(cmd); 890 | encode_raw(cmd_data); 891 | } 892 | 893 | template 894 | auto encode_raw(T data) { 895 | auto const raw_data { reinterpret_cast(&data) }; 896 | for (auto i : range(sizeof(T))) { 897 | this->data.push_back(raw_data[i]); 898 | } 899 | } 900 | 901 | private: 902 | std::vector commands; 903 | std::vector data; 904 | }; 905 | -------------------------------------------------------------------------------- /src/descriptors_cpu.hpp: -------------------------------------------------------------------------------- 1 | #pragma once 2 | 3 | #include "icd.hpp" 4 | 5 | #include 6 | #include 7 | 8 | #include 9 | #include 10 | 11 | using namespace Microsoft::WRL; 12 | 13 | /// A CPU descriptor heap. 14 | template 15 | class heap_cpu_t { 16 | public: 17 | heap_cpu_t(ComPtr heap, UINT handle_size) : 18 | heap { heap }, 19 | occupacy { 0 }, 20 | start { heap->GetCPUDescriptorHandleForHeapStart() }, 21 | handle_size { handle_size } 22 | { } 23 | 24 | auto alloc() -> D3D12_CPU_DESCRIPTOR_HANDLE { 25 | if (full()) { 26 | assert("Heap is full"); 27 | } 28 | 29 | // TODO: possible optimizations via bitscans.. 30 | auto slot { 0 }; 31 | for (; slot < N; ++slot) { 32 | if (!this->occupacy[slot]) { 33 | this->occupacy.set(slot); 34 | break; 35 | } 36 | } 37 | 38 | return D3D12_CPU_DESCRIPTOR_HANDLE { 39 | this->start.ptr + this->handle_size * slot 40 | }; 41 | } 42 | 43 | auto free(D3D12_CPU_DESCRIPTOR_HANDLE handle) -> void { 44 | this->occupacy.reset((handle.ptr - this->start.ptr) / this->handle_size); 45 | } 46 | 47 | auto full() const -> bool { 48 | return this->occupacy.all(); 49 | } 50 | 51 | private: 52 | ComPtr heap; 53 | D3D12_CPU_DESCRIPTOR_HANDLE start; 54 | UINT handle_size; 55 | 56 | std::bitset occupacy; 57 | }; 58 | 59 | using heap_index_t = size_t; 60 | 61 | // CPU descriptor heap manager. 62 | // TODO: multi-thread access 63 | template 64 | class descriptors_cpu_t { 65 | public: 66 | descriptors_cpu_t(ID3D12Device* device) : 67 | device { device }, 68 | heaps { }, 69 | free_list { }, 70 | handle_size { device->GetDescriptorHandleIncrementSize(Ty) } 71 | { } 72 | 73 | auto alloc() -> std::tuple { 74 | do { 75 | if (this->free_list.empty()) { 76 | ComPtr heap { nullptr }; 77 | const D3D12_DESCRIPTOR_HEAP_DESC desc { Ty, N, D3D12_DESCRIPTOR_HEAP_FLAG_NONE, 0 }; 78 | this->device->CreateDescriptorHeap(&desc, IID_PPV_ARGS(&heap)); 79 | 80 | this->heaps.emplace_back(heap, this->handle_size); 81 | this->free_list.push_back(this->heaps.size()-1); 82 | } 83 | 84 | const auto free_heap { free_list.back() }; 85 | if (this->heaps[free_heap].full()) { 86 | free_list.pop_back(); 87 | } else { 88 | return std::make_tuple(this->heaps[free_heap].alloc(), free_heap); 89 | } 90 | } while (true); 91 | } 92 | 93 | auto free(D3D12_CPU_DESCRIPTOR_HANDLE handle, heap_index_t index) { 94 | this->heaps[index].free(handle); 95 | } 96 | private: 97 | ID3D12Device* device; 98 | std::vector> heaps; 99 | std::vector free_list; 100 | UINT handle_size; 101 | }; 102 | 103 | -------------------------------------------------------------------------------- /src/descriptors_gpu.hpp: -------------------------------------------------------------------------------- 1 | #pragma once 2 | 3 | #include "icd.hpp" 4 | 5 | #include 6 | #include 7 | 8 | #include 9 | #include 10 | 11 | #include 12 | 13 | using namespace Microsoft::WRL; 14 | using namespace stdx; 15 | 16 | // TODO: multi-threading 17 | class free_list { 18 | public: 19 | free_list(size_t size) : 20 | _size { size } 21 | { 22 | this->_list.push_back(range(_size)); 23 | } 24 | 25 | auto alloc(size_t num) -> std::optional> { 26 | if (!num) { 27 | return std::optional>(range_t(0u, 0u)); 28 | } 29 | 30 | for (auto it = this->_list.begin(); it != this->_list.end(); ++it) { 31 | if (it->_end >= it->_start + num) { 32 | std::list> tail; 33 | tail.splice(tail.begin(), this->_list, it, this->_list.end()); 34 | 35 | range_t node { tail.front() }; 36 | const range_t allocated { node._start, node._start + num }; 37 | node._start += num; 38 | 39 | if (node._start < node._end) { 40 | this->_list.push_back(node); 41 | } 42 | tail.pop_front(); 43 | this->_list.splice(this->_list.end(), tail); 44 | 45 | return allocated; 46 | } 47 | } 48 | 49 | return std::nullopt; 50 | } 51 | 52 | private: 53 | size_t _size; 54 | std::list> _list; 55 | }; 56 | 57 | using descriptor_cpu_gpu_handle_t = std::tuple; 58 | 59 | /// GPU descriptor heap 60 | template 61 | class descriptors_gpu_t { 62 | public: 63 | descriptors_gpu_t(ID3D12Device* device) : 64 | _handle_size { device->GetDescriptorHandleIncrementSize(Ty) }, 65 | allocator { N } 66 | { 67 | const D3D12_DESCRIPTOR_HEAP_DESC desc { Ty, N, D3D12_DESCRIPTOR_HEAP_FLAG_SHADER_VISIBLE, 0 }; 68 | auto const hr { device->CreateDescriptorHeap(&desc, IID_PPV_ARGS(&this->_heap)) }; 69 | // TODO: error 70 | 71 | this->start_cpu = this->_heap->GetCPUDescriptorHandleForHeapStart(); 72 | this->start_gpu = this->_heap->GetGPUDescriptorHandleForHeapStart(); 73 | } 74 | 75 | /// Allocate a slice of the descriptor heap. 76 | auto alloc(size_t num) -> descriptor_cpu_gpu_handle_t { 77 | const auto range = this->allocator.alloc(num); 78 | if (range) { 79 | return std::make_tuple( 80 | D3D12_CPU_DESCRIPTOR_HANDLE { this->start_cpu.ptr + this->_handle_size * range->_start }, 81 | D3D12_GPU_DESCRIPTOR_HANDLE { this->start_gpu.ptr + this->_handle_size * range->_start } 82 | ); 83 | } else { 84 | // TODO 85 | assert(!"Not enough free descriptors in the allocator"); 86 | return std::make_tuple( 87 | D3D12_CPU_DESCRIPTOR_HANDLE { 0 }, 88 | D3D12_GPU_DESCRIPTOR_HANDLE { 0 } 89 | ); 90 | } 91 | } 92 | 93 | /// Free a descriptor heap slice. 94 | auto free(descriptor_cpu_gpu_handle_t handle, size_t num) -> void { 95 | WARN("GPU descriptor free unimplemented"); 96 | } 97 | 98 | auto handle_size() const -> UINT { 99 | return this->_handle_size; 100 | } 101 | 102 | auto heap() const -> ID3D12DescriptorHeap* { 103 | return this->_heap.Get(); 104 | } 105 | 106 | private: 107 | ComPtr _heap; 108 | D3D12_CPU_DESCRIPTOR_HANDLE start_cpu; 109 | D3D12_GPU_DESCRIPTOR_HANDLE start_gpu; 110 | UINT _handle_size; 111 | 112 | free_list allocator; 113 | }; 114 | -------------------------------------------------------------------------------- /src/descriptors_virtual.hpp: -------------------------------------------------------------------------------- 1 | #pragma once 2 | 3 | #include "descriptors_cpu.hpp" 4 | #include "descriptors_gpu.hpp" 5 | 6 | #include 7 | #include 8 | 9 | 10 | struct placed_descriptor_set_t { 11 | size_t ref_count; 12 | descriptor_cpu_gpu_handle_t start; 13 | size_t num; 14 | }; 15 | 16 | template 17 | struct sampler_heap_t { 18 | public: 19 | sampler_heap_t(ID3D12Device* device) : 20 | heap { device } 21 | { } 22 | 23 | public: 24 | descriptors_gpu_t heap; 25 | std::map placed_sets; 26 | }; 27 | 28 | template 29 | class sampler_heap_cpu_t { 30 | public: 31 | sampler_heap_cpu_t(ID3D12Device* device) : 32 | _handle_size { device->GetDescriptorHandleIncrementSize(D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER) }, 33 | allocator { N } 34 | { 35 | const D3D12_DESCRIPTOR_HEAP_DESC desc { 36 | D3D12_DESCRIPTOR_HEAP_TYPE_SAMPLER, 37 | N, 38 | D3D12_DESCRIPTOR_HEAP_FLAG_NONE, 39 | 0, 40 | }; 41 | auto const hr { device->CreateDescriptorHeap(&desc, IID_PPV_ARGS(&this->_heap)) }; 42 | // TODO: error 43 | 44 | this->start = this->_heap->GetCPUDescriptorHandleForHeapStart(); 45 | } 46 | 47 | auto alloc(size_t num) -> D3D12_CPU_DESCRIPTOR_HANDLE { 48 | const auto range = this->allocator.alloc(num); 49 | if (range) { 50 | return D3D12_CPU_DESCRIPTOR_HANDLE { this->start.ptr + this->_handle_size * range->_start }; 51 | } else { 52 | // TODO 53 | assert(!"Not enough free descriptors in the allocator"); 54 | return D3D12_CPU_DESCRIPTOR_HANDLE { 0 }; 55 | } 56 | } 57 | 58 | auto assign_set(D3D12_CPU_DESCRIPTOR_HANDLE handle, struct descriptor_set_t* set) { 59 | this->_offset_to_set.emplace(static_cast(handle.ptr - this->start.ptr), set); 60 | } 61 | 62 | auto get_set(UINT offset) -> struct descriptor_set_t* { 63 | return this->_offset_to_set[offset]; 64 | } 65 | 66 | auto handle_size() const -> UINT { 67 | return this->_handle_size; 68 | } 69 | 70 | auto heap() const -> ID3D12DescriptorHeap* { 71 | return this->_heap.Get(); 72 | } 73 | 74 | private: 75 | ComPtr _heap; 76 | D3D12_CPU_DESCRIPTOR_HANDLE start; 77 | UINT _handle_size; 78 | 79 | free_list allocator; 80 | std::map _offset_to_set; 81 | }; 82 | 83 | template 84 | class sampler_heaps_t { 85 | public: 86 | sampler_heaps_t(ID3D12Device* device) : 87 | _gpu_heaps { }, 88 | _cpu_heap { device }, 89 | _device { device } 90 | { } 91 | 92 | /// Allocate a slice of the descriptor heap. 93 | /// 94 | /// Only a slice of the underlying CPU heap! 95 | auto alloc(size_t num) -> D3D12_CPU_DESCRIPTOR_HANDLE { 96 | return this->_cpu_heap.alloc(num); 97 | } 98 | 99 | auto add_gpu_heap() -> heap_index_t { 100 | this->_gpu_heaps.push_back( 101 | sampler_heap_t(this->_device) 102 | ); 103 | return this->_gpu_heaps.size() - 1; 104 | } 105 | 106 | auto handle_size() const -> UINT { 107 | return this->_cpu_heap.handle_size(); 108 | } 109 | 110 | auto cpu_heap() const -> ID3D12DescriptorHeap* { 111 | return this->_cpu_heap.heap(); 112 | } 113 | 114 | auto sampler_heap(heap_index_t index) -> sampler_heap_t& { 115 | return this->_gpu_heaps[index]; 116 | } 117 | 118 | auto assign_set(D3D12_CPU_DESCRIPTOR_HANDLE handle, struct descriptor_set_t* set) { 119 | this->_cpu_heap.assign_set(handle, set); 120 | } 121 | 122 | auto get_set(UINT offset) -> struct descriptor_set_t* { 123 | return this->_cpu_heap.get_set(offset); 124 | } 125 | 126 | private: 127 | // Only remove heaps on destruction, we keeping weak indices into this vector. 128 | std::vector> _gpu_heaps; 129 | sampler_heap_cpu_t _cpu_heap; 130 | 131 | ID3D12Device* _device; 132 | }; 133 | -------------------------------------------------------------------------------- /src/device.hpp: -------------------------------------------------------------------------------- 1 | #pragma once 2 | 3 | #include "descriptors_virtual.hpp" 4 | #include "impl.hpp" 5 | 6 | #include 7 | 8 | class device_t final { 9 | public: 10 | device_t(ComPtr device) : 11 | loader_magic { ICD_LOADER_MAGIC }, 12 | descriptors_cpu_rtv { device.Get() }, 13 | descriptors_cpu_dsv { device.Get() }, 14 | descriptors_cpu_cbv_srv_uav { device.Get() }, 15 | descriptors_cpu_sampler { device.Get() }, 16 | descriptors_gpu_cbv_srv_uav { device.Get() }, 17 | descriptors_gpu_sampler { device.Get() }, 18 | pso_buffer_to_image { nullptr }, 19 | signature_buffer_to_image { nullptr }, 20 | device { device } 21 | { } 22 | 23 | ~device_t() {} 24 | 25 | auto operator->() -> ID3D12Device3* { 26 | return this->device.Get(); 27 | } 28 | 29 | auto create_render_target_view( 30 | image_t* image, 31 | VkImageViewType type, 32 | DXGI_FORMAT format, 33 | VkImageSubresourceRange const& range 34 | ) -> std::tuple { 35 | auto resource { image->resource.Get() }; 36 | 37 | const auto layer_count { 38 | range.layerCount == VK_REMAINING_ARRAY_LAYERS ? 39 | image->resource_desc.DepthOrArraySize - range.baseArrayLayer : 40 | range.layerCount 41 | }; 42 | // TODO: multisampling 43 | 44 | auto handle { this->descriptors_cpu_rtv.alloc() }; 45 | 46 | D3D12_RENDER_TARGET_VIEW_DESC desc { format }; 47 | if (type == VK_IMAGE_TYPE_2D && image->resource_desc.SampleDesc.Count != VK_SAMPLE_COUNT_1_BIT) { 48 | desc.ViewDimension = D3D12_RTV_DIMENSION_TEXTURE2DMS; 49 | desc.Texture2DMS = D3D12_TEX2DMS_RTV { }; 50 | } else if (type == VK_IMAGE_VIEW_TYPE_1D && range.baseArrayLayer == 0) { 51 | desc.ViewDimension = D3D12_RTV_DIMENSION_TEXTURE1D; 52 | desc.Texture1D = D3D12_TEX1D_RTV { range.baseMipLevel }; 53 | } else if (type == VK_IMAGE_VIEW_TYPE_1D || type == VK_IMAGE_VIEW_TYPE_1D_ARRAY) { 54 | desc.ViewDimension = D3D12_RTV_DIMENSION_TEXTURE1DARRAY; 55 | desc.Texture1DArray = D3D12_TEX1D_ARRAY_RTV { 56 | range.baseMipLevel, 57 | range.baseArrayLayer, 58 | layer_count 59 | }; 60 | } else if (type == VK_IMAGE_VIEW_TYPE_2D && range.baseArrayLayer == 0) { 61 | desc.ViewDimension = D3D12_RTV_DIMENSION_TEXTURE2D; 62 | desc.Texture2D = D3D12_TEX2D_RTV { range.baseMipLevel, 0 }; 63 | } else if (type == VK_IMAGE_VIEW_TYPE_2D || type == VK_IMAGE_VIEW_TYPE_2D_ARRAY || 64 | type == VK_IMAGE_VIEW_TYPE_CUBE || type == VK_IMAGE_VIEW_TYPE_CUBE_ARRAY) 65 | { 66 | desc.ViewDimension = D3D12_RTV_DIMENSION_TEXTURE2DARRAY; 67 | desc.Texture2DArray = D3D12_TEX2D_ARRAY_RTV { 68 | range.baseMipLevel, 69 | range.baseArrayLayer, 70 | layer_count, 71 | 0 72 | }; 73 | } else if (type == VK_IMAGE_VIEW_TYPE_3D) { 74 | desc.ViewDimension = D3D12_RTV_DIMENSION_TEXTURE3D; 75 | desc.Texture3D = D3D12_TEX3D_RTV { range.baseMipLevel, 0, ~0u }; 76 | } 77 | 78 | this->device->CreateRenderTargetView(resource, &desc, std::get<0>(handle)); 79 | return handle; 80 | } 81 | 82 | auto create_depth_stencil_view( 83 | image_t* image, 84 | VkImageViewType type, 85 | DXGI_FORMAT format, 86 | VkImageSubresourceRange const& range 87 | ) -> std::tuple { 88 | auto resource { image->resource.Get() }; 89 | 90 | const auto layer_count { 91 | range.layerCount == VK_REMAINING_ARRAY_LAYERS ? 92 | image->resource_desc.DepthOrArraySize - range.baseArrayLayer : 93 | range.layerCount 94 | }; 95 | // TODO: multisampling 96 | 97 | auto handle { this->descriptors_cpu_dsv.alloc() }; 98 | 99 | D3D12_DEPTH_STENCIL_VIEW_DESC desc { format }; 100 | if (type == VK_IMAGE_TYPE_2D && image->resource_desc.SampleDesc.Count != VK_SAMPLE_COUNT_1_BIT) { 101 | desc.ViewDimension = D3D12_DSV_DIMENSION_TEXTURE2DMS; 102 | desc.Texture2DMS = D3D12_TEX2DMS_DSV { }; 103 | } else if (type == VK_IMAGE_VIEW_TYPE_1D && range.baseArrayLayer == 0) { 104 | desc.ViewDimension = D3D12_DSV_DIMENSION_TEXTURE1D; 105 | desc.Texture1D = D3D12_TEX1D_DSV { range.baseMipLevel }; 106 | } else if (type == VK_IMAGE_VIEW_TYPE_1D || type == VK_IMAGE_VIEW_TYPE_1D_ARRAY) { 107 | desc.ViewDimension = D3D12_DSV_DIMENSION_TEXTURE1DARRAY; 108 | desc.Texture1DArray = D3D12_TEX1D_ARRAY_DSV { 109 | range.baseMipLevel, 110 | range.baseArrayLayer, 111 | range.layerCount 112 | }; 113 | } else if (type == VK_IMAGE_VIEW_TYPE_2D && range.baseArrayLayer == 0) { 114 | desc.ViewDimension = D3D12_DSV_DIMENSION_TEXTURE2D; 115 | desc.Texture2D = D3D12_TEX2D_DSV { range.baseMipLevel }; 116 | } else if (type == VK_IMAGE_VIEW_TYPE_2D || type == VK_IMAGE_VIEW_TYPE_2D_ARRAY || 117 | type == VK_IMAGE_VIEW_TYPE_CUBE || type == VK_IMAGE_VIEW_TYPE_CUBE_ARRAY) 118 | { 119 | desc.ViewDimension = D3D12_DSV_DIMENSION_TEXTURE2DARRAY; 120 | desc.Texture2DArray = D3D12_TEX2D_ARRAY_DSV { 121 | range.baseMipLevel, 122 | range.baseArrayLayer, 123 | range.layerCount, 124 | }; 125 | } 126 | 127 | this->device->CreateDepthStencilView(resource, &desc, std::get<0>(handle)); 128 | 129 | return handle; 130 | } 131 | 132 | auto create_shader_resource_view( 133 | image_t *image, 134 | VkImageViewType type, 135 | DXGI_FORMAT format, 136 | VkComponentMapping components, 137 | VkImageSubresourceRange const& range 138 | ) -> std::tuple { 139 | auto handle { this->descriptors_cpu_cbv_srv_uav.alloc() }; 140 | auto resource { image->resource.Get() }; 141 | 142 | const auto layer_count { 143 | range.layerCount == VK_REMAINING_ARRAY_LAYERS ? 144 | image->resource_desc.DepthOrArraySize - range.baseArrayLayer : 145 | range.layerCount 146 | }; 147 | const auto level_count { 148 | range.levelCount == VK_REMAINING_MIP_LEVELS ? 149 | image->resource_desc.MipLevels - range.baseMipLevel : 150 | range.levelCount 151 | }; 152 | 153 | // TODO: multisampling 154 | 155 | // TODO: other formats 156 | switch (format) { 157 | case DXGI_FORMAT_D16_UNORM: format = DXGI_FORMAT_R16_UNORM; break; 158 | case DXGI_FORMAT_D32_FLOAT: format = DXGI_FORMAT_R32_FLOAT; break; 159 | case DXGI_FORMAT_D32_FLOAT_S8X24_UINT: { 160 | // Single bit only for SRV (descriptor set only view) 161 | if (range.aspectMask & VK_IMAGE_ASPECT_DEPTH_BIT) { 162 | format = DXGI_FORMAT_R32_FLOAT_X8X24_TYPELESS; 163 | } else { 164 | format = DXGI_FORMAT_X32_TYPELESS_G8X24_UINT; 165 | } 166 | } break; 167 | } 168 | 169 | D3D12_SHADER_RESOURCE_VIEW_DESC desc { format }; 170 | 171 | auto component = [] (VkComponentSwizzle component, D3D12_SHADER_COMPONENT_MAPPING identity) { 172 | switch (component) { 173 | case VK_COMPONENT_SWIZZLE_IDENTITY: return identity; 174 | case VK_COMPONENT_SWIZZLE_ZERO: return D3D12_SHADER_COMPONENT_MAPPING_FORCE_VALUE_0; 175 | case VK_COMPONENT_SWIZZLE_ONE: return D3D12_SHADER_COMPONENT_MAPPING_FORCE_VALUE_1; 176 | case VK_COMPONENT_SWIZZLE_R: return D3D12_SHADER_COMPONENT_MAPPING_FROM_MEMORY_COMPONENT_0; 177 | case VK_COMPONENT_SWIZZLE_G: return D3D12_SHADER_COMPONENT_MAPPING_FROM_MEMORY_COMPONENT_1; 178 | case VK_COMPONENT_SWIZZLE_B: return D3D12_SHADER_COMPONENT_MAPPING_FROM_MEMORY_COMPONENT_2; 179 | case VK_COMPONENT_SWIZZLE_A: return D3D12_SHADER_COMPONENT_MAPPING_FROM_MEMORY_COMPONENT_3; 180 | default: return identity; 181 | } 182 | }; 183 | 184 | desc.Shader4ComponentMapping = D3D12_ENCODE_SHADER_4_COMPONENT_MAPPING( 185 | component(components.r, D3D12_SHADER_COMPONENT_MAPPING_FROM_MEMORY_COMPONENT_0), 186 | component(components.g, D3D12_SHADER_COMPONENT_MAPPING_FROM_MEMORY_COMPONENT_1), 187 | component(components.b, D3D12_SHADER_COMPONENT_MAPPING_FROM_MEMORY_COMPONENT_2), 188 | component(components.a, D3D12_SHADER_COMPONENT_MAPPING_FROM_MEMORY_COMPONENT_3) 189 | ); 190 | 191 | if (type == VK_IMAGE_VIEW_TYPE_1D && range.baseArrayLayer == 0) { 192 | desc.ViewDimension = D3D12_SRV_DIMENSION_TEXTURE1D; 193 | desc.Texture1D = D3D12_TEX1D_SRV { 194 | range.baseMipLevel, 195 | level_count, 196 | 0.0 // TODO: ? 197 | }; 198 | } else if (type == VK_IMAGE_VIEW_TYPE_1D || type == VK_IMAGE_VIEW_TYPE_1D_ARRAY) { 199 | desc.ViewDimension = D3D12_SRV_DIMENSION_TEXTURE1DARRAY; 200 | desc.Texture1DArray = D3D12_TEX1D_ARRAY_SRV { 201 | range.baseMipLevel, 202 | level_count, 203 | range.baseArrayLayer, 204 | layer_count, 205 | 0.0 // TODO: ? 206 | }; 207 | } else if (type == VK_IMAGE_VIEW_TYPE_2D && range.baseArrayLayer == 0) { 208 | desc.ViewDimension = D3D12_SRV_DIMENSION_TEXTURE2D; 209 | desc.Texture2D = D3D12_TEX2D_SRV { 210 | range.baseMipLevel, 211 | level_count, 212 | 0, 213 | 0.0 // TODO: ? 214 | }; 215 | } else if (type == VK_IMAGE_VIEW_TYPE_2D || type == VK_IMAGE_VIEW_TYPE_2D_ARRAY) { 216 | desc.ViewDimension = D3D12_SRV_DIMENSION_TEXTURE2DARRAY; 217 | desc.Texture2DArray = D3D12_TEX2D_ARRAY_SRV { 218 | range.baseMipLevel, 219 | level_count, 220 | range.baseArrayLayer, 221 | layer_count, 222 | 0, 223 | 0.0 // TODO ? 224 | }; 225 | } else if (type == VK_IMAGE_VIEW_TYPE_CUBE && range.baseArrayLayer == 0) { 226 | desc.ViewDimension = D3D12_SRV_DIMENSION_TEXTURECUBE; 227 | desc.TextureCube = D3D12_TEXCUBE_SRV { 228 | range.baseMipLevel, 229 | level_count, 230 | 0.0 // TODO ? 231 | }; 232 | } else if (type == VK_IMAGE_VIEW_TYPE_CUBE || type == VK_IMAGE_VIEW_TYPE_CUBE_ARRAY) { 233 | desc.ViewDimension = D3D12_SRV_DIMENSION_TEXTURECUBEARRAY; 234 | desc.TextureCubeArray = D3D12_TEXCUBE_ARRAY_SRV { 235 | range.baseMipLevel, 236 | level_count, 237 | range.baseArrayLayer, 238 | layer_count / 6, 239 | 0.0 // TODO ? 240 | }; 241 | } else if (type == VK_IMAGE_VIEW_TYPE_3D) { 242 | desc.ViewDimension = D3D12_SRV_DIMENSION_TEXTURE3D; 243 | desc.Texture3D = D3D12_TEX3D_SRV { 244 | range.baseMipLevel, 245 | level_count, 246 | 0.0 // TODO ? 247 | }; 248 | } 249 | 250 | this->device->CreateShaderResourceView(resource, &desc, std::get<0>(handle)); 251 | return handle; 252 | } 253 | 254 | auto create_unordered_access_view( 255 | ID3D12Resource* resource, 256 | VkImageViewType type, 257 | DXGI_FORMAT format, 258 | VkImageSubresourceRange const& range 259 | ) -> std::tuple { 260 | // TODO 261 | assert(range.layerCount != VK_REMAINING_ARRAY_LAYERS); 262 | 263 | auto handle { this->descriptors_cpu_cbv_srv_uav.alloc() }; 264 | 265 | D3D12_UNORDERED_ACCESS_VIEW_DESC desc { format }; 266 | if (type == VK_IMAGE_VIEW_TYPE_1D && range.baseArrayLayer == 0) { 267 | desc.ViewDimension = D3D12_UAV_DIMENSION_TEXTURE1D; 268 | desc.Texture1D = D3D12_TEX1D_UAV { range.baseMipLevel }; 269 | } else if (type == VK_IMAGE_VIEW_TYPE_1D || type == VK_IMAGE_VIEW_TYPE_1D_ARRAY) { 270 | desc.ViewDimension = D3D12_UAV_DIMENSION_TEXTURE1DARRAY; 271 | desc.Texture1DArray = D3D12_TEX1D_ARRAY_UAV { 272 | range.baseMipLevel, 273 | range.baseArrayLayer, 274 | range.layerCount 275 | }; 276 | } else if (type == VK_IMAGE_VIEW_TYPE_2D && range.baseArrayLayer == 0) { 277 | desc.ViewDimension = D3D12_UAV_DIMENSION_TEXTURE2D; 278 | desc.Texture2D = D3D12_TEX2D_UAV { range.baseMipLevel, 0 }; 279 | } else if (type == VK_IMAGE_VIEW_TYPE_2D || type == VK_IMAGE_VIEW_TYPE_2D_ARRAY || 280 | type == VK_IMAGE_VIEW_TYPE_CUBE || type == VK_IMAGE_VIEW_TYPE_CUBE_ARRAY) 281 | { 282 | desc.ViewDimension = D3D12_UAV_DIMENSION_TEXTURE2DARRAY; 283 | desc.Texture2DArray = D3D12_TEX2D_ARRAY_UAV { 284 | range.baseMipLevel, 285 | range.baseArrayLayer, 286 | range.layerCount, 287 | 0 288 | }; 289 | } else if (type == VK_IMAGE_VIEW_TYPE_3D) { 290 | desc.ViewDimension = D3D12_UAV_DIMENSION_TEXTURE3D; 291 | desc.Texture3D = D3D12_TEX3D_UAV { range.baseMipLevel, 0, ~0u }; 292 | } 293 | 294 | this->device->CreateUnorderedAccessView( 295 | resource, 296 | nullptr, // counter 297 | &desc, 298 | std::get<0>(handle) 299 | ); 300 | return handle; 301 | } 302 | 303 | auto destroy_render_target_view(std::tuple view) -> void { 304 | const auto [handle, index] = view; 305 | this->descriptors_cpu_rtv.free(handle, index); 306 | } 307 | 308 | auto destroy_depth_stencil_view(std::tuple view) -> void { 309 | const auto [handle, index] = view; 310 | this->descriptors_cpu_dsv.free(handle, index); 311 | } 312 | 313 | auto destroy_shader_resource_view(std::tuple view) -> void { 314 | const auto [handle, index] = view; 315 | this->descriptors_cpu_cbv_srv_uav.free(handle, index); 316 | } 317 | 318 | auto destroy_unordered_access_view(std::tuple view) -> void { 319 | const auto [handle, index] = view; 320 | this->descriptors_cpu_cbv_srv_uav.free(handle, index); 321 | } 322 | 323 | private: 324 | /// Dispatchable object 325 | uintptr_t loader_magic; 326 | 327 | public: 328 | ComPtr device; 329 | 330 | queue_t present_queue; 331 | std::vector general_queues; 332 | std::vector compute_queues; 333 | std::vector copy_queues; 334 | 335 | span heap_properties; 336 | 337 | // CPU descriptor heaps 338 | descriptors_cpu_t descriptors_cpu_rtv; 339 | descriptors_cpu_t descriptors_cpu_dsv; 340 | descriptors_cpu_t descriptors_cpu_cbv_srv_uav; 341 | descriptors_cpu_t descriptors_cpu_sampler; 342 | 343 | // GPU descriptor heaps 344 | descriptors_gpu_t descriptors_gpu_cbv_srv_uav; 345 | sampler_heaps_t<2048, 1'000'000> descriptors_gpu_sampler; 346 | 347 | // Indirect execution signatures 348 | ComPtr dispatch_indirect; 349 | std::unordered_map> draw_indirect; 350 | std::unordered_map> draw_indexed_indirect; 351 | // Indirect draw/draw_indirect signature access token 352 | std::shared_mutex draw_indirect_access; 353 | 354 | ComPtr pso_buffer_to_image { nullptr }; 355 | ComPtr signature_buffer_to_image { nullptr }; 356 | 357 | ComPtr pso_blit_2d { nullptr }; 358 | ComPtr signature_blit_2d { nullptr }; 359 | }; 360 | -------------------------------------------------------------------------------- /src/formats.cpp: -------------------------------------------------------------------------------- 1 | 2 | #include "icd.hpp" 3 | 4 | // TODO 5 | extern std::array formats { 6 | DXGI_FORMAT_UNKNOWN, // UNDEFINED 7 | DXGI_FORMAT_UNKNOWN, // R4G4_UNORM_PACK8 8 | DXGI_FORMAT_UNKNOWN, // R4G4B4A4_UNORM_PACK16 9 | DXGI_FORMAT_UNKNOWN, // B4G4R4A4_UNORM_PACK16 10 | DXGI_FORMAT_UNKNOWN, // R5G6B5_UNORM_PACK16 11 | DXGI_FORMAT_UNKNOWN, // B5G6R5_UNORM_PACK16 12 | DXGI_FORMAT_UNKNOWN, // R5G5B5A1_UNORM_PACK16 13 | DXGI_FORMAT_UNKNOWN, // B5G5R5A1_UNORM_PACK16 14 | DXGI_FORMAT_UNKNOWN, // A1R5G5B5_UNORM_PACK16 15 | DXGI_FORMAT_R8_UNORM, // R8_UNORM 16 | DXGI_FORMAT_R8_SNORM, // R8_SNORM 17 | DXGI_FORMAT_UNKNOWN, // R8_USCALED 18 | DXGI_FORMAT_UNKNOWN, // R8_SSCALED 19 | DXGI_FORMAT_R8_UINT, // R8_UINT 20 | DXGI_FORMAT_R8_SINT, // R8_SINT 21 | DXGI_FORMAT_UNKNOWN, // R8_SRGB 22 | DXGI_FORMAT_R8G8_UNORM, // R8G8_UNORM 23 | DXGI_FORMAT_R8G8_SNORM, // R8G8_SNORM 24 | DXGI_FORMAT_UNKNOWN, // R8G8_USCALED 25 | DXGI_FORMAT_UNKNOWN, // R8G8_SSCALED 26 | DXGI_FORMAT_R8G8_UINT, // R8G8_UINT 27 | DXGI_FORMAT_R8G8_SINT, // R8G8_SINT 28 | DXGI_FORMAT_UNKNOWN, // R8G8_SRGB 29 | DXGI_FORMAT_UNKNOWN, // R8G8B8_UNORM 30 | DXGI_FORMAT_UNKNOWN, // R8G8B8_SNORM 31 | DXGI_FORMAT_UNKNOWN, // R8G8B8_USCALED 32 | DXGI_FORMAT_UNKNOWN, // R8G8B8_SSCALED 33 | DXGI_FORMAT_UNKNOWN, // R8G8B8_UINT 34 | DXGI_FORMAT_UNKNOWN, // R8G8B8_SINT 35 | DXGI_FORMAT_UNKNOWN, // R8G8B8_SRGB 36 | DXGI_FORMAT_UNKNOWN, // B8G8R8_UNORM 37 | DXGI_FORMAT_UNKNOWN, // B8G8R8_SNORM 38 | DXGI_FORMAT_UNKNOWN, // B8G8R8_USCALED 39 | DXGI_FORMAT_UNKNOWN, // B8G8R8_SSCALED 40 | DXGI_FORMAT_UNKNOWN, // B8G8R8_UINT 41 | DXGI_FORMAT_UNKNOWN, // B8G8R8_SINT 42 | DXGI_FORMAT_UNKNOWN, // B8G8R8_SRGB 43 | DXGI_FORMAT_R8G8B8A8_UNORM, // R8G8B8A8_UNORM 44 | DXGI_FORMAT_R8G8B8A8_SNORM, // R8G8B8A8_SNORM 45 | DXGI_FORMAT_UNKNOWN, // R8G8B8A8_USCALED 46 | DXGI_FORMAT_UNKNOWN, // R8G8B8A8_SSCALED 47 | DXGI_FORMAT_R8G8B8A8_UINT, // R8G8B8A8_UINT 48 | DXGI_FORMAT_R8G8B8A8_SINT, // R8G8B8A8_SINT 49 | DXGI_FORMAT_R8G8B8A8_UNORM_SRGB, // R8G8B8A8_SRGB 50 | DXGI_FORMAT_B8G8R8A8_UNORM, // B8G8R8A8_UNORM 51 | DXGI_FORMAT_UNKNOWN, // B8G8R8A8_SNORM 52 | DXGI_FORMAT_UNKNOWN, // B8G8R8A8_USCALED 53 | DXGI_FORMAT_UNKNOWN, // B8G8R8A8_SSCALED 54 | DXGI_FORMAT_UNKNOWN, // B8G8R8A8_UINT 55 | DXGI_FORMAT_UNKNOWN, // B8G8R8A8_SINT 56 | DXGI_FORMAT_B8G8R8A8_UNORM_SRGB, // B8G8R8A8_SRGB 57 | DXGI_FORMAT_UNKNOWN, // A8B8G8R8_UNORM_PACK32 58 | DXGI_FORMAT_UNKNOWN, // A8B8G8R8_SNORM_PACK32 59 | DXGI_FORMAT_UNKNOWN, // A8B8G8R8_USCALED_PACK32 60 | DXGI_FORMAT_UNKNOWN, // A8B8G8R8_SSCALED_PACK32 61 | DXGI_FORMAT_UNKNOWN, // A8B8G8R8_UINT_PACK32 62 | DXGI_FORMAT_UNKNOWN, // A8B8G8R8_SINT_PACK32 63 | DXGI_FORMAT_UNKNOWN, // A8B8G8R8_SRGB_PACK32 64 | DXGI_FORMAT_UNKNOWN, // A2R10G10B10_UNORM_PACK32 65 | DXGI_FORMAT_UNKNOWN, // A2R10G10B10_SNORM_PACK32 66 | DXGI_FORMAT_UNKNOWN, // A2R10G10B10_USCALED_PACK32 67 | DXGI_FORMAT_UNKNOWN, // A2R10G10B10_SSCALED_PACK32 68 | DXGI_FORMAT_UNKNOWN, // A2R10G10B10_UINT_PACK32 69 | DXGI_FORMAT_UNKNOWN, // A2R10G10B10_SINT_PACK32 70 | DXGI_FORMAT_UNKNOWN, // A2B10G10R10_UNORM_PACK32 71 | DXGI_FORMAT_UNKNOWN, // A2B10G10R10_SNORM_PACK32 72 | DXGI_FORMAT_UNKNOWN, // A2B10G10R10_USCALED_PACK32 73 | DXGI_FORMAT_UNKNOWN, // A2B10G10R10_SSCALED_PACK32 74 | DXGI_FORMAT_UNKNOWN, // A2B10G10R10_UINT_PACK32 75 | DXGI_FORMAT_UNKNOWN, // A2B10G10R10_SINT_PACK32 76 | DXGI_FORMAT_UNKNOWN, // R16_UNORM 77 | DXGI_FORMAT_UNKNOWN, // R16_SNORM 78 | DXGI_FORMAT_UNKNOWN, // R16_USCALED 79 | DXGI_FORMAT_UNKNOWN, // R16_SSCALED 80 | DXGI_FORMAT_UNKNOWN, // R16_UINT 81 | DXGI_FORMAT_UNKNOWN, // R16_SINT 82 | DXGI_FORMAT_UNKNOWN, // R16_SFLOAT 83 | DXGI_FORMAT_UNKNOWN, // R16G16_UNORM 84 | DXGI_FORMAT_UNKNOWN, // R16G16_SNORM 85 | DXGI_FORMAT_UNKNOWN, // R16G16_USCALED 86 | DXGI_FORMAT_UNKNOWN, // R16G16_SSCALED 87 | DXGI_FORMAT_UNKNOWN, // R16G16_UINT 88 | DXGI_FORMAT_UNKNOWN, // R16G16_SINT 89 | DXGI_FORMAT_R16G16_FLOAT, // R16G16_SFLOAT 90 | DXGI_FORMAT_UNKNOWN, // R16G16B16_UNORM 91 | DXGI_FORMAT_UNKNOWN, // R16G16B16_SNORM 92 | DXGI_FORMAT_UNKNOWN, // R16G16B16_USCALED 93 | DXGI_FORMAT_UNKNOWN, // R16G16B16_SSCALED 94 | DXGI_FORMAT_UNKNOWN, // R16G16B16_UINT 95 | DXGI_FORMAT_UNKNOWN, // R16G16B16_SINT 96 | DXGI_FORMAT_UNKNOWN, // R16G16B16_SFLOAT 97 | DXGI_FORMAT_R16G16B16A16_UNORM, // R16G16B16A16_UNORM 98 | DXGI_FORMAT_R16G16B16A16_SNORM, // R16G16B16A16_SNORM 99 | DXGI_FORMAT_UNKNOWN, // R16G16B16A16_USCALED 100 | DXGI_FORMAT_UNKNOWN, // R16G16B16A16_SSCALED 101 | DXGI_FORMAT_R16G16B16A16_UINT, // R16G16B16A16_UINT 102 | DXGI_FORMAT_R16G16B16A16_SINT, // R16G16B16A16_SINT 103 | DXGI_FORMAT_R16G16B16A16_FLOAT, // R16G16B16A16_SFLOAT 104 | DXGI_FORMAT_R32_UINT, // R32_UINT 105 | DXGI_FORMAT_R32_SINT, // R32_SINT 106 | DXGI_FORMAT_R32_FLOAT, // R32_SFLOAT 107 | DXGI_FORMAT_R32G32_UINT, // R32G32_UINT 108 | DXGI_FORMAT_R32G32_SINT, // R32G32_SINT 109 | DXGI_FORMAT_R32G32_FLOAT, // R32G32_SFLOAT 110 | DXGI_FORMAT_R32G32B32_UINT, // R32G32B32_UINT 111 | DXGI_FORMAT_R32G32B32_SINT, // R32G32B32_SINT 112 | DXGI_FORMAT_R32G32B32_FLOAT, // R32G32B32_SFLOAT 113 | DXGI_FORMAT_R32G32B32A32_UINT, // R32G32B32A32_UINT 114 | DXGI_FORMAT_R32G32B32A32_SINT, // R32G32B32A32_SINT 115 | DXGI_FORMAT_R32G32B32A32_FLOAT, // R32G32B32A32_SFLOAT 116 | DXGI_FORMAT_UNKNOWN, // R64_UINT 117 | DXGI_FORMAT_UNKNOWN, // R64_SINT 118 | DXGI_FORMAT_UNKNOWN, // R64_SFLOAT 119 | DXGI_FORMAT_UNKNOWN, // R64G64_UINT 120 | DXGI_FORMAT_UNKNOWN, // R64G64_SINT 121 | DXGI_FORMAT_UNKNOWN, // R64G64_SFLOAT 122 | DXGI_FORMAT_UNKNOWN, // R64G64B64_UINT 123 | DXGI_FORMAT_UNKNOWN, // R64G64B64_SINT 124 | DXGI_FORMAT_UNKNOWN, // R64G64B64_SFLOAT 125 | DXGI_FORMAT_UNKNOWN, // R64G64B64A64_UINT 126 | DXGI_FORMAT_UNKNOWN, // R64G64B64A64_SINT 127 | DXGI_FORMAT_UNKNOWN, // R64G64B64A64_SFLOAT 128 | DXGI_FORMAT_UNKNOWN, // B10G11R11_UFLOAT_PACK32 129 | DXGI_FORMAT_UNKNOWN, // E5B9G9R9_UFLOAT_PACK32 130 | DXGI_FORMAT_D16_UNORM, // D16_UNORM 131 | DXGI_FORMAT_UNKNOWN, // X8_D24_UNORM_PACK32 132 | DXGI_FORMAT_D32_FLOAT, // D32_SFLOAT 133 | DXGI_FORMAT_UNKNOWN, // S8_UINT 134 | DXGI_FORMAT_UNKNOWN, // D16_UNORM_S8_UINT 135 | DXGI_FORMAT_D24_UNORM_S8_UINT, // D24_UNORM_S8_UINT 136 | DXGI_FORMAT_D32_FLOAT_S8X24_UINT, // D32_SFLOAT_S8_UINT 137 | DXGI_FORMAT_BC1_UNORM, // BC1_RGB_UNORM_BLOCK 138 | DXGI_FORMAT_BC1_UNORM_SRGB, // BC1_RGB_SRGB_BLOCK 139 | DXGI_FORMAT_BC1_UNORM, // BC1_RGBA_UNORM_BLOCK 140 | DXGI_FORMAT_BC1_UNORM_SRGB, // BC1_RGBA_SRGB_BLOCK 141 | DXGI_FORMAT_BC2_UNORM, // BC2_UNORM_BLOCK 142 | DXGI_FORMAT_BC2_UNORM_SRGB, // BC2_SRGB_BLOCK 143 | DXGI_FORMAT_BC3_UNORM, // BC3_UNORM_BLOCK 144 | DXGI_FORMAT_BC3_UNORM_SRGB, // BC3_SRGB_BLOCK 145 | DXGI_FORMAT_BC4_UNORM, // BC4_UNORM_BLOCK 146 | DXGI_FORMAT_BC4_SNORM, // BC4_SNORM_BLOCK 147 | DXGI_FORMAT_BC5_UNORM, // BC5_UNORM_BLOCK 148 | DXGI_FORMAT_BC5_SNORM, // BC5_SNORM_BLOCK 149 | DXGI_FORMAT_BC6H_UF16, // BC6H_UFLOAT_BLOCK 150 | DXGI_FORMAT_BC6H_SF16, // BC6H_SFLOAT_BLOCK 151 | DXGI_FORMAT_BC7_UNORM, // BC7_UNORM_BLOCK 152 | DXGI_FORMAT_BC7_UNORM_SRGB, // BC7_SRGB_BLOCK 153 | DXGI_FORMAT_UNKNOWN, // ETC2_R8G8B8_UNORM_BLOCK 154 | DXGI_FORMAT_UNKNOWN, // ETC2_R8G8B8_SRGB_BLOCK 155 | DXGI_FORMAT_UNKNOWN, // ETC2_R8G8B8A1_UNORM_BLOCK 156 | DXGI_FORMAT_UNKNOWN, // ETC2_R8G8B8A1_SRGB_BLOCK 157 | DXGI_FORMAT_UNKNOWN, // ETC2_R8G8B8A8_UNORM_BLOCK 158 | DXGI_FORMAT_UNKNOWN, // ETC2_R8G8B8A8_SRGB_BLOCK 159 | DXGI_FORMAT_UNKNOWN, // EAC_R11_UNORM_BLOCK 160 | DXGI_FORMAT_UNKNOWN, // EAC_R11_SNORM_BLOCK 161 | DXGI_FORMAT_UNKNOWN, // EAC_R11G11_UNORM_BLOCK 162 | DXGI_FORMAT_UNKNOWN, // EAC_R11G11_SNORM_BLOCK 163 | DXGI_FORMAT_UNKNOWN, // ASTC_4x4_UNORM_BLOCK 164 | DXGI_FORMAT_UNKNOWN, // ASTC_4x4_SRGB_BLOCK 165 | DXGI_FORMAT_UNKNOWN, // ASTC_5x4_UNORM_BLOCK 166 | DXGI_FORMAT_UNKNOWN, // ASTC_5x4_SRGB_BLOCK 167 | DXGI_FORMAT_UNKNOWN, // ASTC_5x5_UNORM_BLOCK 168 | DXGI_FORMAT_UNKNOWN, // ASTC_5x5_SRGB_BLOCK 169 | DXGI_FORMAT_UNKNOWN, // ASTC_6x5_UNORM_BLOCK 170 | DXGI_FORMAT_UNKNOWN, // ASTC_6x5_SRGB_BLOCK 171 | DXGI_FORMAT_UNKNOWN, // ASTC_6x6_UNORM_BLOCK 172 | DXGI_FORMAT_UNKNOWN, // ASTC_6x6_SRGB_BLOCK 173 | DXGI_FORMAT_UNKNOWN, // ASTC_8x5_UNORM_BLOCK 174 | DXGI_FORMAT_UNKNOWN, // ASTC_8x5_SRGB_BLOCK 175 | DXGI_FORMAT_UNKNOWN, // ASTC_8x6_UNORM_BLOCK 176 | DXGI_FORMAT_UNKNOWN, // ASTC_8x6_SRGB_BLOCK 177 | DXGI_FORMAT_UNKNOWN, // ASTC_8x8_UNORM_BLOCK 178 | DXGI_FORMAT_UNKNOWN, // ASTC_8x8_SRGB_BLOCK 179 | DXGI_FORMAT_UNKNOWN, // ASTC_10x5_UNORM_BLOCK 180 | DXGI_FORMAT_UNKNOWN, // ASTC_10x5_SRGB_BLOCK 181 | DXGI_FORMAT_UNKNOWN, // ASTC_10x6_UNORM_BLOCK 182 | DXGI_FORMAT_UNKNOWN, // ASTC_10x6_SRGB_BLOCK 183 | DXGI_FORMAT_UNKNOWN, // ASTC_10x8_UNORM_BLOCK 184 | DXGI_FORMAT_UNKNOWN, // ASTC_10x8_SRGB_BLOCK 185 | DXGI_FORMAT_UNKNOWN, // ASTC_10x10_UNORM_BLOCK 186 | DXGI_FORMAT_UNKNOWN, // ASTC_10x10_SRGB_BLOCK 187 | DXGI_FORMAT_UNKNOWN, // ASTC_12x10_UNORM_BLOCK 188 | DXGI_FORMAT_UNKNOWN, // ASTC_12x10_SRGB_BLOCK 189 | DXGI_FORMAT_UNKNOWN, // ASTC_12x12_UNORM_BLOCK 190 | DXGI_FORMAT_UNKNOWN, // ASTC_12x12_SRGB_BLOCK 191 | }; 192 | 193 | std::array formats_block {{ 194 | { 0, 0, 0 }, // UNDEFINED 195 | { }, // R4G4_UNORM_PACK8 196 | { }, // R4G4B4A4_UNORM_PACK16 197 | { }, // B4G4R4A4_UNORM_PACK16 198 | { }, // R5G6B5_UNORM_PACK16 199 | { }, // B5G6R5_UNORM_PACK16 200 | { }, // R5G5B5A1_UNORM_PACK16 201 | { }, // B5G5R5A1_UNORM_PACK16 202 | { }, // A1R5G5B5_UNORM_PACK16 203 | { 1, 1, 8 }, // R8_UNORM 204 | { 1, 1, 8 }, // R8_SNORM 205 | { 1, 1, 8 }, // R8_USCALED 206 | { 1, 1, 8 }, // R8_SSCALED 207 | { 1, 1, 8 }, // R8_UINT 208 | { 1, 1, 8 }, // R8_SINT 209 | { 1, 1, 8 }, // R8_SRGB 210 | { 1, 1, 16 }, // R8G8_UNORM 211 | { 1, 1, 16 }, // R8G8_SNORM 212 | { 1, 1, 16 }, // R8G8_USCALED 213 | { 1, 1, 16 }, // R8G8_SSCALED 214 | { 1, 1, 16 }, // R8G8_UINT 215 | { 1, 1, 16 }, // R8G8_SINT 216 | { 1, 1, 16 }, // R8G8_SRGB 217 | { 1, 1, 24 }, // R8G8B8_UNORM 218 | { 1, 1, 24 }, // R8G8B8_SNORM 219 | { 1, 1, 24 }, // R8G8B8_USCALED 220 | { 1, 1, 24 }, // R8G8B8_SSCALED 221 | { 1, 1, 24 }, // R8G8B8_UINT 222 | { 1, 1, 24 }, // R8G8B8_SINT 223 | { 1, 1, 24 }, // R8G8B8_SRGB 224 | { 1, 1, 24 }, // B8G8R8_UNORM 225 | { 1, 1, 24 }, // B8G8R8_SNORM 226 | { 1, 1, 24 }, // B8G8R8_USCALED 227 | { 1, 1, 24 }, // B8G8R8_SSCALED 228 | { 1, 1, 24 }, // B8G8R8_UINT 229 | { 1, 1, 24 }, // B8G8R8_SINT 230 | { 1, 1, 24 }, // B8G8R8_SRGB 231 | { 1, 1, 32 }, // R8G8B8A8_UNORM 232 | { 1, 1, 32 }, // R8G8B8A8_SNORM 233 | { 1, 1, 32 }, // R8G8B8A8_USCALED 234 | { 1, 1, 32 }, // R8G8B8A8_SSCALED 235 | { 1, 1, 32 }, // R8G8B8A8_UINT 236 | { 1, 1, 32 }, // R8G8B8A8_SINT 237 | { 1, 1, 32 }, // R8G8B8A8_SRGB 238 | { 1, 1, 32 }, // B8G8R8A8_UNORM 239 | { 1, 1, 32 }, // B8G8R8A8_SNORM 240 | { 1, 1, 32 }, // B8G8R8A8_USCALED 241 | { 1, 1, 32 }, // B8G8R8A8_SSCALED 242 | { 1, 1, 32 }, // B8G8R8A8_UINT 243 | { 1, 1, 32 }, // B8G8R8A8_SINT 244 | { 1, 1, 32 }, // B8G8R8A8_SRGB 245 | { 1, 1, 32 }, // A8B8G8R8_UNORM_PACK32 246 | { 1, 1, 32 }, // A8B8G8R8_SNORM_PACK32 247 | { 1, 1, 32 }, // A8B8G8R8_USCALED_PACK32 248 | { 1, 1, 32 }, // A8B8G8R8_SSCALED_PACK32 249 | { 1, 1, 32 }, // A8B8G8R8_UINT_PACK32 250 | { 1, 1, 32 }, // A8B8G8R8_SINT_PACK32 251 | { 1, 1, 32 }, // A8B8G8R8_SRGB_PACK32 252 | { 1, 1, 32 }, // A2R10G10B10_UNORM_PACK32 253 | { 1, 1, 32 }, // A2R10G10B10_SNORM_PACK32 254 | { 1, 1, 32 }, // A2R10G10B10_USCALED_PACK32 255 | { 1, 1, 32 }, // A2R10G10B10_SSCALED_PACK32 256 | { 1, 1, 32 }, // A2R10G10B10_UINT_PACK32 257 | { 1, 1, 32 }, // A2R10G10B10_SINT_PACK32 258 | { 1, 1, 32 }, // A2B10G10R10_UNORM_PACK32 259 | { 1, 1, 32 }, // A2B10G10R10_SNORM_PACK32 260 | { 1, 1, 32 }, // A2B10G10R10_USCALED_PACK32 261 | { 1, 1, 32 }, // A2B10G10R10_SSCALED_PACK32 262 | { 1, 1, 32 }, // A2B10G10R10_UINT_PACK32 263 | { 1, 1, 32 }, // A2B10G10R10_SINT_PACK32 264 | { 1, 1, 16 }, // R16_UNORM 265 | { 1, 1, 16 }, // R16_SNORM 266 | { 1, 1, 16 }, // R16_USCALED 267 | { 1, 1, 16 }, // R16_SSCALED 268 | { 1, 1, 16 }, // R16_UINT 269 | { 1, 1, 16 }, // R16_SINT 270 | { 1, 1, 16 }, // R16_SFLOAT 271 | { 1, 1, 32 }, // R16G16_UNORM 272 | { 1, 1, 32 }, // R16G16_SNORM 273 | { 1, 1, 32 }, // R16G16_USCALED 274 | { 1, 1, 32 }, // R16G16_SSCALED 275 | { 1, 1, 32 }, // R16G16_UINT 276 | { 1, 1, 32 }, // R16G16_SINT 277 | { 1, 1, 32 }, // R16G16_SFLOAT 278 | { 1, 1, 48 }, // R16G16B16_UNORM 279 | { 1, 1, 48 }, // R16G16B16_SNORM 280 | { 1, 1, 48 }, // R16G16B16_USCALED 281 | { 1, 1, 48 }, // R16G16B16_SSCALED 282 | { 1, 1, 48 }, // R16G16B16_UINT 283 | { 1, 1, 48 }, // R16G16B16_SINT 284 | { 1, 1, 48 }, // R16G16B16_SFLOAT 285 | { 1, 1, 64 }, // R16G16B16A16_UNORM 286 | { 1, 1, 64 }, // R16G16B16A16_SNORM 287 | { 1, 1, 64 }, // R16G16B16A16_USCALED 288 | { 1, 1, 64 }, // R16G16B16A16_SSCALED 289 | { 1, 1, 64 }, // R16G16B16A16_UINT 290 | { 1, 1, 64 }, // R16G16B16A16_SINT 291 | { 1, 1, 64 }, // R16G16B16A16_SFLOAT 292 | { 1, 1, 32 }, // R32_UINT 293 | { 1, 1, 32 }, // R32_SINT 294 | { 1, 1, 32 }, // R32_SFLOAT 295 | { 1, 1, 64 }, // R32G32_UINT 296 | { 1, 1, 64 }, // R32G32_SINT 297 | { 1, 1, 64 }, // R32G32_SFLOAT 298 | { 1, 1, 96 }, // R32G32B32_UINT 299 | { 1, 1, 96 }, // R32G32B32_SINT 300 | { 1, 1, 96 }, // R32G32B32_SFLOAT 301 | { 1, 1, 128 }, // R32G32B32A32_UINT 302 | { 1, 1, 128 }, // R32G32B32A32_SINT 303 | { 1, 1, 128 }, // R32G32B32A32_SFLOAT 304 | { 1, 1, 64 }, // R64_UINT 305 | { 1, 1, 64 }, // R64_SINT 306 | { 1, 1, 64 }, // R64_SFLOAT 307 | { }, // R64G64_UINT 308 | { }, // R64G64_SINT 309 | { }, // R64G64_SFLOAT 310 | { }, // R64G64B64_UINT 311 | { }, // R64G64B64_SINT 312 | { }, // R64G64B64_SFLOAT 313 | { }, // R64G64B64A64_UINT 314 | { }, // R64G64B64A64_SINT 315 | { }, // R64G64B64A64_SFLOAT 316 | { }, // B10G11R11_UFLOAT_PACK32 317 | { }, // E5B9G9R9_UFLOAT_PACK32 318 | { }, // D16_UNORM 319 | { }, // X8_D24_UNORM_PACK32 320 | { 1, 1, 32 }, // D32_SFLOAT 321 | { }, // S8_UINT 322 | { }, // D16_UNORM_S8_UINT 323 | { }, // D24_UNORM_S8_UINT 324 | { }, // D32_SFLOAT_S8_UINT 325 | { 4, 4, 64 }, // BC1_RGB_UNORM_BLOCK 326 | { 4, 4, 64 }, // BC1_RGB_SRGB_BLOCK 327 | { 4, 4, 64 }, // BC1_RGBA_UNORM_BLOCK 328 | { }, // BC1_RGBA_SRGB_BLOCK 329 | { 4, 4, 128 }, // BC2_UNORM_BLOCK 330 | { 4, 4, 128 }, // BC2_SRGB_BLOCK 331 | { 4, 4, 128 }, // BC3_UNORM_BLOCK 332 | { 4, 4, 128 }, // BC3_SRGB_BLOCK 333 | { 4, 4, 64 }, // BC4_UNORM_BLOCK 334 | { 4, 4, 64 }, // BC4_SNORM_BLOCK 335 | { 4, 4, 128 }, // BC5_UNORM_BLOCK 336 | { 4, 4, 128 }, // BC5_SNORM_BLOCK 337 | { 4, 4, 128 }, // BC6H_UFLOAT_BLOCK 338 | { 4, 4, 128 }, // BC6H_SFLOAT_BLOCK 339 | { 4, 4, 128 }, // BC7_UNORM_BLOCK 340 | { 4, 4, 128 }, // BC7_SRGB_BLOCK 341 | { }, // ETC2_R8G8B8_UNORM_BLOCK 342 | { }, // ETC2_R8G8B8_SRGB_BLOCK 343 | { }, // ETC2_R8G8B8A1_UNORM_BLOCK 344 | { }, // ETC2_R8G8B8A1_SRGB_BLOCK 345 | { }, // ETC2_R8G8B8A8_UNORM_BLOCK 346 | { }, // ETC2_R8G8B8A8_SRGB_BLOCK 347 | { }, // EAC_R11_UNORM_BLOCK 348 | { }, // EAC_R11_SNORM_BLOCK 349 | { }, // EAC_R11G11_UNORM_BLOCK 350 | { }, // EAC_R11G11_SNORM_BLOCK 351 | { }, // ASTC_4x4_UNORM_BLOCK 352 | { }, // ASTC_4x4_SRGB_BLOCK 353 | { }, // ASTC_5x4_UNORM_BLOCK 354 | { }, // ASTC_5x4_SRGB_BLOCK 355 | { }, // ASTC_5x5_UNORM_BLOCK 356 | { }, // ASTC_5x5_SRGB_BLOCK 357 | { }, // ASTC_6x5_UNORM_BLOCK 358 | { }, // ASTC_6x5_SRGB_BLOCK 359 | { }, // ASTC_6x6_UNORM_BLOCK 360 | { }, // ASTC_6x6_SRGB_BLOCK 361 | { }, // ASTC_8x5_UNORM_BLOCK 362 | { }, // ASTC_8x5_SRGB_BLOCK 363 | { }, // ASTC_8x6_UNORM_BLOCK 364 | { }, // ASTC_8x6_SRGB_BLOCK 365 | { }, // ASTC_8x8_UNORM_BLOCK 366 | { }, // ASTC_8x8_SRGB_BLOCK 367 | { }, // ASTC_10x5_UNORM_BLOCK 368 | { }, // ASTC_10x5_SRGB_BLOCK 369 | { }, // ASTC_10x6_UNORM_BLOCK 370 | { }, // ASTC_10x6_SRGB_BLOCK 371 | { }, // ASTC_10x8_UNORM_BLOCK 372 | { }, // ASTC_10x8_SRGB_BLOCK 373 | { }, // ASTC_10x10_UNORM_BLOCK 374 | { }, // ASTC_10x10_SRGB_BLOCK 375 | { }, // ASTC_12x10_UNORM_BLOCK 376 | { }, // ASTC_12x10_SRGB_BLOCK 377 | { }, // ASTC_12x12_UNORM_BLOCK 378 | { }, // ASTC_12x12_SRGB_BLOCK 379 | }}; 380 | 381 | // TODO: mirroring required format support by vulkan but not what d3d12 actually supports.. 382 | std::array formats_property {{ 383 | { 0, 0, 0 }, // UNDEFINED 384 | { }, // R4G4_UNORM_PACK8 385 | { }, // R4G4B4A4_UNORM_PACK16 386 | { }, // B4G4R4A4_UNORM_PACK16 387 | { }, // R5G6B5_UNORM_PACK16 388 | { }, // B5G6R5_UNORM_PACK16 389 | { }, // R5G5B5A1_UNORM_PACK16 390 | { }, // B5G5R5A1_UNORM_PACK16 391 | { }, // A1R5G5B5_UNORM_PACK16 392 | { // R8_UNORM 393 | 0, 394 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT | 395 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT, 396 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 397 | }, 398 | { // R8_SNORM 399 | 0, 400 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 401 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 402 | }, 403 | { }, // R8_USCALED 404 | { }, // R8_SSCALED 405 | { 406 | 0, 407 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | 408 | VK_FORMAT_FEATURE_BLIT_DST_BIT, 409 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 410 | }, // R8_UINT 411 | { 412 | 0, 413 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | 414 | VK_FORMAT_FEATURE_BLIT_DST_BIT, 415 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 416 | }, // R8_SINT 417 | { }, // R8_SRGB 418 | { 419 | 0, 420 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT | 421 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT, 422 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 423 | }, // R8G8_UNORM 424 | { 425 | 0, 426 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 427 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 428 | }, // R8G8_SNORM 429 | { }, // R8G8_USCALED 430 | { }, // R8G8_SSCALED 431 | { 432 | 0, 433 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | 434 | VK_FORMAT_FEATURE_BLIT_DST_BIT, 435 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 436 | }, // R8G8_UINT 437 | { 438 | 0, 439 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | 440 | VK_FORMAT_FEATURE_BLIT_DST_BIT, 441 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 442 | }, // R8G8_SINT 443 | { }, // R8G8_SRGB 444 | { }, // R8G8B8_UNORM 445 | { }, // R8G8B8_SNORM 446 | { }, // R8G8B8_USCALED 447 | { }, // R8G8B8_SSCALED 448 | { }, // R8G8B8_UINT 449 | { }, // R8G8B8_SINT 450 | { }, // R8G8B8_SRGB 451 | { }, // B8G8R8_UNORM 452 | { }, // B8G8R8_SNORM 453 | { }, // B8G8R8_USCALED 454 | { }, // B8G8R8_SSCALED 455 | { }, // B8G8R8_UINT 456 | { }, // B8G8R8_SINT 457 | { }, // B8G8R8_SRGB 458 | { // R8G8B8A8_UNORM 459 | 0, 460 | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | 461 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT, 462 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT, 463 | }, 464 | { 465 | 0, 466 | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT, 467 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT, 468 | }, // R8G8B8A8_SNORM 469 | { }, // R8G8B8A8_USCALED 470 | { }, // R8G8B8A8_SSCALED 471 | { }, // R8G8B8A8_UINT 472 | { }, // R8G8B8A8_SINT 473 | { // R8G8B8A8_SRGB 474 | 0, 475 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT | 476 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT, 477 | 0, 478 | }, 479 | { 480 | 0, 481 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT | 482 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT, 483 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 484 | }, // B8G8R8A8_UNORM 485 | { }, // B8G8R8A8_SNORM 486 | { }, // B8G8R8A8_USCALED 487 | { }, // B8G8R8A8_SSCALED 488 | { }, // B8G8R8A8_UINT 489 | { }, // B8G8R8A8_SINT 490 | { }, // B8G8R8A8_SRGB 491 | { }, // A8B8G8R8_UNORM_PACK32 492 | { }, // A8B8G8R8_SNORM_PACK32 493 | { }, // A8B8G8R8_USCALED_PACK32 494 | { }, // A8B8G8R8_SSCALED_PACK32 495 | { }, // A8B8G8R8_UINT_PACK32 496 | { }, // A8B8G8R8_SINT_PACK32 497 | { }, // A8B8G8R8_SRGB_PACK32 498 | { }, // A2R10G10B10_UNORM_PACK32 499 | { }, // A2R10G10B10_SNORM_PACK32 500 | { }, // A2R10G10B10_USCALED_PACK32 501 | { }, // A2R10G10B10_SSCALED_PACK32 502 | { }, // A2R10G10B10_UINT_PACK32 503 | { }, // A2R10G10B10_SINT_PACK32 504 | { 505 | 0, 506 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT | 507 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT, 508 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 509 | }, // A2B10G10R10_UNORM_PACK32 510 | { }, // A2B10G10R10_SNORM_PACK32 511 | { }, // A2B10G10R10_USCALED_PACK32 512 | { }, // A2B10G10R10_SSCALED_PACK32 513 | { }, // A2B10G10R10_UINT_PACK32 514 | { }, // A2B10G10R10_SINT_PACK32 515 | { 516 | 0, 517 | 0, 518 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT, 519 | }, // R16_UNORM 520 | { }, // R16_SNORM 521 | { }, // R16_USCALED 522 | { }, // R16_SSCALED 523 | { }, // R16_UINT 524 | { }, // R16_SINT 525 | { 526 | 0, 527 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT | 528 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT, 529 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 530 | }, // R16_SFLOAT 531 | { 532 | 0, 533 | 0, 534 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT, 535 | }, // R16G16_UNORM 536 | { 537 | 0, 538 | 0, 539 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT, 540 | }, // R16G16_SNORM 541 | { }, // R16G16_USCALED 542 | { }, // R16G16_SSCALED 543 | { }, // R16G16_UINT 544 | { }, // R16G16_SINT 545 | { 546 | 0, 547 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT | 548 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT, 549 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT, 550 | }, // R16G16_SFLOAT 551 | { }, // R16G16B16_UNORM 552 | { }, // R16G16B16_SNORM 553 | { }, // R16G16B16_USCALED 554 | { }, // R16G16B16_SSCALED 555 | { }, // R16G16B16_UINT 556 | { }, // R16G16B16_SINT 557 | { }, // R16G16B16_SFLOAT 558 | { 559 | 0, 560 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT| VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT | 561 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT, // D3D12 specific 562 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT, 563 | }, // R16G16B16A16_UNORM 564 | { 565 | 0, 566 | 0, 567 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT, 568 | }, // R16G16B16A16_SNORM 569 | { }, // R16G16B16A16_USCALED 570 | { }, // R16G16B16A16_SSCALED 571 | { }, // R16G16B16A16_UINT 572 | { }, // R16G16B16A16_SINT 573 | { 574 | 0, 575 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT | 576 | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT | 577 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BLEND_BIT, 578 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT, 579 | }, // R16G16B16A16_SFLOAT 580 | { 581 | 0, 582 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | 583 | VK_FORMAT_FEATURE_STORAGE_IMAGE_ATOMIC_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT, 584 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT | 585 | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_ATOMIC_BIT, 586 | }, // R32_UINT 587 | { 588 | 0, 589 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | 590 | VK_FORMAT_FEATURE_STORAGE_IMAGE_ATOMIC_BIT | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT, 591 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT | 592 | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_ATOMIC_BIT, 593 | }, // R32_SINT 594 | { 595 | 0, 596 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | 597 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT, 598 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT, 599 | }, // R32_SFLOAT !! 600 | { 601 | 0, 602 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | 603 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT, 604 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT, 605 | }, // R32G32_UINT 606 | { 607 | 0, 608 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | 609 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT, 610 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT, 611 | }, // R32G32_SINT 612 | { 613 | 0, 614 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | 615 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT, 616 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT, 617 | }, // R32G32_SFLOAT 618 | { }, // R32G32B32_UINT 619 | { }, // R32G32B32_SINT 620 | { }, // R32G32B32_SFLOAT 621 | { 622 | 0, 623 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | 624 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT, 625 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT, 626 | }, // R32G32B32A32_UINT 627 | { 628 | 0, 629 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | 630 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT, 631 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT, 632 | }, // R32G32B32A32_SINT 633 | { 634 | 0, 635 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT | 636 | VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT | VK_FORMAT_FEATURE_BLIT_DST_BIT, 637 | VK_FORMAT_FEATURE_VERTEX_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT, 638 | }, // R32G32B32A32_SFLOAT 639 | { }, // R64_UINT 640 | { }, // R64_SINT 641 | { }, // R64_SFLOAT 642 | { }, // R64G64_UINT 643 | { }, // R64G64_SINT 644 | { }, // R64G64_SFLOAT 645 | { }, // R64G64B64_UINT 646 | { }, // R64G64B64_SINT 647 | { }, // R64G64B64_SFLOAT 648 | { }, // R64G64B64A64_UINT 649 | { }, // R64G64B64A64_SINT 650 | { }, // R64G64B64A64_SFLOAT 651 | { }, // B10G11R11_UFLOAT_PACK32 652 | { }, // E5B9G9R9_UFLOAT_PACK32 653 | { 654 | 0, 655 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT, 656 | 0, 657 | }, // D16_UNORM 658 | { }, // X8_D24_UNORM_PACK32 659 | { // D32_SFLOAT // TODO 660 | 0, 661 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT, 662 | 0, 663 | }, 664 | { }, // S8_UINT 665 | { 666 | 0, 667 | 0, 668 | 0, 669 | }, // D16_UNORM_S8_UINT 670 | { 671 | 0, 672 | VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT, 673 | 0, 674 | }, // D24_UNORM_S8_UINT 675 | { 676 | 0, 677 | VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT, 678 | 0, 679 | }, // D32_SFLOAT_S8_UINT 680 | { 681 | 0, 682 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 683 | 0, 684 | }, // BC1_RGB_UNORM_BLOCK 685 | { 686 | 0, 687 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 688 | 0, 689 | }, // BC1_RGB_SRGB_BLOCK 690 | { 691 | 0, 692 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 693 | 0, 694 | }, // BC1_RGBA_UNORM_BLOCK 695 | { 696 | 0, 697 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 698 | 0, 699 | }, // BC1_RGBA_SRGB_BLOCK 700 | { 701 | 0, 702 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 703 | 0, 704 | }, // BC2_UNORM_BLOCK 705 | { 706 | 0, 707 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 708 | 0, 709 | }, // BC2_SRGB_BLOCK 710 | { 711 | 0, 712 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 713 | 0, 714 | }, // BC3_UNORM_BLOCK 715 | { 716 | 0, 717 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 718 | 0, 719 | }, // BC3_SRGB_BLOCK 720 | { 721 | 0, 722 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 723 | 0, 724 | }, // BC4_UNORM_BLOCK 725 | { 726 | 0, 727 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 728 | 0, 729 | }, // BC4_SNORM_BLOCK 730 | { 731 | 0, 732 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 733 | 0, 734 | }, // BC5_UNORM_BLOCK 735 | { 736 | 0, 737 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 738 | 0, 739 | }, // BC5_SNORM_BLOCK 740 | { 741 | 0, 742 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 743 | 0, 744 | }, // BC6H_UFLOAT_BLOCK 745 | { 746 | 0, 747 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 748 | 0, 749 | }, // BC6H_SFLOAT_BLOCK 750 | { 751 | 0, 752 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 753 | 0, 754 | }, // BC7_UNORM_BLOCK 755 | { 756 | 0, 757 | VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_BLIT_SRC_BIT | VK_FORMAT_FEATURE_SAMPLED_IMAGE_FILTER_LINEAR_BIT, 758 | 0, 759 | }, // BC7_SRGB_BLOCK 760 | { }, // ETC2_R8G8B8_UNORM_BLOCK 761 | { }, // ETC2_R8G8B8_SRGB_BLOCK 762 | { }, // ETC2_R8G8B8A1_UNORM_BLOCK 763 | { }, // ETC2_R8G8B8A1_SRGB_BLOCK 764 | { }, // ETC2_R8G8B8A8_UNORM_BLOCK 765 | { }, // ETC2_R8G8B8A8_SRGB_BLOCK 766 | { }, // EAC_R11_UNORM_BLOCK 767 | { }, // EAC_R11_SNORM_BLOCK 768 | { }, // EAC_R11G11_UNORM_BLOCK 769 | { }, // EAC_R11G11_SNORM_BLOCK 770 | { }, // ASTC_4x4_UNORM_BLOCK 771 | { }, // ASTC_4x4_SRGB_BLOCK 772 | { }, // ASTC_5x4_UNORM_BLOCK 773 | { }, // ASTC_5x4_SRGB_BLOCK 774 | { }, // ASTC_5x5_UNORM_BLOCK 775 | { }, // ASTC_5x5_SRGB_BLOCK 776 | { }, // ASTC_6x5_UNORM_BLOCK 777 | { }, // ASTC_6x5_SRGB_BLOCK 778 | { }, // ASTC_6x6_UNORM_BLOCK 779 | { }, // ASTC_6x6_SRGB_BLOCK 780 | { }, // ASTC_8x5_UNORM_BLOCK 781 | { }, // ASTC_8x5_SRGB_BLOCK 782 | { }, // ASTC_8x6_UNORM_BLOCK 783 | { }, // ASTC_8x6_SRGB_BLOCK 784 | { }, // ASTC_8x8_UNORM_BLOCK 785 | { }, // ASTC_8x8_SRGB_BLOCK 786 | { }, // ASTC_10x5_UNORM_BLOCK 787 | { }, // ASTC_10x5_SRGB_BLOCK 788 | { }, // ASTC_10x6_UNORM_BLOCK 789 | { }, // ASTC_10x6_SRGB_BLOCK 790 | { }, // ASTC_10x8_UNORM_BLOCK 791 | { }, // ASTC_10x8_SRGB_BLOCK 792 | { }, // ASTC_10x10_UNORM_BLOCK 793 | { }, // ASTC_10x10_SRGB_BLOCK 794 | { }, // ASTC_12x10_UNORM_BLOCK 795 | { }, // ASTC_12x10_SRGB_BLOCK 796 | { }, // ASTC_12x12_UNORM_BLOCK 797 | { }, // ASTC_12x12_SRGB_BLOCK 798 | }}; 799 | -------------------------------------------------------------------------------- /src/icd.hpp: -------------------------------------------------------------------------------- 1 | #pragma once 2 | 3 | #define NOMINMAX 4 | 5 | #include 6 | 7 | #define VK_USE_PLATFORM_WIN32_KHR 8 | #define VK_USE_PLATFORM_UWP_RKZ 9 | #include 10 | 11 | #include 12 | 13 | #include 14 | #include 15 | 16 | static auto init_logging() { 17 | static auto initialized { false }; 18 | if (initialized) { 19 | return; 20 | } 21 | 22 | initialized = true; 23 | // spdlog::basic_logger_mt("rostkatze", "test.log"); 24 | spdlog::create("rostkatze"); 25 | spdlog::set_level(spdlog::level::trace); 26 | } 27 | 28 | static auto log() { 29 | return spdlog::get("rostkatze"); 30 | } 31 | 32 | #define TRACE(...) log()->trace(__VA_ARGS__) 33 | #define DEBUG(...) log()->debug(__VA_ARGS__) 34 | #define INFO(...) log()->info(__VA_ARGS__) 35 | #define WARN(...) log()->warn(__VA_ARGS__) 36 | #define ERR(...) log()->error(__VA_ARGS__) 37 | 38 | struct format_block_t { 39 | uint8_t width; 40 | uint8_t height; 41 | uint16_t bits; 42 | }; 43 | 44 | extern std::array formats; 45 | extern std::array formats_block; 46 | extern std::array formats_property; 47 | 48 | #define VK_FNC(name) \ 49 | if (!strcmp(pName, #name)) { \ 50 | return reinterpret_cast(&name); \ 51 | } 52 | 53 | #define VK_INSTANCE_FNC() \ 54 | VK_FNC(vkCreateInstance) \ 55 | VK_FNC(vkDestroyInstance) \ 56 | VK_FNC(vkEnumerateInstanceExtensionProperties) \ 57 | VK_FNC(vkEnumerateDeviceExtensionProperties) \ 58 | VK_FNC(vkGetDeviceProcAddr) \ 59 | VK_FNC(vkEnumeratePhysicalDevices) \ 60 | VK_FNC(vkCreateDevice) \ 61 | VK_FNC(vkCreateWin32SurfaceKHR) \ 62 | VK_FNC(vkCreateUWPSurfaceRKZ) 63 | 64 | #define VK_PHYSICAL_DEVICE_FNC() \ 65 | VK_FNC(vkGetPhysicalDeviceFeatures) \ 66 | VK_FNC(vkGetPhysicalDeviceFormatProperties) \ 67 | VK_FNC(vkGetPhysicalDeviceImageFormatProperties) \ 68 | VK_FNC(vkGetPhysicalDeviceProperties) \ 69 | VK_FNC(vkGetPhysicalDeviceMemoryProperties) \ 70 | VK_FNC(vkGetPhysicalDeviceQueueFamilyProperties) \ 71 | VK_FNC(vkGetPhysicalDeviceSparseImageFormatProperties) \ 72 | VK_FNC(vkGetPhysicalDeviceSurfaceSupportKHR) \ 73 | VK_FNC(vkGetPhysicalDeviceSurfaceCapabilitiesKHR) \ 74 | VK_FNC(vkGetPhysicalDeviceSurfaceFormatsKHR) \ 75 | VK_FNC(vkGetPhysicalDeviceSurfacePresentModesKHR) \ 76 | VK_FNC(vkGetPhysicalDeviceFeatures2KHR) \ 77 | VK_FNC(vkGetPhysicalDeviceProperties2KHR) \ 78 | VK_FNC(vkGetPhysicalDeviceFormatProperties2KHR) \ 79 | VK_FNC(vkGetPhysicalDeviceImageFormatProperties2KHR) \ 80 | VK_FNC(vkGetPhysicalDeviceQueueFamilyProperties2KHR) \ 81 | VK_FNC(vkGetPhysicalDeviceMemoryProperties2KHR) \ 82 | VK_FNC(vkGetPhysicalDeviceSparseImageFormatProperties2KHR) \ 83 | VK_FNC(vkGetPhysicalDeviceUWPPresentationSupportRKZ) 84 | 85 | #define VK_DEVICE_FNC() \ 86 | VK_FNC(vkDestroyDevice) \ 87 | VK_FNC(vkGetDeviceQueue) \ 88 | VK_FNC(vkQueueSubmit) \ 89 | VK_FNC(vkQueueWaitIdle) \ 90 | VK_FNC(vkDeviceWaitIdle) \ 91 | VK_FNC(vkAllocateMemory) \ 92 | VK_FNC(vkFreeMemory) \ 93 | VK_FNC(vkMapMemory) \ 94 | VK_FNC(vkUnmapMemory) \ 95 | VK_FNC(vkFlushMappedMemoryRanges) \ 96 | VK_FNC(vkInvalidateMappedMemoryRanges) \ 97 | VK_FNC(vkGetDeviceMemoryCommitment) \ 98 | VK_FNC(vkBindBufferMemory) \ 99 | VK_FNC(vkBindImageMemory) \ 100 | VK_FNC(vkGetBufferMemoryRequirements) \ 101 | VK_FNC(vkGetImageMemoryRequirements) \ 102 | VK_FNC(vkGetImageSparseMemoryRequirements) \ 103 | VK_FNC(vkQueueBindSparse) \ 104 | VK_FNC(vkCreateFence) \ 105 | VK_FNC(vkDestroyFence) \ 106 | VK_FNC(vkResetFences) \ 107 | VK_FNC(vkGetFenceStatus) \ 108 | VK_FNC(vkWaitForFences) \ 109 | VK_FNC(vkCreateSemaphore) \ 110 | VK_FNC(vkDestroySemaphore) \ 111 | VK_FNC(vkCreateEvent) \ 112 | VK_FNC(vkDestroyEvent) \ 113 | VK_FNC(vkGetEventStatus) \ 114 | VK_FNC(vkSetEvent) \ 115 | VK_FNC(vkResetEvent) \ 116 | VK_FNC(vkCreateQueryPool) \ 117 | VK_FNC(vkDestroyQueryPool) \ 118 | VK_FNC(vkGetQueryPoolResults) \ 119 | VK_FNC(vkCreateBuffer) \ 120 | VK_FNC(vkDestroyBuffer) \ 121 | VK_FNC(vkCreateBufferView) \ 122 | VK_FNC(vkDestroyBufferView) \ 123 | VK_FNC(vkCreateImage) \ 124 | VK_FNC(vkDestroyImage) \ 125 | VK_FNC(vkGetImageSubresourceLayout) \ 126 | VK_FNC(vkCreateImageView) \ 127 | VK_FNC(vkDestroyImageView) \ 128 | VK_FNC(vkCreateShaderModule) \ 129 | VK_FNC(vkDestroyShaderModule) \ 130 | VK_FNC(vkCreatePipelineCache) \ 131 | VK_FNC(vkDestroyPipelineCache) \ 132 | VK_FNC(vkGetPipelineCacheData) \ 133 | VK_FNC(vkMergePipelineCaches) \ 134 | VK_FNC(vkCreateGraphicsPipelines) \ 135 | VK_FNC(vkCreateComputePipelines) \ 136 | VK_FNC(vkDestroyPipeline) \ 137 | VK_FNC(vkCreatePipelineLayout) \ 138 | VK_FNC(vkDestroyPipelineLayout) \ 139 | VK_FNC(vkCreateSampler) \ 140 | VK_FNC(vkDestroySampler) \ 141 | VK_FNC(vkCreateDescriptorSetLayout) \ 142 | VK_FNC(vkDestroyDescriptorSetLayout) \ 143 | VK_FNC(vkCreateDescriptorPool) \ 144 | VK_FNC(vkDestroyDescriptorPool) \ 145 | VK_FNC(vkResetDescriptorPool) \ 146 | VK_FNC(vkAllocateDescriptorSets) \ 147 | VK_FNC(vkFreeDescriptorSets) \ 148 | VK_FNC(vkUpdateDescriptorSets) \ 149 | VK_FNC(vkCreateFramebuffer) \ 150 | VK_FNC(vkDestroyFramebuffer) \ 151 | VK_FNC(vkCreateRenderPass) \ 152 | VK_FNC(vkDestroyRenderPass) \ 153 | VK_FNC(vkGetRenderAreaGranularity) \ 154 | VK_FNC(vkCreateCommandPool) \ 155 | VK_FNC(vkDestroyCommandPool) \ 156 | VK_FNC(vkResetCommandPool) \ 157 | VK_FNC(vkAllocateCommandBuffers) \ 158 | VK_FNC(vkFreeCommandBuffers) \ 159 | VK_FNC(vkCreateSwapchainKHR) \ 160 | VK_FNC(vkDestroySwapchainKHR) \ 161 | VK_FNC(vkGetSwapchainImagesKHR) \ 162 | VK_FNC(vkAcquireNextImageKHR) \ 163 | VK_FNC(vkQueuePresentKHR) \ 164 | VK_FNC(vkBeginCommandBuffer) \ 165 | VK_FNC(vkEndCommandBuffer) \ 166 | VK_FNC(vkResetCommandBuffer) \ 167 | VK_FNC(vkCmdBindPipeline) \ 168 | VK_FNC(vkCmdSetViewport) \ 169 | VK_FNC(vkCmdSetScissor) \ 170 | VK_FNC(vkCmdSetLineWidth) \ 171 | VK_FNC(vkCmdSetDepthBias) \ 172 | VK_FNC(vkCmdSetBlendConstants) \ 173 | VK_FNC(vkCmdSetDepthBounds) \ 174 | VK_FNC(vkCmdSetStencilCompareMask) \ 175 | VK_FNC(vkCmdSetStencilWriteMask) \ 176 | VK_FNC(vkCmdSetStencilReference) \ 177 | VK_FNC(vkCmdBindDescriptorSets) \ 178 | VK_FNC(vkCmdBindIndexBuffer) \ 179 | VK_FNC(vkCmdBindVertexBuffers) \ 180 | VK_FNC(vkCmdDraw) \ 181 | VK_FNC(vkCmdDrawIndexed) \ 182 | VK_FNC(vkCmdDrawIndirect) \ 183 | VK_FNC(vkCmdDrawIndexedIndirect) \ 184 | VK_FNC(vkCmdDispatch) \ 185 | VK_FNC(vkCmdDispatchIndirect) \ 186 | VK_FNC(vkCmdCopyBuffer) \ 187 | VK_FNC(vkCmdCopyImage) \ 188 | VK_FNC(vkCmdBlitImage) \ 189 | VK_FNC(vkCmdCopyBufferToImage) \ 190 | VK_FNC(vkCmdCopyImageToBuffer) \ 191 | VK_FNC(vkCmdUpdateBuffer) \ 192 | VK_FNC(vkCmdFillBuffer) \ 193 | VK_FNC(vkCmdClearColorImage) \ 194 | VK_FNC(vkCmdClearDepthStencilImage) \ 195 | VK_FNC(vkCmdClearAttachments) \ 196 | VK_FNC(vkCmdResolveImage) \ 197 | VK_FNC(vkCmdSetEvent) \ 198 | VK_FNC(vkCmdResetEvent) \ 199 | VK_FNC(vkCmdWaitEvents) \ 200 | VK_FNC(vkCmdPipelineBarrier) \ 201 | VK_FNC(vkCmdBeginQuery) \ 202 | VK_FNC(vkCmdEndQuery) \ 203 | VK_FNC(vkCmdResetQueryPool) \ 204 | VK_FNC(vkCmdWriteTimestamp) \ 205 | VK_FNC(vkCmdCopyQueryPoolResults) \ 206 | VK_FNC(vkCmdPushConstants) \ 207 | VK_FNC(vkCmdBeginRenderPass) \ 208 | VK_FNC(vkCmdNextSubpass) \ 209 | VK_FNC(vkCmdEndRenderPass) \ 210 | VK_FNC(vkCmdExecuteCommands) 211 | -------------------------------------------------------------------------------- /src/impl.hpp: -------------------------------------------------------------------------------- 1 | #pragma once 2 | 3 | #include "icd.hpp" 4 | #include "descriptors_cpu.hpp" 5 | #include "descriptors_gpu.hpp" 6 | 7 | #include 8 | #include 9 | #include 10 | #include 11 | #include 12 | #include 13 | #include 14 | 15 | #include 16 | 17 | #include 18 | #include 19 | 20 | #include 21 | #include 22 | #include 23 | #include 24 | 25 | #include 26 | 27 | using namespace Microsoft::WRL; 28 | using namespace gsl; 29 | using namespace stdx; 30 | 31 | // Forward declerations 32 | class instance_t; 33 | class device_t; 34 | class queue_t; 35 | 36 | struct physical_device_t; 37 | struct semaphore_t; 38 | struct surface_t; 39 | struct device_memory_t; 40 | struct image_t; 41 | struct render_pass_t; 42 | struct framebuffer_t; 43 | struct image_view_t; 44 | 45 | static const size_t MAX_VERTEX_BUFFER_SLOTS = 16; 46 | 47 | // Separation of the 64 root signature entries: 48 | // 49 | // * 32 for push constants 50 | // * 8 for dynamic uniform buffer offsets 51 | // * 4 for dynamic storage buffer offsets 52 | // * 8 for cbv/srv/uav descriptor tables 53 | // * 8 for sampler descriptor tables 54 | static const size_t MAX_PUSH_CONSTANTS_SIZE = 128; // 4 * 32 fields 55 | static const size_t MAX_UNIFORM_BUFFERS_DYNAMIC = 8; 56 | static const size_t MAX_STORAGE_BUFFERS_DYNAMIC = 4; 57 | static const size_t MAX_BUFFERS_DYNAMIC = MAX_UNIFORM_BUFFERS_DYNAMIC + MAX_STORAGE_BUFFERS_DYNAMIC; 58 | 59 | enum queue_family { 60 | QUEUE_FAMILY_GENERAL_PRESENT = 0, 61 | QUEUE_FAMILY_GENERAL, 62 | QUEUE_FAMILY_COMPUTE, 63 | QUEUE_FAMILY_COPY, 64 | }; 65 | 66 | 67 | static auto init_debug_interface() -> void { 68 | static auto initialized { false }; 69 | if (initialized) { 70 | return; 71 | } 72 | 73 | initialized = true; 74 | ComPtr debug_controller {}; 75 | auto hr = ::D3D12GetDebugInterface(IID_PPV_ARGS(&debug_controller)); 76 | if (SUCCEEDED(hr)) { 77 | debug_controller->EnableDebugLayer(); 78 | } 79 | } 80 | 81 | struct heap_properties_t { 82 | D3D12_CPU_PAGE_PROPERTY page_property; 83 | D3D12_MEMORY_POOL memory_pool; 84 | }; 85 | 86 | // https://msdn.microsoft.com/de-de/library/windows/desktop/dn788678(v=vs.85).aspx 87 | static const std::array HEAPS_NUMA = {{ 88 | // DEFAULT 89 | { D3D12_CPU_PAGE_PROPERTY_NOT_AVAILABLE, D3D12_MEMORY_POOL_L1 }, 90 | // UPLOAD 91 | { D3D12_CPU_PAGE_PROPERTY_WRITE_COMBINE, D3D12_MEMORY_POOL_L0 }, 92 | // READBACK 93 | { D3D12_CPU_PAGE_PROPERTY_WRITE_BACK, D3D12_MEMORY_POOL_L0 } 94 | }}; 95 | 96 | static const std::array HEAPS_UMA = {{ 97 | // DEFAULT 98 | { D3D12_CPU_PAGE_PROPERTY_NOT_AVAILABLE, D3D12_MEMORY_POOL_L0 }, 99 | // UPLOAD 100 | { D3D12_CPU_PAGE_PROPERTY_WRITE_COMBINE, D3D12_MEMORY_POOL_L0 }, 101 | // READBACK 102 | { D3D12_CPU_PAGE_PROPERTY_WRITE_BACK, D3D12_MEMORY_POOL_L0 } 103 | }}; 104 | 105 | static const std::array HEAPS_CCUMA = {{ 106 | // DEFAULT 107 | { D3D12_CPU_PAGE_PROPERTY_NOT_AVAILABLE, D3D12_MEMORY_POOL_L0 }, 108 | // UPLOAD 109 | { D3D12_CPU_PAGE_PROPERTY_WRITE_BACK, D3D12_MEMORY_POOL_L0 }, 110 | //READBACK 111 | { D3D12_CPU_PAGE_PROPERTY_WRITE_BACK, D3D12_MEMORY_POOL_L0 } 112 | }}; 113 | 114 | struct physical_device_properties_t { 115 | uint32_t api_version; 116 | uint32_t driver_version; 117 | uint32_t vendor_id; 118 | uint32_t device_id; 119 | VkPhysicalDeviceType device_type; 120 | std::array device_name; 121 | }; 122 | 123 | struct physical_device_t { 124 | ComPtr adapter; 125 | 126 | physical_device_properties_t properties; 127 | std::optional conservative_properties; 128 | VkPhysicalDeviceMemoryProperties memory_properties; 129 | span heap_properties; 130 | 131 | VkPhysicalDeviceLimits limits; 132 | }; 133 | 134 | class instance_t final { 135 | public: 136 | instance_t() { 137 | init_logging(); 138 | init_debug_interface(); 139 | 140 | DEBUG("creating instance"); 141 | 142 | // Create dxgi factory 143 | { 144 | IDXGIFactory5 *factory; 145 | auto hr { ::CreateDXGIFactory2( 146 | DXGI_CREATE_FACTORY_DEBUG, 147 | IID_PPV_ARGS(&factory) 148 | )}; 149 | 150 | this->dxgi_factory = factory; 151 | } 152 | 153 | struct adapter_info_t { 154 | ComPtr adapter; 155 | physical_device_properties_t properties; 156 | std::optional conservative_properties; 157 | VkPhysicalDeviceMemoryProperties memory_properties; 158 | span heap_properties; 159 | VkPhysicalDeviceLimits limits; 160 | }; 161 | 162 | // 163 | std::vector adapters; 164 | auto i { 0 }; 165 | while (true) { 166 | // Query adapter 167 | ComPtr dxgi_adapter {}; 168 | if (this->dxgi_factory->EnumAdapters1(i, &dxgi_adapter) == DXGI_ERROR_NOT_FOUND) { 169 | break; 170 | } 171 | i += 1; 172 | 173 | ComPtr adapter; 174 | if (FAILED(dxgi_adapter.As(&adapter))) { 175 | ERR("Couldn't convert adapter to `IDXGIAdapter4`!"); 176 | continue; 177 | } 178 | 179 | DXGI_ADAPTER_DESC3 adapter_desc; 180 | adapter->GetDesc3(&adapter_desc); 181 | 182 | physical_device_properties_t properties { 183 | VK_MAKE_VERSION(1, 1, 0), 184 | 0, 185 | adapter_desc.VendorId, 186 | adapter_desc.DeviceId, 187 | VK_PHYSICAL_DEVICE_TYPE_OTHER, // TODO 188 | }; 189 | static_assert(VK_MAX_PHYSICAL_DEVICE_NAME_SIZE >= sizeof(WCHAR) * 128); 190 | WideCharToMultiByte( 191 | CP_UTF8, 192 | WC_NO_BEST_FIT_CHARS, 193 | adapter_desc.Description, 194 | 128, 195 | properties.device_name.data(), 196 | VK_MAX_PHYSICAL_DEVICE_NAME_SIZE, 197 | nullptr, 198 | nullptr 199 | ); 200 | 201 | // temporary device 202 | ComPtr device; 203 | { 204 | const auto hr { ::D3D12CreateDevice( 205 | adapter.Get(), 206 | D3D_FEATURE_LEVEL_11_0, 207 | IID_PPV_ARGS(&device) 208 | )}; 209 | 210 | if (FAILED(hr)) { 211 | // Doesn't support dx12 212 | continue; 213 | } 214 | } 215 | 216 | D3D12_FEATURE_DATA_D3D12_OPTIONS feature_options { 0 }; 217 | { 218 | const auto hr { 219 | device->CheckFeatureSupport( 220 | D3D12_FEATURE_D3D12_OPTIONS, 221 | &feature_options, 222 | sizeof(feature_options) 223 | ) 224 | }; 225 | // TODO: error handling 226 | } 227 | std::optional conservative_properties { std::nullopt }; 228 | if (feature_options.ConservativeRasterizationTier) { 229 | VkPhysicalDeviceConservativeRasterizationPropertiesEXT conservative; 230 | conservative.maxExtraPrimitiveOverestimationSize = 0.0; 231 | conservative.conservativePointAndLineRasterization = false; 232 | conservative.degenerateLinesRasterized = false; // not supported anyways? 233 | conservative.conservativeRasterizationPostDepthCoverage = false; // TODO: check again 234 | 235 | switch (feature_options.ConservativeRasterizationTier) { 236 | case D3D12_CONSERVATIVE_RASTERIZATION_TIER_1: { 237 | conservative.primitiveOverestimationSize = 0.5; 238 | conservative.primitiveUnderestimation = false; 239 | conservative.degenerateTrianglesRasterized = false; 240 | conservative.fullyCoveredFragmentShaderInputVariable = false; 241 | 242 | } break; 243 | case D3D12_CONSERVATIVE_RASTERIZATION_TIER_2: { 244 | conservative.primitiveOverestimationSize = 1.0 / 256.0f; 245 | conservative.primitiveUnderestimation = false; 246 | conservative.degenerateTrianglesRasterized = true; 247 | conservative.fullyCoveredFragmentShaderInputVariable = false; 248 | } break; 249 | case D3D12_CONSERVATIVE_RASTERIZATION_TIER_3: { 250 | conservative.primitiveOverestimationSize = 1.0 / 256.0f; 251 | conservative.primitiveUnderestimation = true; 252 | conservative.degenerateTrianglesRasterized = true; 253 | conservative.fullyCoveredFragmentShaderInputVariable = true; // TODO: SPIRV-Cross support 254 | } break; 255 | } 256 | conservative_properties = conservative; 257 | } 258 | 259 | 260 | D3D12_FEATURE_DATA_ARCHITECTURE feature_architecture { 0 }; 261 | { 262 | const auto hr { 263 | device->CheckFeatureSupport( 264 | D3D12_FEATURE_ARCHITECTURE, 265 | &feature_architecture, 266 | sizeof(feature_architecture) 267 | ) 268 | }; 269 | if (FAILED(hr)) { 270 | ERR("Couldn't query feature architcture data"); 271 | // TODO 272 | } 273 | } 274 | const auto uma { feature_architecture.UMA }; 275 | const auto cc_uma { feature_architecture.CacheCoherentUMA }; 276 | 277 | VkPhysicalDeviceMemoryProperties memory_properties; 278 | 279 | span heap_properties; 280 | 281 | // Memory types 282 | // TODO: HEAP TIER 1 283 | memory_properties.memoryTypeCount = 3; 284 | if (uma && cc_uma) { 285 | heap_properties = span{ HEAPS_CCUMA }; 286 | 287 | // Default 288 | memory_properties.memoryTypes[0] = { 289 | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, 290 | 0 291 | }; 292 | // Upload 293 | memory_properties.memoryTypes[1] = { 294 | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | 295 | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT, 296 | 0 297 | }; 298 | // Readback 299 | memory_properties.memoryTypes[2] = { 300 | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | 301 | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT, 302 | 0 303 | }; 304 | } else if (uma && !cc_uma) { 305 | heap_properties = span{ HEAPS_UMA }; 306 | 307 | // Default 308 | memory_properties.memoryTypes[0] = { 309 | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, 310 | 0 311 | }; 312 | // Upload 313 | memory_properties.memoryTypes[1] = { 314 | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT, 315 | 0 316 | }; 317 | // Readback 318 | memory_properties.memoryTypes[2] = { 319 | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT, 320 | 0 321 | }; 322 | } else { // NUMA 323 | heap_properties = span{ HEAPS_NUMA }; 324 | 325 | // Default 326 | memory_properties.memoryTypes[0] = { 327 | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, 328 | 0 329 | }; 330 | // Upload 331 | memory_properties.memoryTypes[1] = { 332 | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT, 333 | 1 334 | }; 335 | // Readback 336 | memory_properties.memoryTypes[2] = { 337 | VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT, 338 | 1 339 | }; 340 | } 341 | 342 | // Memory heaps 343 | auto query_memory = [&] (DXGI_MEMORY_SEGMENT_GROUP segment) { 344 | DXGI_QUERY_VIDEO_MEMORY_INFO memory_info; 345 | const auto hr { adapter->QueryVideoMemoryInfo(0, segment, &memory_info) }; 346 | // TODO: error handling 347 | return memory_info.Budget; 348 | }; 349 | 350 | const auto memory_local_size { query_memory(DXGI_MEMORY_SEGMENT_GROUP_LOCAL) }; 351 | memory_properties.memoryHeaps[0] = { memory_local_size, VK_MEMORY_HEAP_DEVICE_LOCAL_BIT }; 352 | 353 | if (!uma) { 354 | memory_properties.memoryHeapCount = 2; 355 | memory_properties.memoryHeaps[1] = { query_memory(DXGI_MEMORY_SEGMENT_GROUP_NON_LOCAL), 0 }; 356 | } else { 357 | memory_properties.memoryHeapCount = 1; 358 | } 359 | 360 | // Limits 361 | VkPhysicalDeviceLimits limits { 0 }; 362 | limits.maxImageDimension1D = D3D12_REQ_TEXTURE1D_U_DIMENSION; 363 | limits.maxImageDimension2D = D3D12_REQ_TEXTURE2D_U_OR_V_DIMENSION; 364 | limits.maxImageDimension3D = D3D12_REQ_TEXTURE3D_U_V_OR_W_DIMENSION; 365 | limits.maxImageDimensionCube = D3D12_REQ_TEXTURECUBE_DIMENSION; 366 | // TODO: missing fields 367 | limits.maxPushConstantsSize = MAX_PUSH_CONSTANTS_SIZE; 368 | // TODO: missing fields 369 | limits.maxComputeSharedMemorySize = D3D12_CS_THREAD_LOCAL_TEMP_REGISTER_POOL; 370 | // TODO: missing fields 371 | limits.maxSamplerAnisotropy = 16; 372 | // TODO: missing fields 373 | limits.minUniformBufferOffsetAlignment = 256; 374 | // TODO: missing fields 375 | limits.framebufferColorSampleCounts = VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT; // TODO 376 | limits.framebufferDepthSampleCounts = VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT; // TODO 377 | limits.framebufferStencilSampleCounts = VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT; // TODO 378 | limits.framebufferNoAttachmentsSampleCounts = VK_SAMPLE_COUNT_1_BIT | VK_SAMPLE_COUNT_4_BIT; // TODO 379 | 380 | adapters.emplace_back( 381 | adapter_info_t { 382 | adapter, 383 | properties, 384 | conservative_properties, 385 | memory_properties, 386 | heap_properties, 387 | limits, 388 | } 389 | ); 390 | } 391 | 392 | const auto num_adapters { adapters.size() }; 393 | this->_adapters = std::vector(num_adapters); 394 | for (auto i : range(num_adapters)) { 395 | auto& adapter { this->_adapters[i] }; 396 | adapter.adapter = adapters[i].adapter; 397 | adapter.properties = adapters[i].properties; 398 | adapter.conservative_properties = adapters[i].conservative_properties; 399 | adapter.memory_properties = adapters[i].memory_properties; 400 | adapter.heap_properties = adapters[i].heap_properties; 401 | adapter.limits = adapters[i].limits; 402 | } 403 | } 404 | 405 | ~instance_t() { 406 | 407 | } 408 | 409 | auto adapters() const -> span { 410 | return make_span(this->_adapters); 411 | } 412 | 413 | public: 414 | ComPtr dxgi_factory; 415 | std::vector _adapters; 416 | }; 417 | 418 | class queue_t final { 419 | public: 420 | queue_t() : queue_t { nullptr, nullptr } { } 421 | queue_t(queue_t && tmp) : 422 | loader_magic { ICD_LOADER_MAGIC }, 423 | queue { tmp.queue }, 424 | idle_fence { tmp.idle_fence }, 425 | idle_event { tmp.idle_event } 426 | { 427 | tmp.queue = nullptr; 428 | tmp.idle_fence = nullptr; 429 | tmp.idle_event = 0; 430 | } 431 | 432 | queue_t(queue_t const&) = delete; 433 | 434 | queue_t(ComPtr queue, ComPtr fence) : 435 | loader_magic { ICD_LOADER_MAGIC }, 436 | queue { queue }, 437 | idle_fence { fence }, 438 | idle_event { ::CreateEvent(nullptr, TRUE, FALSE, nullptr) } // Event with manual reset 439 | { } 440 | 441 | ~queue_t() { 442 | if (this->idle_event) { 443 | ::CloseHandle(this->idle_event); 444 | } 445 | } 446 | 447 | auto operator= (queue_t const&) -> queue_t& = delete; 448 | auto operator= (queue_t && tmp) -> queue_t& { 449 | this->queue = std::move(tmp.queue); 450 | this->idle_fence = std::move(tmp.idle_fence); 451 | this->idle_event = std::move(tmp.idle_event); 452 | 453 | tmp.queue = nullptr; 454 | tmp.idle_fence = nullptr; 455 | tmp.idle_event = 0; 456 | 457 | return *this; 458 | } 459 | 460 | auto operator->() -> ID3D12CommandQueue* { 461 | return this->queue.Get(); 462 | } 463 | 464 | private: 465 | /// Dispatchable object 466 | uintptr_t loader_magic; 467 | 468 | public: 469 | ComPtr queue; 470 | ComPtr idle_fence; 471 | HANDLE idle_event; 472 | }; 473 | 474 | struct command_pool_t { 475 | D3D12_COMMAND_LIST_TYPE type; 476 | ComPtr allocator; 477 | }; 478 | 479 | struct image_t { 480 | ComPtr resource; 481 | D3D12_RESOURCE_ALLOCATION_INFO allocation_info; 482 | D3D12_RESOURCE_DESC resource_desc; 483 | format_block_t block_data; 484 | VkImageUsageFlags usage; 485 | }; 486 | 487 | // TODO: the whole structure could be stripped a bit 488 | // and maybe precompute some more information on creation. 489 | struct render_pass_t { 490 | struct subpass_t { 491 | std::vector input_attachments; 492 | std::vector color_attachments; 493 | std::vector resolve_attachments; 494 | VkAttachmentReference depth_attachment; 495 | std::vector preserve_attachments; 496 | }; 497 | 498 | struct attachment_t { 499 | VkAttachmentDescription desc; 500 | std::optional first_use; 501 | }; 502 | 503 | std::vector subpasses; 504 | std::vector attachments; 505 | std::vector dependencies; 506 | }; 507 | 508 | struct framebuffer_t { 509 | std::vector attachments; 510 | }; 511 | 512 | struct image_view_t { 513 | ID3D12Resource* image; 514 | std::optional> rtv; 515 | std::optional> dsv; 516 | std::optional> srv; 517 | std::optional> uav; // TODO: destroy 518 | }; 519 | 520 | struct shader_module_t { 521 | // Raw SPIR-V code 522 | std::vector spirv; 523 | }; 524 | 525 | struct semaphore_t { 526 | ComPtr fence; 527 | }; 528 | 529 | struct surface_t { 530 | ComPtr dxgi_factory; 531 | std::variant handle; 532 | }; 533 | 534 | struct swapchain_t { 535 | ComPtr swapchain; 536 | std::vector images; 537 | }; 538 | 539 | struct buffer_t { 540 | ComPtr resource; 541 | VkMemoryRequirements memory_requirements; 542 | D3D12_RESOURCE_FLAGS usage_flags; 543 | }; 544 | 545 | struct device_memory_t { 546 | ComPtr heap; 547 | ComPtr buffer; 548 | VkDeviceSize size; 549 | }; 550 | 551 | struct pipeline_cache_t { 552 | 553 | }; 554 | 555 | struct fence_t { 556 | ComPtr fence; 557 | }; 558 | 559 | struct descriptor_set_layout_t { 560 | struct binding_t { 561 | uint32_t binding; 562 | VkDescriptorType type; 563 | uint32_t descriptor_count; 564 | VkShaderStageFlags stage_flags; 565 | std::vector immutable_samplers; 566 | }; 567 | 568 | std::vector layouts; 569 | }; 570 | 571 | /// CbvSrvUav placed descriptor sets 572 | struct descriptor_set_placed_t { 573 | // A descriptor sets is a subslice fo a descriptor pool. 574 | // Needed when binding descriptor sets. 575 | D3D12_GPU_DESCRIPTOR_HANDLE gpu_start; 576 | }; 577 | 578 | /// Sampler virtual descriptor set (dynamic gpu heap placement) 579 | struct descriptor_set_virtual_t { 580 | D3D12_CPU_DESCRIPTOR_HANDLE cpu_handle; 581 | UINT num_descriptors; // Required for uploading 582 | std::set heaps_placed; 583 | }; 584 | 585 | struct descriptor_set_t { 586 | using binding_index = size_t; 587 | 588 | struct binding_t { 589 | D3D12_CPU_DESCRIPTOR_HANDLE start_cbv_srv_uav; 590 | D3D12_CPU_DESCRIPTOR_HANDLE start_sampler; 591 | size_t num_descriptors; 592 | }; 593 | 594 | std::optional set_cbv_srv_uav; 595 | std::optional set_sampler; 596 | 597 | // Each binding of the descriptor set is again a slice of the descriptor set slice. 598 | // Needed for updating descriptor sets. 599 | std::map bindings; 600 | }; 601 | 602 | struct descriptor_pool_t { 603 | public: 604 | descriptor_pool_t(size_t num_cbv_srv_uav, size_t num_sampler) : 605 | slice_cbv_srv_uav { num_cbv_srv_uav }, 606 | slice_sampler { num_sampler } 607 | { } 608 | 609 | auto alloc(size_t num_cbv_srv_uav, size_t num_sampler) 610 | -> std::tuple 611 | { 612 | const auto start_cbv_srv_uav = this->slice_cbv_srv_uav.alloc(num_cbv_srv_uav); 613 | const auto start_sampler = this->slice_sampler.alloc(num_sampler); 614 | 615 | return std::make_tuple(start_cbv_srv_uav, start_sampler); 616 | } 617 | 618 | public: 619 | struct slice_t { 620 | public: 621 | slice_t(size_t num) : 622 | allocator { num } 623 | {} 624 | 625 | auto alloc(size_t num) -> descriptor_cpu_gpu_handle_t { 626 | const auto range = this->allocator.alloc(num); 627 | if (range) { 628 | const auto [start_cpu, start_gpu] = this->start; 629 | return std::make_tuple( 630 | D3D12_CPU_DESCRIPTOR_HANDLE { start_cpu.ptr + this->handle_size * range->_start }, 631 | D3D12_GPU_DESCRIPTOR_HANDLE { start_gpu.ptr + this->handle_size * range->_start } 632 | ); 633 | } else { 634 | // TODO 635 | assert(!"Not enough free descriptors in the allocator"); 636 | return std::make_tuple( 637 | D3D12_CPU_DESCRIPTOR_HANDLE { 0 }, 638 | D3D12_GPU_DESCRIPTOR_HANDLE { 0 } 639 | ); 640 | } 641 | } 642 | public: 643 | descriptor_cpu_gpu_handle_t start; 644 | UINT handle_size; 645 | free_list allocator; 646 | }; 647 | 648 | slice_t slice_cbv_srv_uav; 649 | slice_t slice_sampler; 650 | }; 651 | 652 | enum root_table_flags { 653 | TABLE_CBV_SRV_UAV = 0x1, 654 | TABLE_SAMPLER = 0x2, 655 | }; 656 | 657 | struct pipeline_layout_t { 658 | ComPtr signature; 659 | std::vector tables; // root_table_flags 660 | std::vector root_constants; 661 | size_t num_root_constants; // Number of root constants (32bit) used for push constants 662 | size_t num_dynamic_offsets; 663 | size_t num_signature_entries; 664 | }; 665 | 666 | struct sampler_t { 667 | std::tuple sampler; 668 | }; 669 | 670 | enum dynamic_state_flags { 671 | DYNAMIC_STATE_DEPTH_BIAS = 0x1, 672 | DYNAMIC_STATE_STENCIL_COMPARE_MASK = 0x2, 673 | DYNAMIC_STATE_STENCIL_WRITE_MASK = 0x4, 674 | DYNAMIC_STATE_PRIMITIVE_RESTART = 0x8, 675 | }; 676 | using dynamic_states_t = uint32_t; 677 | 678 | struct dynamic_state_t { 679 | dynamic_state_t() : 680 | depth_bias { 0 }, 681 | depth_bias_clamp { 0.0f }, 682 | depth_bias_slope { 0.0f }, 683 | stencil_read_mask { 0 }, 684 | stencil_write_mask { 0 }, 685 | strip_cut { D3D12_INDEX_BUFFER_STRIP_CUT_VALUE_DISABLED } 686 | { } 687 | 688 | INT depth_bias; 689 | FLOAT depth_bias_clamp; 690 | FLOAT depth_bias_slope; 691 | UINT8 stencil_read_mask; 692 | UINT8 stencil_write_mask; 693 | D3D12_INDEX_BUFFER_STRIP_CUT_VALUE strip_cut; 694 | 695 | auto operator== (dynamic_state_t const& rhs) const -> bool { 696 | return 697 | this->depth_bias == rhs.depth_bias && 698 | this->depth_bias_clamp == rhs.depth_bias_clamp && 699 | this->depth_bias_slope == rhs.depth_bias_slope && 700 | this->stencil_read_mask == rhs.stencil_read_mask && 701 | this->stencil_write_mask == rhs.stencil_write_mask && 702 | this->strip_cut == rhs.strip_cut; 703 | } 704 | }; 705 | 706 | namespace std { 707 | template<> 708 | struct hash { 709 | auto operator()(dynamic_state_t const& v) const -> std::size_t { 710 | std::size_t hash { 0 }; 711 | stdx::hash_combine( 712 | hash, 713 | v.depth_bias, 714 | v.depth_bias_clamp, 715 | v.depth_bias_slope, 716 | v.stencil_read_mask, 717 | v.stencil_write_mask, 718 | v.strip_cut 719 | ); 720 | return hash; 721 | } 722 | }; 723 | } 724 | 725 | struct blend_factors_t { 726 | FLOAT factors[4]; 727 | }; 728 | 729 | struct pipeline_t { 730 | pipeline_t() : 731 | signature { nullptr }, 732 | num_signature_entries { 0 }, 733 | num_root_constants { 0 }, 734 | num_dynamic_offsets { 0 }, 735 | dynamic_states { 0 }, 736 | static_viewports { std::nullopt }, 737 | static_scissors { std::nullopt }, 738 | static_blend_factors { std::nullopt }, 739 | static_depth_bounds { std::nullopt }, 740 | static_stencil_reference { std::nullopt } 741 | {} 742 | 743 | // There is only a single pipeline, no need for dynamic creation 744 | // Compute pipelines and graphics pipeline without primitive restart and dynamic states 745 | struct unique_pso_t { 746 | ComPtr pipeline; 747 | }; 748 | 749 | // Pipeline with dynamic states 750 | // - Viewport and scissor are dynamic per se 751 | // - Line width must be 1.0 752 | // - Blend constants are dynamic per se 753 | // - Depth bounds are dynamic per se (if supported) 754 | // - Stencil ref is dynamic per se 755 | // - Index primitive restart value must be set dynamically 756 | // 757 | // Access to these pipelines need to be ensured by `pso_access`. 758 | struct dynamic_pso_t { 759 | std::unordered_map> pipelines; 760 | 761 | // desc for dynamic pipeline creation 762 | ComPtr vertex_shader; 763 | ComPtr domain_shader; 764 | ComPtr hull_shader; 765 | ComPtr geometry_shader; 766 | ComPtr pixel_shader; 767 | D3D12_BLEND_DESC blend_state; 768 | UINT sample_mask; 769 | D3D12_RASTERIZER_DESC rasterizer_state; 770 | D3D12_DEPTH_STENCIL_DESC depth_stencil_state; 771 | std::vector input_elements; 772 | UINT num_render_targets; 773 | DXGI_FORMAT rtv_formats[8]; 774 | DXGI_FORMAT dsv_format; 775 | DXGI_SAMPLE_DESC sample_desc; 776 | D3D12_PRIMITIVE_TOPOLOGY_TYPE topology_type; 777 | }; 778 | 779 | // Shared by compute and graphics 780 | std::variant pso; 781 | // Required for accessing 782 | std::shared_mutex pso_access; 783 | ID3D12RootSignature* signature; 784 | 785 | size_t num_signature_entries; 786 | size_t num_root_constants; // Number of root constants (32bit) in the root signature for push constants 787 | size_t num_dynamic_offsets; 788 | std::vector root_constants; 789 | 790 | // Graphics only 791 | D3D12_PRIMITIVE_TOPOLOGY topology; 792 | std::array vertex_strides; 793 | 794 | bool primitive_restart; 795 | dynamic_states_t dynamic_states; 796 | 797 | dynamic_state_t static_state; 798 | std::optional> static_viewports; 799 | std::optional> static_scissors; 800 | std::optional static_blend_factors; 801 | std::optional> static_depth_bounds; 802 | std::optional static_stencil_reference; 803 | }; 804 | 805 | static auto up_align(UINT v, UINT alignment) -> UINT { 806 | return (v + alignment - 1) & ~(alignment - 1); 807 | } 808 | 809 | auto create_command_signature( 810 | ID3D12Device* device, 811 | D3D12_INDIRECT_ARGUMENT_TYPE type, 812 | UINT stride 813 | ) -> ComPtr; 814 | -------------------------------------------------------------------------------- /src/lib.cpp: -------------------------------------------------------------------------------- 1 | 2 | #include "icd.hpp" 3 | 4 | #include 5 | #include 6 | #include 7 | #include 8 | 9 | extern "C" { 10 | 11 | #undef VKAPI_ATTR 12 | #define VKAPI_ATTR __declspec(dllexport) 13 | 14 | VKAPI_ATTR PFN_vkVoidFunction VKAPI_CALL vk_icdGetInstanceProcAddr(VkInstance instance, const char *pName) { 15 | return vkGetInstanceProcAddr(instance, pName); 16 | } 17 | 18 | VKAPI_ATTR PFN_vkVoidFunction VKAPI_CALL vk_icdGetPhysicalDeviceProcAddr(VkInstance instance, const char *pName) { 19 | VK_PHYSICAL_DEVICE_FNC() 20 | 21 | return nullptr; 22 | } 23 | 24 | VKAPI_ATTR VkResult VKAPI_CALL vk_icdNegotiateLoaderICDInterfaceVersion(uint32_t *pSupportedVersion) { 25 | *pSupportedVersion = std::min(*pSupportedVersion, CURRENT_LOADER_ICD_INTERFACE_VERSION); 26 | return VK_SUCCESS; 27 | } 28 | 29 | }; // extern "C" -------------------------------------------------------------------------------- /src/wsi.cpp: -------------------------------------------------------------------------------- 1 | //! Windows system interface related implementations 2 | 3 | #include "device.hpp" 4 | #include "impl.hpp" 5 | 6 | #include 7 | 8 | #include 9 | 10 | VKAPI_ATTR VkResult VKAPI_CALL vkCreateSwapchainKHR( 11 | VkDevice _device, 12 | const VkSwapchainCreateInfoKHR* pCreateInfo, 13 | const VkAllocationCallbacks* pAllocator, 14 | VkSwapchainKHR* pSwapchain 15 | ) { 16 | TRACE("vkCreateSwapchainKHR"); 17 | 18 | auto const& info { *pCreateInfo }; 19 | auto device { reinterpret_cast(_device) }; 20 | auto surface { reinterpret_cast(info.surface) }; 21 | auto old_swapchain { reinterpret_cast(info.oldSwapchain) }; 22 | 23 | if (old_swapchain) { 24 | for (auto& image : old_swapchain->images) { 25 | image.resource.Reset(); 26 | } 27 | old_swapchain->swapchain.Reset(); 28 | } 29 | 30 | auto vk_format = info.imageFormat; 31 | // Flip model swapchain doesn't allow srgb image formats, requires workaround with UNORM. 32 | switch (vk_format) { 33 | case VK_FORMAT_B8G8R8A8_SRGB: vk_format = VK_FORMAT_B8G8R8A8_UNORM; break; 34 | case VK_FORMAT_R8G8B8A8_SRGB: vk_format = VK_FORMAT_R8G8B8A8_UNORM; break; 35 | } 36 | auto const format { formats[vk_format] }; 37 | auto swapchain { new swapchain_t }; 38 | 39 | ComPtr dxgi_swapchain { nullptr }; 40 | const DXGI_SWAP_CHAIN_DESC1 desc { 41 | info.imageExtent.width, 42 | info.imageExtent.height, 43 | format, 44 | FALSE, 45 | { 1, 0 }, 46 | 0, // TODO: usage 47 | info.minImageCount, 48 | DXGI_SCALING_NONE, 49 | DXGI_SWAP_EFFECT_FLIP_DISCARD, // TODO 50 | DXGI_ALPHA_MODE_UNSPECIFIED, // TODO 51 | 0, // TODO: flags 52 | }; 53 | 54 | const auto hr { 55 | std::visit( 56 | stdx::match( 57 | [&] (HWND hwnd) { 58 | return surface->dxgi_factory->CreateSwapChainForHwnd( 59 | device->present_queue.queue.Get(), 60 | hwnd, 61 | &desc, 62 | nullptr, // TODO: fullscreen 63 | nullptr, // TODO: restrict? 64 | &dxgi_swapchain 65 | ); 66 | }, 67 | [&] (IUnknown* window) { 68 | return surface->dxgi_factory->CreateSwapChainForCoreWindow( 69 | device->present_queue.queue.Get(), 70 | window, 71 | &desc, 72 | nullptr, // TODO: restrict? 73 | &dxgi_swapchain 74 | ); 75 | } 76 | ), 77 | surface->handle 78 | ) 79 | }; 80 | 81 | if (FAILED(hr)) { 82 | assert(false); 83 | } 84 | // TODO: errror 85 | 86 | if (FAILED(dxgi_swapchain.As(&(swapchain->swapchain)))) { 87 | ERR("Couldn't convert swapchain to `IDXGISwapChain3`!"); 88 | } 89 | 90 | for (auto i : range(info.minImageCount)) { 91 | ComPtr resource { nullptr }; 92 | swapchain->swapchain->GetBuffer(i, IID_PPV_ARGS(&resource)); 93 | 94 | const D3D12_RESOURCE_DESC desc { 95 | D3D12_RESOURCE_DIMENSION_TEXTURE2D, 96 | 0, 97 | info.imageExtent.width, 98 | info.imageExtent.height, 99 | 1, 100 | 1, 101 | format, 102 | { 1, 0 }, 103 | D3D12_TEXTURE_LAYOUT_UNKNOWN, // TODO 104 | D3D12_RESOURCE_FLAG_NONE, // TODO 105 | }; 106 | 107 | swapchain->images.emplace_back( 108 | image_t { 109 | resource, 110 | { }, 111 | desc, 112 | formats_block[vk_format], 113 | info.imageUsage, 114 | } 115 | ); 116 | } 117 | 118 | *pSwapchain = reinterpret_cast(swapchain); 119 | 120 | return VK_SUCCESS; 121 | } 122 | 123 | VKAPI_ATTR void VKAPI_CALL vkDestroySwapchainKHR( 124 | VkDevice device, 125 | VkSwapchainKHR swapchain, 126 | const VkAllocationCallbacks* pAllocator 127 | ) { 128 | WARN("vkDestroySwapchainKHR unimplemented"); 129 | } 130 | 131 | VKAPI_ATTR VkResult VKAPI_CALL vkGetSwapchainImagesKHR( 132 | VkDevice device, 133 | VkSwapchainKHR _swapchain, 134 | uint32_t* pSwapchainImageCount, 135 | VkImage* pSwapchainImages 136 | ) { 137 | TRACE("vkGetSwapchainImagesKHR"); 138 | 139 | auto swapchain { reinterpret_cast(_swapchain) }; 140 | auto num_images { static_cast(swapchain->images.size()) }; 141 | 142 | if (!pSwapchainImages) { 143 | *pSwapchainImageCount = num_images; 144 | return VK_SUCCESS; 145 | } 146 | 147 | auto result { VK_SUCCESS }; 148 | if (*pSwapchainImageCount < num_images) { 149 | num_images = *pSwapchainImageCount; 150 | result = VK_INCOMPLETE; 151 | } 152 | 153 | for (auto i : range(num_images)) { 154 | pSwapchainImages[i] = reinterpret_cast(&swapchain->images[i]); 155 | } 156 | 157 | return VK_SUCCESS; 158 | } 159 | 160 | VKAPI_ATTR VkResult VKAPI_CALL vkAcquireNextImageKHR( 161 | VkDevice device, 162 | VkSwapchainKHR _swapchain, 163 | uint64_t timeout, 164 | VkSemaphore semaphore, 165 | VkFence fence, 166 | uint32_t* pImageIndex 167 | ) { 168 | TRACE("vkAcquireNextImageKHR"); 169 | 170 | // TODO: single in order images atm with additional limitations, looking into alternatives.. 171 | // TODO: sync stuff 172 | 173 | auto swapchain { reinterpret_cast(_swapchain) }; 174 | *pImageIndex = swapchain->swapchain->GetCurrentBackBufferIndex(); 175 | 176 | return VK_SUCCESS; 177 | } 178 | 179 | VKAPI_ATTR VkResult VKAPI_CALL vkQueuePresentKHR( 180 | VkQueue queue, 181 | const VkPresentInfoKHR* pPresentInfo 182 | ) { 183 | TRACE("vkQueuePresentKHR"); 184 | 185 | auto const& info { *pPresentInfo }; 186 | auto wait_semaphores { span(info.pWaitSemaphores, info.waitSemaphoreCount) }; 187 | auto swapchains { span(info.pSwapchains, info.swapchainCount) }; 188 | auto image_indices { span(info.pImageIndices, info.swapchainCount) }; 189 | 190 | // TODO: image indices, results, semaphores 191 | 192 | for (auto i : range(info.swapchainCount)) { 193 | auto const& swapchain { reinterpret_cast(swapchains[i]) }; 194 | 195 | swapchain->swapchain->Present(0, 0); // TODO: vsync? 196 | } 197 | 198 | return VK_SUCCESS; 199 | } 200 | 201 | VKAPI_ATTR void VKAPI_CALL vkDestroySurfaceKHR( 202 | VkInstance instance, 203 | VkSurfaceKHR surface, 204 | const VkAllocationCallbacks* pAllocator 205 | ) { 206 | TRACE("vkDestroySurfaceKHR unimplemented"); 207 | } 208 | 209 | VKAPI_ATTR VkResult VKAPI_CALL vkGetPhysicalDeviceSurfaceSupportKHR( 210 | VkPhysicalDevice physicalDevice, 211 | uint32_t queueFamilyIndex, 212 | VkSurfaceKHR surface, 213 | VkBool32* pSupported 214 | ) { 215 | TRACE("vkGetPhysicalDeviceSurfaceSupportKHR"); 216 | 217 | *pSupported = queueFamilyIndex == QUEUE_FAMILY_GENERAL_PRESENT; 218 | 219 | return VK_SUCCESS; 220 | } 221 | 222 | VKAPI_ATTR VkResult VKAPI_CALL vkGetPhysicalDeviceSurfaceCapabilitiesKHR( 223 | VkPhysicalDevice physicalDevice, 224 | VkSurfaceKHR _surface, 225 | VkSurfaceCapabilitiesKHR* pSurfaceCapabilities 226 | ) { 227 | TRACE("vkGetPhysicalDeviceSurfaceCapabilitiesKHR"); 228 | 229 | auto surface { reinterpret_cast(_surface) }; 230 | 231 | const auto hwnd { 232 | std::visit( 233 | stdx::match( 234 | [&] (HWND hwnd) { 235 | return hwnd; 236 | }, 237 | [&] (IUnknown *window) { 238 | ComPtr core_window { nullptr }; 239 | window->QueryInterface(core_window.GetAddressOf()); 240 | 241 | HWND hwnd; 242 | core_window->get_WindowHandle(&hwnd); 243 | return hwnd; 244 | } 245 | ), 246 | surface->handle 247 | ) 248 | }; 249 | 250 | RECT rect; 251 | if (!::GetClientRect(hwnd, &rect)) { 252 | // TODO 253 | ERR("Couldn't get size of window"); 254 | } 255 | 256 | VkExtent2D extent { 257 | static_cast(rect.right - rect.left), 258 | static_cast(rect.bottom - rect.top), 259 | }; 260 | 261 | *pSurfaceCapabilities = { 262 | // Image count due to FLIP_DISCARD 263 | 2, // minImageCount 264 | 16, // maxImageCount 265 | extent, // currentExtent 266 | extent, // minImageExtent 267 | extent, // maxImageExtent 268 | 1, // maxImageArrayLayers // TODO 269 | VK_SURFACE_TRANSFORM_IDENTITY_BIT_KHR, // supportedTransforms // TODO 270 | VK_SURFACE_TRANSFORM_IDENTITY_BIT_KHR, // currentTransform // TODO 271 | VK_COMPOSITE_ALPHA_OPAQUE_BIT_KHR, // supportedCompositeAlpha // TODO 272 | VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT, // supportedUsageFlags // TODO 273 | }; 274 | 275 | return VK_SUCCESS; 276 | } 277 | 278 | VKAPI_ATTR VkResult VKAPI_CALL vkGetPhysicalDeviceSurfaceFormatsKHR( 279 | VkPhysicalDevice physicalDevice, 280 | VkSurfaceKHR surface, 281 | uint32_t* pSurfaceFormatCount, 282 | VkSurfaceFormatKHR* pSurfaceFormats 283 | ) { 284 | TRACE("vkGetPhysicalDeviceSurfaceFormatsKHR"); 285 | 286 | // TODO: more formats 287 | const std::array formats {{ 288 | { VK_FORMAT_B8G8R8A8_SRGB, VK_COLOR_SPACE_SRGB_NONLINEAR_KHR }, 289 | { VK_FORMAT_B8G8R8A8_UNORM, VK_COLOR_SPACE_SRGB_NONLINEAR_KHR } 290 | }}; 291 | 292 | auto num_formats { static_cast(formats.size()) }; 293 | 294 | if (!pSurfaceFormats) { 295 | *pSurfaceFormatCount = num_formats; 296 | return VK_SUCCESS; 297 | } 298 | 299 | auto result { VK_SUCCESS }; 300 | if (*pSurfaceFormatCount < num_formats) { 301 | num_formats = *pSurfaceFormatCount; 302 | result = VK_INCOMPLETE; 303 | } 304 | 305 | for (auto i : range(num_formats)) { 306 | pSurfaceFormats[i] = formats[i]; 307 | } 308 | 309 | return result; 310 | } 311 | 312 | VKAPI_ATTR VkResult VKAPI_CALL vkGetPhysicalDeviceSurfacePresentModesKHR( 313 | VkPhysicalDevice physicalDevice, 314 | VkSurfaceKHR surface, 315 | uint32_t* pPresentModeCount, 316 | VkPresentModeKHR* pPresentModes 317 | ) { 318 | TRACE("vkGetPhysicalDeviceSurfacePresentModesKHR"); 319 | 320 | // TODO 321 | const std::array modes { 322 | VK_PRESENT_MODE_FIFO_KHR 323 | }; 324 | 325 | auto num_modes { static_cast(modes.size()) }; 326 | 327 | if (!pPresentModes) { 328 | *pPresentModeCount = num_modes; 329 | return VK_SUCCESS; 330 | } 331 | 332 | auto result { VK_SUCCESS }; 333 | if (*pPresentModeCount < num_modes) { 334 | num_modes = *pPresentModeCount; 335 | result = VK_INCOMPLETE; 336 | } 337 | 338 | for (auto i : range(num_modes)) { 339 | pPresentModes[i] = modes[i]; 340 | } 341 | 342 | return result; 343 | } 344 | 345 | VKAPI_ATTR VkResult VKAPI_CALL vkCreateWin32SurfaceKHR( 346 | VkInstance _instance, 347 | const VkWin32SurfaceCreateInfoKHR* pCreateInfo, 348 | const VkAllocationCallbacks* pAllocator, 349 | VkSurfaceKHR* pSurface 350 | ) { 351 | TRACE("vkCreateWin32SurfaceKHR unimplemented"); 352 | 353 | auto instance { reinterpret_cast(_instance) }; 354 | auto const& info { *pCreateInfo }; 355 | 356 | *pSurface = reinterpret_cast( 357 | new surface_t { instance->dxgi_factory, info.hwnd} 358 | ); 359 | 360 | return VK_SUCCESS; 361 | } 362 | 363 | VKAPI_ATTR VkResult VKAPI_CALL vkCreateUWPSurfaceRKZ( 364 | VkInstance _instance, 365 | const VkUWPSurfaceCreateInfoRKZ* pCreateInfo, 366 | const VkAllocationCallbacks* pAllocator, 367 | VkSurfaceKHR* pSurface 368 | ) { 369 | TRACE("vkCreateUWPSurfaceRKZ"); 370 | 371 | auto instance { reinterpret_cast(_instance) }; 372 | auto const& info { *pCreateInfo }; 373 | 374 | *pSurface = reinterpret_cast( 375 | new surface_t { instance->dxgi_factory, info.pWindow} 376 | ); 377 | 378 | return VK_SUCCESS; 379 | } 380 | 381 | VKAPI_ATTR VkBool32 VKAPI_CALL vkGetPhysicalDeviceUWPPresentationSupportRKZ( 382 | VkPhysicalDevice physicalDevice, 383 | uint32_t queueFamilyIndex 384 | ) { 385 | return queueFamilyIndex == QUEUE_FAMILY_GENERAL_PRESENT; 386 | } 387 | --------------------------------------------------------------------------------