diff --git a/neo/renderer/ImageManager.cpp b/neo/renderer/ImageManager.cpp index 64d65a97..fc39db36 100644 --- a/neo/renderer/ImageManager.cpp +++ b/neo/renderer/ImageManager.cpp @@ -40,7 +40,7 @@ idImageManager imageManager; idImageManager* globalImages = &imageManager; extern DeviceManager* deviceManager; -extern idCVar r_uploadBufferSizeMB; +extern idCVar r_vkUploadBufferSizeMB; idCVar preLoad_Images( "preLoad_Images", "1", CVAR_SYSTEM | CVAR_BOOL, "preload images during beginlevelload" ); @@ -886,7 +886,7 @@ int idImageManager::LoadLevelImages( bool pacifier ) if( deviceManager->GetGraphicsAPI() == nvrhi::GraphicsAPI::VULKAN ) { // SRS - set upload buffer size to avoid Vulkan staging buffer fragmentation - size_t maxBufferSize = ( size_t )( r_uploadBufferSizeMB.GetInteger() * 1024 * 1024 ); + size_t maxBufferSize = ( size_t )( r_vkUploadBufferSizeMB.GetInteger() * 1024 * 1024 ); params.setUploadChunkSize( maxBufferSize ); } commandList = deviceManager->GetDevice()->createCommandList( params ); @@ -1009,7 +1009,7 @@ void idImageManager::LoadDeferredImages( nvrhi::ICommandList* _commandList ) if( deviceManager->GetGraphicsAPI() == nvrhi::GraphicsAPI::VULKAN ) { // SRS - set upload buffer size to avoid Vulkan staging buffer fragmentation - size_t maxBufferSize = ( size_t )( r_uploadBufferSizeMB.GetInteger() * 1024 * 1024 ); + size_t maxBufferSize = ( size_t )( r_vkUploadBufferSizeMB.GetInteger() * 1024 * 1024 ); params.setUploadChunkSize( maxBufferSize ); } commandList = deviceManager->GetDevice()->createCommandList( params ); diff --git a/neo/renderer/ModelManager.cpp b/neo/renderer/ModelManager.cpp index 85842ceb..5739f22a 100644 --- a/neo/renderer/ModelManager.cpp +++ b/neo/renderer/ModelManager.cpp @@ -37,7 +37,7 @@ If you have questions concerning this license or the applicable additional terms #include extern DeviceManager* deviceManager; -extern idCVar r_uploadBufferSizeMB; +extern idCVar r_vkUploadBufferSizeMB; idCVar binaryLoadRenderModels( "binaryLoadRenderModels", "1", 0, "enable binary load/write of render models" ); idCVar preload_MapModels( "preload_MapModels", "1", CVAR_SYSTEM | CVAR_BOOL, "preload models during begin or end levelload" ); @@ -248,7 +248,7 @@ void idRenderModelManagerLocal::Init() if( deviceManager->GetGraphicsAPI() == nvrhi::GraphicsAPI::VULKAN ) { // SRS - set upload buffer size to avoid Vulkan staging buffer fragmentation - size_t maxBufferSize = ( size_t )( r_uploadBufferSizeMB.GetInteger() * 1024 * 1024 ); + size_t maxBufferSize = ( size_t )( r_vkUploadBufferSizeMB.GetInteger() * 1024 * 1024 ); params.setUploadChunkSize( maxBufferSize ); } commandList = deviceManager->GetDevice()->createCommandList( params ); diff --git a/neo/renderer/NVRHI/RenderBackend_NVRHI.cpp b/neo/renderer/NVRHI/RenderBackend_NVRHI.cpp index 4cbed29f..59a793a3 100644 --- a/neo/renderer/NVRHI/RenderBackend_NVRHI.cpp +++ b/neo/renderer/NVRHI/RenderBackend_NVRHI.cpp @@ -56,7 +56,7 @@ idCVar stereoRender_warpTargetFraction( "stereoRender_warpTargetFraction", "1.0" idCVar r_showSwapBuffers( "r_showSwapBuffers", "0", CVAR_BOOL, "Show timings from GL_BlockingSwapBuffers" ); idCVar r_syncEveryFrame( "r_syncEveryFrame", "1", CVAR_BOOL, "Don't let the GPU buffer execution past swapbuffers" ); -idCVar r_uploadBufferSizeMB( "r_uploadBufferSizeMB", "64", CVAR_INTEGER | CVAR_INIT, "Size of gpu upload buffer (Vulkan only)" ); +idCVar r_vkUploadBufferSizeMB( "r_vkUploadBufferSizeMB", "64", CVAR_INTEGER | CVAR_INIT, "Size of gpu upload buffer (Vulkan only)" ); constexpr std::size_t MAX_IMAGE_PARMS = 16; @@ -222,7 +222,7 @@ void idRenderBackend::Init() if( deviceManager->GetGraphicsAPI() == nvrhi::GraphicsAPI::VULKAN ) { // SRS - set upload buffer size to avoid Vulkan staging buffer fragmentation - size_t maxBufferSize = ( size_t )( r_uploadBufferSizeMB.GetInteger() * 1024 * 1024 ); + size_t maxBufferSize = ( size_t )( r_vkUploadBufferSizeMB.GetInteger() * 1024 * 1024 ); params.setUploadChunkSize( maxBufferSize ); } commandList = deviceManager->GetDevice()->createCommandList( params ); diff --git a/neo/sys/DeviceManager_DX12.cpp b/neo/sys/DeviceManager_DX12.cpp index bf3badd9..4669a3af 100644 --- a/neo/sys/DeviceManager_DX12.cpp +++ b/neo/sys/DeviceManager_DX12.cpp @@ -48,7 +48,7 @@ using nvrhi::RefCountPtr; #define HR_RETURN(hr) if(FAILED(hr)) return false idCVar r_graphicsAdapter( "r_graphicsAdapter", "", CVAR_RENDERER | CVAR_INIT | CVAR_ARCHIVE, "Substring in the name the DXGI graphics adapter to select a certain GPU" ); -idCVar r_maxFrameLatency( "r_maxFrameLatency", "2", CVAR_RENDERER | CVAR_INIT | CVAR_ARCHIVE | CVAR_INTEGER, "Maximum frame latency for DXGI swap chains (DX12 only)", 0, NUM_FRAME_DATA ); +idCVar r_dxMaxFrameLatency( "r_dxMaxFrameLatency", "2", CVAR_RENDERER | CVAR_INIT | CVAR_ARCHIVE | CVAR_INTEGER, "Maximum frame latency for DXGI swap chains (DX12 only)", 0, NUM_FRAME_DATA ); class DeviceManager_DX12 : public DeviceManager { @@ -314,7 +314,7 @@ bool DeviceManager_DX12::CreateDeviceAndSwapChain() m_SwapChainDesc.BufferCount = m_DeviceParams.swapChainBufferCount; m_SwapChainDesc.SwapEffect = DXGI_SWAP_EFFECT_FLIP_DISCARD; m_SwapChainDesc.Flags = ( m_DeviceParams.allowModeSwitch ? DXGI_SWAP_CHAIN_FLAG_ALLOW_MODE_SWITCH : 0 ) | - ( r_maxFrameLatency.GetInteger() > 0 ? DXGI_SWAP_CHAIN_FLAG_FRAME_LATENCY_WAITABLE_OBJECT : 0 ); + ( r_dxMaxFrameLatency.GetInteger() > 0 ? DXGI_SWAP_CHAIN_FLAG_FRAME_LATENCY_WAITABLE_OBJECT : 0 ); // Special processing for sRGB swap chain formats. // DXGI will not create a swap chain with an sRGB format, but its contents will be interpreted as sRGB. @@ -433,9 +433,9 @@ bool DeviceManager_DX12::CreateDeviceAndSwapChain() hr = pSwapChain1->QueryInterface( IID_PPV_ARGS( &m_SwapChain ) ); HR_RETURN( hr ); - if( r_maxFrameLatency.GetInteger() > 0 ) + if( r_dxMaxFrameLatency.GetInteger() > 0 ) { - hr = m_SwapChain->SetMaximumFrameLatency( r_maxFrameLatency.GetInteger() ); + hr = m_SwapChain->SetMaximumFrameLatency( r_dxMaxFrameLatency.GetInteger() ); HR_RETURN( hr ); m_frameLatencyWaitableObject = m_SwapChain->GetFrameLatencyWaitableObject(); diff --git a/neo/sys/DeviceManager_VK.cpp b/neo/sys/DeviceManager_VK.cpp index c2e683f5..2dc0b6ef 100644 --- a/neo/sys/DeviceManager_VK.cpp +++ b/neo/sys/DeviceManager_VK.cpp @@ -68,7 +68,7 @@ idCVar r_vmaDeviceLocalMemoryMB( "r_vmaDeviceLocalMemoryMB", "256", CVAR_INTEGER | CVAR_INIT, "Size of VMA allocation block for gpu memory." ); #endif -idCVar r_preferFastSync( "r_preferFastSync", "1", CVAR_RENDERER | CVAR_ARCHIVE | CVAR_BOOL, "Prefer Fast Sync/no-tearing in place of VSync off/tearing (Vulkan only)" ); +idCVar r_vkPreferFastSync( "r_vkPreferFastSync", "1", CVAR_RENDERER | CVAR_ARCHIVE | CVAR_BOOL, "Prefer Fast Sync/no-tearing in place of VSync off/tearing" ); // Define the Vulkan dynamic dispatcher - this needs to occur in exactly one cpp file in the program. VULKAN_HPP_DEFAULT_DISPATCH_LOADER_DYNAMIC_STORAGE @@ -1178,7 +1178,7 @@ bool DeviceManager_VK::createSwapChain() switch( m_DeviceParams.vsyncEnabled ) { case 0: - presentMode = enablePModeMailbox && r_preferFastSync.GetBool() ? vk::PresentModeKHR::eMailbox : + presentMode = enablePModeMailbox && r_vkPreferFastSync.GetBool() ? vk::PresentModeKHR::eMailbox : ( enablePModeImmediate ? vk::PresentModeKHR::eImmediate : vk::PresentModeKHR::eFifo ); break; case 1: