From a5c1b63574e5c184d8130d27e10805288aee7a6b Mon Sep 17 00:00:00 2001 From: Simon Date: Wed, 3 Feb 2021 22:13:37 +0000 Subject: [PATCH] Multiple improvements - Single movement speed cvar - Weapon Haptics on weapon controller - Proper compositor layer for menu - Weapon hand thumbstick up and down to switch weapons - Removed weapon angle adjustment as default is fine --- Projects/Android/Android.iml | 6 +- Projects/Android/android.debug.keystore | Bin 0 -> 2159 bytes Projects/Android/jni/Android.mk | 3 +- .../QuakeQuestSrc/QuakeQuest_SurfaceView.c | 920 +++++++++--------- Projects/Android/jni/QuakeQuestSrc/VrCommon.h | 38 + .../Android/jni/QuakeQuestSrc/VrCompositor.c | 595 +++++++++++ .../Android/jni/QuakeQuestSrc/VrCompositor.h | 216 ++++ Projects/Android/jni/darkplaces/cl_input.c | 34 +- Projects/Android/jni/darkplaces/cl_main.c | 4 +- Projects/Android/jni/darkplaces/client.h | 4 +- Projects/Android/jni/darkplaces/gl_rmain.c | 4 +- Projects/Android/jni/darkplaces/menu.c | 28 +- Projects/Android/jni/darkplaces/sbar.c | 4 +- Projects/Android/jni/darkplaces/vid_shared.c | 8 +- Projects/Android/jni/darkplaces/view.c | 2 +- 15 files changed, 1331 insertions(+), 535 deletions(-) create mode 100644 Projects/Android/android.debug.keystore create mode 100644 Projects/Android/jni/QuakeQuestSrc/VrCommon.h create mode 100644 Projects/Android/jni/QuakeQuestSrc/VrCompositor.c create mode 100644 Projects/Android/jni/QuakeQuestSrc/VrCompositor.h diff --git a/Projects/Android/Android.iml b/Projects/Android/Android.iml index 7c00c54..2791cf7 100644 --- a/Projects/Android/Android.iml +++ b/Projects/Android/Android.iml @@ -37,15 +37,15 @@ - + - - + + diff --git a/Projects/Android/android.debug.keystore b/Projects/Android/android.debug.keystore new file mode 100644 index 0000000000000000000000000000000000000000..cdc70bfeb00af8065111e13c6027d0f90864b709 GIT binary patch literal 2159 zcmbtUX*kr49-aTpAlVyR*=5Pn_|H(1EWM2-Buk{S#Y_w`*_TL$8GFbQN!A+4yd-P( z>|By9rtu1cMg}E`G_L#J`{Dh5Kb#Ne{GR9coadb9Icyf21%W`I2LS$tz)fGQKLL-$ z;@pBfym4U%x)0jEDg*-IfPzFg8|3Fati}nz0ad6F0B}IyMEJ@+-B&lT#)oT?pmO=7 z;Zey2*%!{S8W3U4)tZjla|U}(|&Im-i)LpW{ylIu?Un%Kyx z)_XXi$<6sLq~uWhM2dF>5@+Jcg`*^Y#$0XQ^Ds+M_IygI)PEFMRZ?XiJdlDg|9b4O z{}nZEI_6!18%c@$sZ2Ny_Pz2%M!)(UvQstYTh_{Jn!WcCizXj+c`C%`&zkzbnDpO{ zpWs?=WwWcg)er4xJgp!OEn7>LcATr&zF#3>QOrkwoYu(s<2WURG08~^#u20Hfz^+V0+SqW zOwVK0x|$S*HwAOO;GabMjmJut?>Ikni+3}TJ|;5Ah|;qtD|a--sr2_A6UTEkw*;BoedexraR0Jw=alO;AdPX20L)Ei5fAnR*1+o@T z04;e7GS`>P)6dWXeP;> zolR&vx~r;qPUPYt0Cz&qtmCtgCk9r&yG9leDj&@`9W(UeIgE+*V{+`=#Xrg3&^K=O zGl??2u27hU294J3TUnGK#rz>eKYk^HF;Z+d8m0DQrtF%D|MZ6qK>BvPLnGXLj}>7Z zQ8*blz|LP7F^H7$>=ZeU_|>(O+v3eBI;+Sr!}4L0RziGlq%OJL)yU?|64lottjT1O zpmjDc9wMIIFmWg0o0Ivg$fSJ!$}K{eSjT0n+PvlblkIPYV+@97)Stzo$F1L&Q#T8z zZb!3RO8Wy$9H=)nauUw02!{TpE;>5;q2-pESY*JISjUsxHCH=khdVv9HBwHCrK(zo zRf=}aI@Dxl=8m|K-Nr;Nl1fdV@fU`Y23aB7)VSo5%fCX_yM_5KCIwjT`ZT;+zq(V{ zFej_l?jAGQhWX>}hku!zec*`{V!$9i(7cF^>u2B2E9Z<1CP>U4pAIR%2ALjPO`cy> zoar<^r}o~c5xpN&qwwObZMvTbo`&idjbnXF*x09X^@u||wTZH;eQDPnR#%6@FaTT# zD4;$hezc02x0L$zt&aNG7Vq7v;P%-t(xX>=ILOj{nY{ipw&A0sdzHhw+)_hirkCZm z)$m-#S|{3Xx}%)>$HFQ-C5hMRO-+3@of3F&T8UoQ6w4jca)`bBYrh6oE$g3_+qM2# zcPW_GPkgz@*Oy4S7eovGNC~|bYV&E)>YeOHSv}9xvfFjRn>h^2H(2P-f)`UiP=O}x zqswrO_bPzOG3=wZiUW!zCxuGG$_DbMAe7lCFvF*$kPu!`8!~dY!B#IXdTxZ|g+8MB za=4y8R4d-7DiXOk*3G}&;yfu~;Lb1@NExwcQqI9>lBv|L%5Iz5vpRl__E)ZShpl9s zFF5z(0ddf5tdz3fUN{gPnA_SzVfk;i9%pFJ{owKY`VjTxQx)j{FPAij4)>Zf2 z>Q#m53ZU&U?}P*KdxPaX`>qgp63%txj3ig3vZ*UJ-jN2DOy1m{k5WK;4s9{})hKtMf|I4H^qvxD;>1g>7R zJgI#Njk#iW#saiFjp7G+|8s#MIBfB^3BD*HP~gA(5Fyy#EeP*}1*OIKP&%MCS{u~X z1U0pVJrR)qr{RM=hYzFl6CBW?^ediRd{V&0e*W-PHcVRl#A*w2f|BmA@oIG9sx`qs!RwRTi*4b& z>3Wj>T6Nhp4eKZ-lh|0+jQnchotanWpVQFAv%hi7zk`blDLcSzxapk5hzXyTnM;~~ z)nc8SF8{TPT^Z(BjvUWIMFv$mR|Z<$EA~9H_4xA7px@7m>)Q%v+hTZn&I7ogSy61m zJ>;XC59i#X<+I2|6TjIhT$^ZbPsdID<(?F?*w$`!T!w2E=JcApu*%Vf!MPZDo!iBx_ z-bwhG>g|UL)zDhcX*&1PgYEtq;z8>NAp#ItXnMJarCJhF$;}zHw&xL9nJBoc((LpT zmEbv}dNHxn_OS^irTZ`gD;4>`IgND{m0L#~lT+y<+FVNq@|nb$E9sh>|}g!P~Tm>1ULB;1{ftY7jPDMm #include #include -#include #include #include #include @@ -35,23 +34,21 @@ Copyright : Copyright 2015 Oculus VR, LLC. All Rights reserved. #include "../darkplaces/quakedef.h" #include "../darkplaces/menu.h" -#include +#include "VrApi.h" +#include "VrApi_Helpers.h" +#include "VrApi_SystemUtils.h" +#include "VrApi_Input.h" +#include "VrApi_Types.h" + +#include "VrCompositor.h" + + +#include "VrCommon.h" #if !defined( EGL_OPENGL_ES3_BIT_KHR ) #define EGL_OPENGL_ES3_BIT_KHR 0x0040 #endif -#define PITCH 0 -#define YAW 1 -#define ROLL 2 - - -#define SOURCE_GAMEPAD 0x00000401 -#define SOURCE_JOYSTICK 0x01000010 - - -float maximumSupportedFramerate=60.0; //The lowest default framerate - // EXT_texture_border_clamp #ifndef GL_CLAMP_TO_BORDER #define GL_CLAMP_TO_BORDER 0x812D @@ -89,29 +86,13 @@ PFNEGLSIGNALSYNCKHRPROC eglSignalSyncKHR; PFNEGLGETSYNCATTRIBKHRPROC eglGetSyncAttribKHR; #endif -#include "VrApi.h" -#include "VrApi_Helpers.h" -#include "VrApi_SystemUtils.h" -#include "VrApi_Input.h" - -#ifndef NDEBUG -#define DEBUG 1 -#endif - -#define LOG_TAG "QuakeQuest" - -#define ALOGE(...) __android_log_print( ANDROID_LOG_ERROR, LOG_TAG, __VA_ARGS__ ) -#if DEBUG -#define ALOGV(...) __android_log_print( ANDROID_LOG_VERBOSE, LOG_TAG, __VA_ARGS__ ) -#else -#define ALOGV(...) -#endif - -int CPU_LEVEL = 2; -int GPU_LEVEL = 3; +//Let's go to the maximum! +int CPU_LEVEL = 4; +int GPU_LEVEL = 4; int NUM_MULTI_SAMPLES = 1; +float SS_MULTIPLIER = 1.2f; -float SS_MULTIPLIER = 1.0f; +float maximumSupportedFramerate=60.0; //The lowest default framerate extern float worldPosition[3]; float hmdPosition[3]; @@ -120,7 +101,7 @@ float positionDeltaThisFrame[3]; extern cvar_t vr_worldscale; extern cvar_t r_lasersight; -extern cvar_t cl_forwardspeed; +extern cvar_t cl_movementspeed; extern cvar_t cl_walkdirection; extern cvar_t cl_controllerdeadzone; extern cvar_t cl_righthanded; @@ -129,17 +110,19 @@ extern cvar_t slowmo; extern cvar_t bullettime; extern cvar_t cl_trackingmode; -extern int key_consoleactive; +ovrDeviceID controllerIDs[2]; -static float radians(float deg) { + +jclass clazz; + +float radians(float deg) { return (deg * M_PI) / 180.0; } -static float degrees(float rad) { +float degrees(float rad) { return (rad * 180.0) / M_PI; } - /* global arg_xxx structs */ struct arg_dbl *ss; struct arg_int *cpu; @@ -154,22 +137,22 @@ int argc=0; /* ================================================================================ -System Clock Time +System Clock Time in millis ================================================================================ */ -static double GetTimeInSeconds() +double GetTimeInMilliSeconds() { struct timespec now; clock_gettime( CLOCK_MONOTONIC, &now ); - return ( now.tv_sec * 1e9 + now.tv_nsec ) * 0.000000001; + return ( now.tv_sec * 1e9 + now.tv_nsec ) * (double)(1e-6); } /* ================================================================================ -QQUEST Stuff +QuakeQuest Stuff ================================================================================ */ @@ -267,23 +250,22 @@ long delta=0; int curtime; int Sys_Milliseconds (void) { - struct timeval tp; - struct timezone tzp; - static int secbase; + struct timeval tp; + struct timezone tzp; + static int secbase; - gettimeofday(&tp, &tzp); + gettimeofday(&tp, &tzp); - if (!secbase) - { - secbase = tp.tv_sec; - return tp.tv_usec/1000; - } + if (!secbase) + { + secbase = tp.tv_sec; + return tp.tv_usec/1000; + } - curtime = (tp.tv_sec - secbase)*1000 + tp.tv_usec/1000; + curtime = (tp.tv_sec - secbase)*1000 + tp.tv_usec/1000; - return curtime; + return curtime; } - int runStatus = -1; void QC_exit(int exitCode) { @@ -296,20 +278,20 @@ float weaponOffset[3]; float weaponVelocity[3]; float vrFOV; -float vrWidth; -float vrHeight; + +int hmdType; int bigScreen = 1; -ovrMatrix4f modelScreen; -ovrMatrix4f rotation; +extern client_static_t cls; + +qboolean useScreenLayer() +{ + //TODO + return (bigScreen != 0 || cls.demoplayback || key_consoleactive); +} void BigScreenMode(int mode) { - if (mode == 1 && key_consoleactive == 0) // bit of a hack, but only way rotation set when menu first invoked - { - rotation = ovrMatrix4f_CreateRotation( 0.0f, radians(hmdorientation[YAW]), 0.0f ); - } - if (bigScreen != 2) { bigScreen = mode; @@ -317,7 +299,6 @@ void BigScreenMode(int mode) } extern int stereoMode; -qboolean demoplayback; static void UnEscapeQuotes( char *arg ) { @@ -461,43 +442,6 @@ static const char * GlFrameBufferStatusString( GLenum status ) } } -#define CHECK_GL_ERRORS -#ifdef CHECK_GL_ERRORS - -static const char * GlErrorString( GLenum error ) -{ - switch ( error ) - { - case GL_NO_ERROR: return "GL_NO_ERROR"; - case GL_INVALID_ENUM: return "GL_INVALID_ENUM"; - case GL_INVALID_VALUE: return "GL_INVALID_VALUE"; - case GL_INVALID_OPERATION: return "GL_INVALID_OPERATION"; - case GL_INVALID_FRAMEBUFFER_OPERATION: return "GL_INVALID_FRAMEBUFFER_OPERATION"; - case GL_OUT_OF_MEMORY: return "GL_OUT_OF_MEMORY"; - default: return "unknown"; - } -} - -static void GLCheckErrors( int line ) -{ - for ( int i = 0; i < 10; i++ ) - { - const GLenum error = glGetError(); - if ( error == GL_NO_ERROR ) - { - break; - } - ALOGE( "GL error on line %d: %s", line, GlErrorString( error ) ); - } -} - -#define GL( func ) func; GLCheckErrors( __LINE__ ); - -#else // CHECK_GL_ERRORS - -#define GL( func ) func; - -#endif // CHECK_GL_ERRORS /* ================================================================================ @@ -685,17 +629,6 @@ ovrFramebuffer ================================================================================ */ -typedef struct -{ - int Width; - int Height; - int Multisamples; - int TextureSwapChainLength; - int TextureSwapChainIndex; - ovrTextureSwapChain * ColorTextureSwapChain; - GLuint * DepthBuffers; - GLuint * FrameBuffers; -} ovrFramebuffer; static void ovrFramebuffer_Clear( ovrFramebuffer * frameBuffer ) { @@ -721,7 +654,7 @@ static bool ovrFramebuffer_Create( ovrFramebuffer * frameBuffer, const GLenum co PFNGLFRAMEBUFFERTEXTUREMULTISAMPLEMULTIVIEWOVRPROC glFramebufferTextureMultisampleMultiviewOVR = (PFNGLFRAMEBUFFERTEXTUREMULTISAMPLEMULTIVIEWOVRPROC) eglGetProcAddress( "glFramebufferTextureMultisampleMultiviewOVR" ); - frameBuffer->Width = width; + frameBuffer->Width = width; frameBuffer->Height = height; frameBuffer->Multisamples = multisamples; @@ -745,10 +678,10 @@ static bool ovrFramebuffer_Create( ovrFramebuffer * frameBuffer, const GLenum co } else { - // Just clamp to edge. However, this requires manually clearing the border - // around the layer to clear the edge texels. - GL( glTexParameteri( colorTextureTarget, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE ) ); - GL( glTexParameteri( colorTextureTarget, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE ) ); + // Just clamp to edge. However, this requires manually clearing the border + // around the layer to clear the edge texels. + GL( glTexParameteri( colorTextureTarget, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE ) ); + GL( glTexParameteri( colorTextureTarget, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE ) ); } GL( glTexParameteri( colorTextureTarget, GL_TEXTURE_MIN_FILTER, GL_LINEAR ) ); GL( glTexParameteri( colorTextureTarget, GL_TEXTURE_MAG_FILTER, GL_LINEAR ) ); @@ -804,7 +737,7 @@ static bool ovrFramebuffer_Create( ovrFramebuffer * frameBuffer, const GLenum co return true; } -static void ovrFramebuffer_Destroy( ovrFramebuffer * frameBuffer ) +void ovrFramebuffer_Destroy( ovrFramebuffer * frameBuffer ) { GL( glDeleteFramebuffers( frameBuffer->TextureSwapChainLength, frameBuffer->FrameBuffers ) ); GL( glDeleteRenderbuffers( frameBuffer->TextureSwapChainLength, frameBuffer->DepthBuffers ) ); @@ -817,57 +750,58 @@ static void ovrFramebuffer_Destroy( ovrFramebuffer * frameBuffer ) ovrFramebuffer_Clear( frameBuffer ); } -static void ovrFramebuffer_SetCurrent( ovrFramebuffer * frameBuffer ) +void ovrFramebuffer_SetCurrent( ovrFramebuffer * frameBuffer ) { - GL( glBindFramebuffer( GL_DRAW_FRAMEBUFFER, frameBuffer->FrameBuffers[frameBuffer->TextureSwapChainIndex] ) ); + GL( glBindFramebuffer( GL_DRAW_FRAMEBUFFER, frameBuffer->FrameBuffers[frameBuffer->TextureSwapChainIndex] ) ); } -static void ovrFramebuffer_SetNone() +void ovrFramebuffer_SetNone() { - GL( glBindFramebuffer( GL_DRAW_FRAMEBUFFER, 0 ) ); + GL( glBindFramebuffer( GL_DRAW_FRAMEBUFFER, 0 ) ); } -static void ovrFramebuffer_Resolve( ovrFramebuffer * frameBuffer ) +void ovrFramebuffer_Resolve( ovrFramebuffer * frameBuffer ) { // Discard the depth buffer, so the tiler won't need to write it back out to memory. const GLenum depthAttachment[1] = { GL_DEPTH_ATTACHMENT }; glInvalidateFramebuffer( GL_DRAW_FRAMEBUFFER, 1, depthAttachment ); - // Flush this frame worth of commands. - glFlush(); + // Flush this frame worth of commands. + glFlush(); } -static void ovrFramebuffer_Advance( ovrFramebuffer * frameBuffer ) +void ovrFramebuffer_Advance( ovrFramebuffer * frameBuffer ) { // Advance to the next texture from the set. frameBuffer->TextureSwapChainIndex = ( frameBuffer->TextureSwapChainIndex + 1 ) % frameBuffer->TextureSwapChainLength; } -static void ovrFramebuffer_ClearEdgeTexels( ovrFramebuffer * frameBuffer ) +void ovrFramebuffer_ClearEdgeTexels( ovrFramebuffer * frameBuffer ) { - GL( glEnable( GL_SCISSOR_TEST ) ); - GL( glViewport( 0, 0, frameBuffer->Width, frameBuffer->Height ) ); + GL( glEnable( GL_SCISSOR_TEST ) ); + GL( glViewport( 0, 0, frameBuffer->Width, frameBuffer->Height ) ); - // Explicitly clear the border texels to black because OpenGL-ES does not support GL_CLAMP_TO_BORDER. - // Clear to fully opaque black. - GL( glClearColor( 0.0f, 0.0f, 0.0f, 1.0f ) ); + // Explicitly clear the border texels to black because OpenGL-ES does not support GL_CLAMP_TO_BORDER. + // Clear to fully opaque black. + GL( glClearColor( 0.0f, 0.0f, 0.0f, 1.0f ) ); - // bottom - GL( glScissor( 0, 0, frameBuffer->Width, 1 ) ); - GL( glClear( GL_COLOR_BUFFER_BIT ) ); - // top - GL( glScissor( 0, frameBuffer->Height - 1, frameBuffer->Width, 1 ) ); - GL( glClear( GL_COLOR_BUFFER_BIT ) ); - // left - GL( glScissor( 0, 0, 1, frameBuffer->Height ) ); - GL( glClear( GL_COLOR_BUFFER_BIT ) ); - // right - GL( glScissor( frameBuffer->Width - 1, 0, 1, frameBuffer->Height ) ); - GL( glClear( GL_COLOR_BUFFER_BIT ) ); + // bottom + GL( glScissor( 0, 0, frameBuffer->Width, 1 ) ); + GL( glClear( GL_COLOR_BUFFER_BIT ) ); + // top + GL( glScissor( 0, frameBuffer->Height - 1, frameBuffer->Width, 1 ) ); + GL( glClear( GL_COLOR_BUFFER_BIT ) ); + // left + GL( glScissor( 0, 0, 1, frameBuffer->Height ) ); + GL( glClear( GL_COLOR_BUFFER_BIT ) ); + // right + GL( glScissor( frameBuffer->Width - 1, 0, 1, frameBuffer->Height ) ); + GL( glClear( GL_COLOR_BUFFER_BIT ) ); - GL( glScissor( 0, 0, 0, 0 ) ); - GL( glDisable( GL_SCISSOR_TEST ) ); + + GL( glScissor( 0, 0, 0, 0 ) ); + GL( glDisable( GL_SCISSOR_TEST ) ); } @@ -880,171 +814,41 @@ ovrRenderer */ -typedef struct -{ - ovrFramebuffer FrameBuffer[VRAPI_FRAME_LAYER_EYE_MAX]; - ovrFramebuffer QuakeFrameBuffer; - ovrMatrix4f ProjectionMatrix; - int NumBuffers; -} ovrRenderer; - -static void ovrRenderer_Clear( ovrRenderer * renderer ) +void ovrRenderer_Clear( ovrRenderer * renderer ) { for ( int eye = 0; eye < VRAPI_FRAME_LAYER_EYE_MAX; eye++ ) { ovrFramebuffer_Clear( &renderer->FrameBuffer[eye] ); } - ovrFramebuffer_Clear( &renderer->QuakeFrameBuffer ); - renderer->ProjectionMatrix = ovrMatrix4f_CreateIdentity(); + renderer->NumBuffers = VRAPI_FRAME_LAYER_EYE_MAX; } -static const char VERTEX_SHADER[] = - "uniform mat4 u_MVPMatrix;" - "attribute vec4 a_Position;" - "attribute vec2 a_texCoord;" - "varying vec2 v_texCoord;" - "void main() {" - " gl_Position = u_MVPMatrix * a_Position;" - " v_texCoord = a_texCoord;" - "}"; - - -static const char FRAGMENT_SHADER[] = - "precision mediump float;" - "varying vec2 v_texCoord;" - "uniform sampler2D s_texture;" - "uniform sampler2D s_texture1;" - "void main() {" - " gl_FragColor = texture2D( s_texture, v_texCoord ) * texture2D( s_texture1, v_texCoord );" - "}"; - -int loadShader(int type, const char * shaderCode){ - int shader = glCreateShader(type); - GLint length = 0; - GL( glShaderSource(shader, 1, &shaderCode, 0)); - GL( glCompileShader(shader)); - return shader; -} - -int BuildScreenVignetteTexture( ) -{ - static const int scale = 8; - static const int width = 10 * scale; - static const int height = 8 * scale; - unsigned char buffer[width * height]; - memset( buffer, 255, sizeof( buffer ) ); - for ( int i = 0; i < width; i++ ) - { - buffer[i] = 0; - buffer[width*height - 1 - i] = 0; - } - for ( int i = 0; i < height; i++ ) - { - buffer[i * width] = 0; - buffer[i * width + width - 1] = 0; - } - GLuint texId = 0; - glGenTextures( 1, &texId ); - glBindTexture( GL_TEXTURE_2D, texId ); - glTexImage2D( GL_TEXTURE_2D, 0, GL_LUMINANCE, width, height, 0, GL_LUMINANCE, GL_UNSIGNED_BYTE, buffer ); - glTexParameteri( GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE ); - glTexParameteri( GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE ); - glTexParameteri( GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR ); - glTexParameteri( GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR ); - glBindTexture( GL_TEXTURE_2D, 0 ); - - return texId; -} - -static short indices[6] = {0, 1, 2, 0, 2, 3}; - -static float uvs[8] = { - 0.0f, 1.0f, - 0.0f, 0.0f, - 1.0f, 0.0f, - 1.0f, 1.0f -}; - -static float SCREEN_COORDS[12] = { - -1.0f, 0.75f, 0.0f, - -1.0f, -0.75f, 0.0f, - 1.0f, -0.75f, 0.0f, - 1.0f, 0.75f, 0.0f -}; - -int vignetteTexture = 0; - -int sp_Image = 0; -int positionParam = 0; -int texCoordParam = 0; -int samplerParam = 0; -int vignetteParam = 0; -int modelViewProjectionParam = 0; - -static void ovrRenderer_Create( ovrRenderer * renderer, const ovrJava * java ) +void ovrRenderer_Create( int width, int height, ovrRenderer * renderer, const ovrJava * java ) { renderer->NumBuffers = VRAPI_FRAME_LAYER_EYE_MAX; - // Create the shaders, images - int vertexShader = loadShader(GL_VERTEX_SHADER, VERTEX_SHADER); - int fragmentShader = loadShader(GL_FRAGMENT_SHADER, FRAGMENT_SHADER); - - - sp_Image = glCreateProgram(); // create empty OpenGL ES Program - GL( glAttachShader(sp_Image, vertexShader)); // add the vertex shader to program - GL( glAttachShader(sp_Image, fragmentShader)); // add the fragment shader to program - GL( glLinkProgram(sp_Image)); // creates OpenGL ES program executable - positionParam = GL( glGetAttribLocation(sp_Image, "a_Position")); - texCoordParam = GL( glGetAttribLocation(sp_Image, "a_texCoord")); - modelViewProjectionParam = GL( glGetUniformLocation(sp_Image, "u_MVPMatrix")); - samplerParam = GL( glGetUniformLocation(sp_Image, "s_texture")); - vignetteParam = GL( glGetUniformLocation(sp_Image, "s_texture1")); - - vignetteTexture = BuildScreenVignetteTexture(); - - modelScreen = ovrMatrix4f_CreateIdentity(); - rotation = ovrMatrix4f_CreateIdentity(); - ovrMatrix4f translation = ovrMatrix4f_CreateTranslation( 0, 0, -1.5f ); - modelScreen = ovrMatrix4f_Multiply( &modelScreen, &translation ); - - vrFOV = vrapi_GetSystemPropertyInt( java, VRAPI_SYS_PROP_SUGGESTED_EYE_FOV_DEGREES_Y); - // Create the render Textures. for ( int eye = 0; eye < VRAPI_FRAME_LAYER_EYE_MAX; eye++ ) { ovrFramebuffer_Create( &renderer->FrameBuffer[eye], GL_RGBA8, - (int)vrWidth, - (int)vrHeight, + width, + height, NUM_MULTI_SAMPLES ); } - - ovrFramebuffer_Create( &renderer->QuakeFrameBuffer, - GL_RGBA8, - (int)vrWidth, - (int)vrHeight, - NUM_MULTI_SAMPLES ); - - // Setup the projection matrix. - renderer->ProjectionMatrix = ovrMatrix4f_CreateProjectionFov( - vrFOV, - vrFOV, - 0.0f, 0.0f, 1.0f, 0.0f ); - } -static void ovrRenderer_Destroy( ovrRenderer * renderer ) +void ovrRenderer_Destroy( ovrRenderer * renderer ) { for ( int eye = 0; eye < renderer->NumBuffers; eye++ ) { ovrFramebuffer_Destroy( &renderer->FrameBuffer[eye] ); } - ovrFramebuffer_Destroy( &renderer->QuakeFrameBuffer ); - renderer->ProjectionMatrix = ovrMatrix4f_CreateIdentity(); } + #ifndef EPSILON #define EPSILON 0.001f #endif @@ -1067,10 +871,10 @@ static ovrVector3f normalizeVec(ovrVector3f vec) { void NormalizeAngles(vec3_t angles) { - while (angles[0] >= 90) angles[0] -= 360; + while (angles[0] >= 90) angles[0] -= 180; while (angles[1] >= 180) angles[1] -= 360; while (angles[2] >= 180) angles[2] -= 360; - while (angles[0] < -90) angles[0] += 360; + while (angles[0] < -90) angles[0] += 180; while (angles[1] < -180) angles[1] += 360; while (angles[2] < -180) angles[2] += 360; } @@ -1129,10 +933,10 @@ void QuatToYawPitchRoll(ovrQuatf q, float pitchAdjust, vec3_t out) { ovrMatrix4f mat = ovrMatrix4f_CreateFromQuaternion( &q ); if (pitchAdjust != 0.0f) - { - ovrMatrix4f rot = ovrMatrix4f_CreateRotation(radians(pitchAdjust), 0.0f, 0.0f); - mat = ovrMatrix4f_Multiply(&mat, &rot); - } + { + ovrMatrix4f rot = ovrMatrix4f_CreateRotation(radians(pitchAdjust), 0.0f, 0.0f); + mat = ovrMatrix4f_Multiply(&mat, &rot); + } ovrVector4f v1 = {0, 0, -1, 0}; ovrVector4f v2 = {1, 0, 0, 0}; @@ -1154,15 +958,6 @@ void QuatToYawPitchRoll(ovrQuatf q, float pitchAdjust, vec3_t out) { return; } -static void adjustYaw(float adjustment, float* out) -{ - *out -= adjustment; - if (*out > 180.0f) - *out -= 360.0f; - if (*out < -180.0f) - *out += 360.0f; -} - void setWorldPosition( float x, float y, float z ) { positionDeltaThisFrame[0] = (worldPosition[0] - x); @@ -1174,18 +969,56 @@ void setWorldPosition( float x, float y, float z ) worldPosition[2] = z; } -void setHMDPosition( float x, float y, float z ) +extern float playerYaw; +void setHMDPosition( float x, float y, float z, float yaw ) { - hmdPosition[0] = x; - hmdPosition[1] = y; - if (bigScreen) + static bool s_useScreen = false; + + VectorSet(hmdPosition, x, y, z); + + if (s_useScreen != useScreenLayer()) { + s_useScreen = useScreenLayer(); + + //Record player height on transition playerHeight = y; } - hmdPosition[2] = z; + + if (!useScreenLayer()) + { + playerYaw = yaw; + } } -static ovrLayerProjection2 ovrRenderer_RenderFrame( ovrRenderer * renderer, const ovrJava * java, +qboolean isMultiplayer() +{ + return Cvar_VariableValue("maxclients") > 1; +} + + +/* +======================== +Android_Vibrate +======================== +*/ + +//0 = left, 1 = right +float vibration_channel_duration[2] = {0.0f, 0.0f}; +float vibration_channel_intensity[2] = {0.0f, 0.0f}; + +void Android_Vibrate( float duration, int channel, float intensity ) +{ + if (vibration_channel_duration[channel] > 0.0f) + return; + + if (vibration_channel_duration[channel] == -1.0f && duration != 0.0f) + return; + + vibration_channel_duration[channel] = duration; + vibration_channel_intensity[channel] = intensity; +} + +static void ovrRenderer_RenderFrame( ovrRenderer * renderer, const ovrJava * java, const ovrTracking2 * tracking, ovrMobile * ovr ) { ovrTracking2 updatedTracking = *tracking; @@ -1196,7 +1029,7 @@ static ovrLayerProjection2 ovrRenderer_RenderFrame( ovrRenderer * renderer, cons const ovrQuatf quatHmd = tracking->HeadPose.Pose.Orientation; const ovrVector3f positionHmd = tracking->HeadPose.Pose.Position; QuatToYawPitchRoll(quatHmd, 0.0f, hmdorientation); - setHMDPosition(positionHmd.x, positionHmd.y, positionHmd.z); + setHMDPosition(positionHmd.x, positionHmd.y, positionHmd.z, hmdorientation[YAW]); //Use hmd position for world position setWorldPosition(positionHmd.x, positionHmd.y, positionHmd.z); @@ -1213,43 +1046,14 @@ static ovrLayerProjection2 ovrRenderer_RenderFrame( ovrRenderer * renderer, cons //Set everything up QC_BeginFrame(/* true to stop time if needed in future */ false); - - ovrLayerProjection2 layer = vrapi_DefaultLayerProjection2(); - layer.HeadPose = updatedTracking.HeadPose; - for ( int eye = 0; eye < VRAPI_FRAME_LAYER_EYE_MAX; eye++ ) - { - ovrFramebuffer * frameBuffer = &renderer->FrameBuffer[renderer->NumBuffers == 1 ? 0 : eye]; - layer.Textures[eye].ColorSwapChain = frameBuffer->ColorTextureSwapChain; - layer.Textures[eye].SwapChainIndex = frameBuffer->TextureSwapChainIndex; - - ovrMatrix4f projectionMatrix; - projectionMatrix = ovrMatrix4f_CreateProjectionFov(vrFOV, vrFOV, - 0.0f, 0.0f, 0.1f, 0.0f); - - layer.Textures[eye].TexCoordsFromTanAngles = ovrMatrix4f_TanAngleMatrixFromProjection(&projectionMatrix); - - layer.Textures[eye].TextureRect.x = 0; - layer.Textures[eye].TextureRect.y = 0; - layer.Textures[eye].TextureRect.width = 1.0f; - layer.Textures[eye].TextureRect.height = 1.0f; - } - layer.Header.Flags |= VRAPI_FRAME_LAYER_FLAG_CHROMATIC_ABERRATION_CORRECTION; - // Render the eye images. for ( int eye = 0; eye < renderer->NumBuffers; eye++ ) { ovrFramebuffer * frameBuffer = &renderer->FrameBuffer[eye]; - if (bigScreen != 0 || demoplayback) - frameBuffer = &renderer->QuakeFrameBuffer; ovrFramebuffer_SetCurrent( frameBuffer ); //If showing the menu, then render mono, easier to navigate menu - if (eye > 0 && m_state != m_none) - { - //Do nothing, we just use the left eye again, render in Mono - } - else { GL( glEnable( GL_SCISSOR_TEST ) ); GL( glDepthMask( GL_TRUE ) ); @@ -1271,72 +1075,7 @@ static ovrLayerProjection2 ovrRenderer_RenderFrame( ovrRenderer * renderer, cons //Clear edge to prevent smearing ovrFramebuffer_ClearEdgeTexels( frameBuffer ); - if (bigScreen != 0 || demoplayback || key_consoleactive) - { - frameBuffer = &renderer->FrameBuffer[eye]; - ovrFramebuffer_SetCurrent( frameBuffer ); - // Apply the projection and view transformation - ovrMatrix4f modelView = ovrMatrix4f_Multiply( &updatedTracking.Eye[0].ViewMatrix, &rotation ); - modelView = ovrMatrix4f_Multiply( &modelView, &modelScreen ); - ovrMatrix4f modelViewProjection = ovrMatrix4f_Multiply( &renderer->ProjectionMatrix, &modelView ); - - GLint originalTex0 = 0; - GLint originalTex1 = 0; - GL( glEnable( GL_SCISSOR_TEST ) ); - GL( glDepthMask( GL_TRUE ) ); - GL( glEnable( GL_DEPTH_TEST ) ); - GL( glDepthFunc( GL_LEQUAL ) ); - GL( glViewport( 0, 0, frameBuffer->Width, frameBuffer->Height ) ); - GL( glScissor( 0, 0, frameBuffer->Width, frameBuffer->Height ) ); - GL( glClearColor( 0.01f, 0.0f, 0.0f, 1.0f ) ); - GL( glClear( GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT ) ); - GL( glDisable(GL_SCISSOR_TEST)); - - GL( glUseProgram(sp_Image)); - - // Set the position of the screen - GL( glVertexAttribPointer(positionParam, 3, GL_FLOAT, GL_FALSE, 0, SCREEN_COORDS)); - - // Prepare the texture coordinates - GL( glVertexAttribPointer(texCoordParam, 2, GL_FLOAT, GL_FALSE, 0, uvs) ); - - // Apply the projection and view transformation - modelView = ovrMatrix4f_Multiply( &modelView, &modelScreen ); - - GL( glUniformMatrix4fv(modelViewProjectionParam, 1, GL_TRUE, (const GLfloat *)modelViewProjection.M[0]) ); - - GL( glActiveTexture(GL_TEXTURE0) ); - - GLuint colorTexture = vrapi_GetTextureSwapChainHandle( - renderer->QuakeFrameBuffer.ColorTextureSwapChain, - renderer->QuakeFrameBuffer.TextureSwapChainIndex); - glGetIntegerv(GL_TEXTURE_BINDING_2D, &originalTex0); - GL( glBindTexture( GL_TEXTURE_2D, colorTexture ) ); - - // Set the sampler texture unit to our fbo's color texture - GL( glUniform1i(samplerParam, 0) ); - - //Bind vignette texture to texture 1 - GL( glActiveTexture( GL_TEXTURE1 ) ); - glGetIntegerv(GL_TEXTURE_BINDING_2D, &originalTex1); - GL( glBindTexture( GL_TEXTURE_2D, vignetteTexture) ); - GL( glUniform1i(vignetteParam, 1) ); - - // Draw the triangles - GL( glDrawElements(GL_TRIANGLES, 6, GL_UNSIGNED_SHORT, indices) ); - - - //Restore previously active textures - GL( glBindTexture(GL_TEXTURE_2D, originalTex1) ); - GL( glActiveTexture(GL_TEXTURE0) ); - GL( glBindTexture(GL_TEXTURE_2D, originalTex0) ); - - - //Clear edge to prevent smearing - ovrFramebuffer_ClearEdgeTexels( frameBuffer ); - } - ovrFramebuffer_Resolve( frameBuffer ); ovrFramebuffer_Advance( frameBuffer ); } @@ -1344,10 +1083,9 @@ static ovrLayerProjection2 ovrRenderer_RenderFrame( ovrRenderer * renderer, cons QC_EndFrame(); ovrFramebuffer_SetNone(); - - return layer; } + /* ================================================================================ @@ -1372,6 +1110,7 @@ typedef struct ANativeWindow * NativeWindow; bool Resumed; ovrMobile * Ovr; + ovrScene Scene; long long FrameIndex; double DisplayTime; int SwapInterval; @@ -1379,6 +1118,8 @@ typedef struct int GpuLevel; int MainThreadTid; int RenderThreadTid; + ovrLayer_Union2 Layers[ovrMaxLayerCount]; + int LayerCount; ovrRenderer Renderer; } ovrApp; @@ -1387,8 +1128,6 @@ static void ovrApp_Clear( ovrApp * app ) app->Java.Vm = NULL; app->Java.Env = NULL; app->Java.ActivityObject = NULL; - app->NativeWindow = NULL; - app->Resumed = false; app->Ovr = NULL; app->FrameIndex = 1; app->DisplayTime = 0; @@ -1400,6 +1139,7 @@ static void ovrApp_Clear( ovrApp * app ) ovrEgl_Clear( &app->Egl ); + ovrScene_Clear( &app->Scene ); ovrRenderer_Clear( &app->Renderer ); } @@ -1666,6 +1406,105 @@ float length(float x, float y) return sqrtf(powf(x, 2.0f) + powf(y, 2.0f)); } +static void weaponHaptics() +{ + if (useScreenLayer()) + { + return; + } + + qboolean isFirePressed = false; + if (cl_righthanded.integer) { + //Fire + isFirePressed = rightTrackedRemoteState_new.Buttons & ovrButton_Trigger; + } else { + isFirePressed = leftTrackedRemoteState_new.Buttons & ovrButton_Trigger; + } + + if (isFirePressed) + { + static double timeLastHaptic = 0; + double timeNow = GetTimeInMilliSeconds(); + + float hapticInterval = 0; + float hapticLevel = 0; + float hapticLength = 0; + + switch (cl.stats[STAT_ACTIVEWEAPON]) + { + case IT_SHOTGUN: + { + hapticInterval = 500; + hapticLevel = 0.7f; + hapticLength = 150; + } + break; + case IT_SUPER_SHOTGUN: + { + hapticInterval = 700; + hapticLevel = 0.8f; + hapticLength = 200; + } + break; + case IT_NAILGUN: + { + hapticInterval = 100; + hapticLevel = 0.6f; + hapticLength = 50; + } + break; + case IT_SUPER_NAILGUN: + { + hapticInterval = 80; + hapticLevel = 0.9f; + hapticLength = 50; + } + break; + case IT_GRENADE_LAUNCHER: + { + hapticInterval = 600; + hapticLevel = 0.7f; + hapticLength = 100; + } + break; + case IT_ROCKET_LAUNCHER: + { + hapticInterval = 800; + hapticLevel = 1.0f; + hapticLength = 300; + } + break; + case IT_LIGHTNING: + { + hapticInterval = 10; + hapticLevel = lhrandom(0.0, 0.8f); + hapticLength = 10; + } + break; + case IT_SUPER_LIGHTNING: + { + hapticInterval = 10; + hapticLevel = lhrandom(0.3, 1.0f); + hapticLength = 10; + } + break; + case IT_AXE: + { + hapticInterval = 500; + hapticLevel = 0.6f; + hapticLength = 100; + } + break; + } + + if ((timeNow - timeLastHaptic) > hapticInterval) + { + timeLastHaptic = timeNow; + Android_Vibrate(hapticLength, cl_righthanded.integer ? 1 : 0, hapticLevel); + } + } +} + static void ovrApp_HandleInput( ovrApp * app ) { float remote_movementSideways = 0.0f; @@ -1701,9 +1540,11 @@ static void ovrApp_HandleInput( ovrApp * app ) if (remoteCapabilities.ControllerCapabilities & ovrControllerCaps_RightHand) { rightTrackedRemoteState_new = trackedRemoteState; rightRemoteTracking = remoteTracking; + controllerIDs[1] = cap.DeviceID; } else{ leftTrackedRemoteState_new = trackedRemoteState; leftRemoteTracking = remoteTracking; + controllerIDs[0] = cap.DeviceID; } } } @@ -1787,9 +1628,11 @@ static void ovrApp_HandleInput( ovrApp * app ) right_char, right_char); } - //Menu button + //Menu button - could be on left or right controller handleTrackedControllerButton(&leftTrackedRemoteState_new, &leftTrackedRemoteState_old, ovrButton_Enter, K_ESCAPE); + handleTrackedControllerButton(&rightTrackedRemoteState_new, &rightTrackedRemoteState_old, + ovrButton_Enter, K_ESCAPE); if (textInput) { @@ -1850,20 +1693,6 @@ static void ovrApp_HandleInput( ovrApp * app ) controllerYawHeading = controllerAngles[YAW] - gunangles[YAW] + yawOffset; hmdYawHeading = hmdorientation[YAW] - gunangles[YAW] + yawOffset; - -#ifndef NDEBUG - //Change heading mode on click of off=hand joystick - if ((offHandTrackedRemoteState->Buttons & ovrButton_Joystick) && bigScreen == 0 && - (offHandTrackedRemoteState->Buttons & ovrButton_Joystick) != - (offHandTrackedRemoteStateOld->Buttons & ovrButton_Joystick)) { - Cvar_SetValueQuick(&cl_walkdirection, 1 - cl_walkdirection.integer); - if (cl_walkdirection.integer == 1) { - SCR_CenterPrint("Heading Mode: Direction of HMD"); - } else { - SCR_CenterPrint("Heading Mode: Direction of off-hand controller"); - } - } -#endif } //Right-hand specific stuff @@ -1876,7 +1705,7 @@ static void ovrApp_HandleInput( ovrApp * app ) //This section corrects for the fact that the controller actually controls direction of movement, but we want to move relative to the direction the //player is facing for positional tracking float multiplier = /*arbitrary value that works ->*/ - 2300.0f / (cl_forwardspeed.value * ((offHandTrackedRemoteState->Buttons & ovrButton_Trigger) ? cl_movespeedkey.value : 1.0f)); + 2300.0f / (cl_movementspeed.value * ((offHandTrackedRemoteState->Buttons & ovrButton_Trigger) ? cl_movespeedkey.value : 1.0f)); vec2_t v; rotateAboutOrigin(-positionDeltaThisFrame[0] * multiplier, @@ -1928,13 +1757,20 @@ static void ovrApp_HandleInput( ovrApp * app ) if ((rightTrackedRemoteState_new.Buttons & ovrButton_B) && (rightTrackedRemoteState_new.Buttons & ovrButton_B) != (rightTrackedRemoteState_old.Buttons & ovrButton_B)) { - float newPitchAdjust = vr_weaponpitchadjust.value + 5.0f; - if (newPitchAdjust > 10.0f) { - newPitchAdjust = -40.0f; - } - Cvar_SetValueQuick(&vr_weaponpitchadjust, newPitchAdjust); - ALOGV("Pitch Adjust: %f", newPitchAdjust); + //Unused + } + + { + //Weapon/Inventory Chooser + int rightJoyState = (rightTrackedRemoteState_new.Joystick.y < -0.7f ? 1 : 0); + if (rightJoyState != (rightTrackedRemoteState_old.Joystick.y < -0.7f ? 1 : 0)) { + QC_KeyEvent(rightJoyState, '/', 0); + } + rightJoyState = (rightTrackedRemoteState_new.Joystick.y > 0.7f ? 1 : 0); + if (rightJoyState != (rightTrackedRemoteState_old.Joystick.y > 0.7f ? 1 : 0)) { + QC_KeyEvent(rightJoyState, '#', 0); + } } } @@ -1950,10 +1786,6 @@ static void ovrApp_HandleInput( ovrApp * app ) ovrButton_Trigger, K_SHIFT); } - //Next Weapon - handleTrackedControllerButton(&rightTrackedRemoteState_new, - &rightTrackedRemoteState_old, ovrButton_GripTrigger, '/'); - rightTrackedRemoteState_old = rightTrackedRemoteState_new; } @@ -2017,10 +1849,6 @@ static void ovrApp_HandleInput( ovrApp * app ) ovrButton_Trigger, K_MOUSE1); } - //Prev Weapon - handleTrackedControllerButton(&leftTrackedRemoteState_new, &leftTrackedRemoteState_old, - ovrButton_GripTrigger, '#'); - #ifndef NDEBUG //Give all weapons and all ammo and god mode if ((leftTrackedRemoteState_new.Buttons & ovrButton_X) && @@ -2263,14 +2091,41 @@ typedef struct { JavaVM * JavaVm; jobject ActivityObject; + jclass ActivityClass; pthread_t Thread; ovrMessageQueue MessageQueue; ANativeWindow * NativeWindow; } ovrAppThread; long shutdownCountdown; + +int m_width; +int m_height; static ovrJava java; +qboolean R_SetMode( void ); + +void Quest_GetScreenRes(int *width, int *height) +{ + *width = m_width; + *height = m_height; +} + +int Quest_GetRefresh() +{ + return vrapi_GetSystemPropertyInt( &java, VRAPI_SYS_PROP_DISPLAY_REFRESH_RATE ); +} + +float getFOV() +{ + return vrapi_GetSystemPropertyFloat( &java, VRAPI_SYS_PROP_SUGGESTED_EYE_FOV_DEGREES_Y ); +} + +void Quest_MessageBox(const char *title, const char *text) +{ + ALOGE("%s %s", title, text); +} + float GetFOV() { return vrapi_GetSystemPropertyFloat( &java, VRAPI_SYS_PROP_SUGGESTED_EYE_FOV_DEGREES_Y ); @@ -2307,12 +2162,9 @@ void * AppThreadFunction( void * parm ) // This app will handle android gamepad events itself. vrapi_SetPropertyInt( &appState.Java, VRAPI_EAT_NATIVE_GAMEPAD_EVENTS, 0 ); - //Using a symmetrical render target - vrWidth=vrapi_GetSystemPropertyInt( &java, VRAPI_SYS_PROP_SUGGESTED_EYE_TEXTURE_WIDTH ) * SS_MULTIPLIER; - vrHeight=vrWidth; - - - + //Using a symmetrical render target + m_width=vrapi_GetSystemPropertyInt( &java, VRAPI_SYS_PROP_SUGGESTED_EYE_TEXTURE_WIDTH ) * SS_MULTIPLIER; + m_height=m_width; //AmmarkoV : Query Refresh rates and select maximum..! //----------------------------------------------------------------------------------------------------------- int numberOfRefreshRates = vrapi_GetSystemPropertyInt(&java,VRAPI_SYS_PROP_NUM_SUPPORTED_DISPLAY_REFRESH_RATES); @@ -2320,7 +2172,7 @@ void * AppThreadFunction( void * parm ) if (numberOfRefreshRates > 16 ) { numberOfRefreshRates = 16; } vrapi_GetSystemPropertyFloatArray(&java, VRAPI_SYS_PROP_SUPPORTED_DISPLAY_REFRESH_RATES,&refreshRatesArray[0], numberOfRefreshRates); for (int i = 0; i < numberOfRefreshRates; i++) { - ALOGV("Supported refresh rate : %s Hz", refreshRatesArray[i]); + //ALOGV("Supported refresh rate : %s Hz", refreshRatesArray[i]); if (maximumSupportedFramerate 0.0f || + vibration_channel_duration[i] == -1.0f) { + vrapi_SetHapticVibrationSimple(appState.Ovr, controllerIDs[i], + vibration_channel_intensity[i]); + + if (vibration_channel_duration[i] != -1.0f) { + vibration_channel_duration[i] -= frametime; + + if (vibration_channel_duration[i] < 0.0f) { + vibration_channel_duration[i] = 0.0f; + vibration_channel_intensity[i] = 0.0f; + } + } + } else { + vrapi_SetHapticVibrationSimple(appState.Ovr, controllerIDs[i], 0.0f); + } + } + if (runStatus == -1) { #ifndef NDEBUG if (appState.FrameIndex > 10800) @@ -2471,40 +2355,112 @@ void * AppThreadFunction( void * parm ) runStatus = 0; } #endif - // Get the HMD pose, predicted for the middle of the time period during which - // the new eye images will be displayed. The number of frames predicted ahead - // depends on the pipeline depth of the engine and the synthesis rate. - // The better the prediction, the less black will be pulled in at the edges. - const double predictedDisplayTime = vrapi_GetPredictedDisplayTime(appState.Ovr, - appState.FrameIndex); - const ovrTracking2 tracking = vrapi_GetPredictedTracking2(appState.Ovr, - predictedDisplayTime); - appState.DisplayTime = predictedDisplayTime; + //Set 90hz mode for Quest 2 + if (hmdType == VRAPI_DEVICE_TYPE_OCULUSQUEST2) { + vrapi_SetDisplayRefreshRate(appState.Ovr, 90); + } + + // Get the HMD pose, predicted for the middle of the time period during which + // the new eye images will be displayed. The number of frames predicted ahead + // depends on the pipeline depth of the engine and the synthesis rate. + // The better the prediction, the less black will be pulled in at the edges. + const double predictedDisplayTime = vrapi_GetPredictedDisplayTime(appState.Ovr, + appState.FrameIndex); + const ovrTracking2 tracking = vrapi_GetPredictedTracking2(appState.Ovr, + predictedDisplayTime); + + appState.DisplayTime = predictedDisplayTime; ovrApp_HandleInput(&appState); + weaponHaptics(); - // Render eye images and setup the primary layer using ovrTracking2. - const ovrLayerProjection2 worldLayer = ovrRenderer_RenderFrame(&appState.Renderer, - &appState.Java, - &tracking, appState.Ovr); + ovrSubmitFrameDescription2 frameDesc = { 0 }; + if (!useScreenLayer()) { - const ovrLayerHeader2 *layers[] = - { - &worldLayer.Header - }; + ovrLayerProjection2 layer = vrapi_DefaultLayerProjection2(); + layer.HeadPose = tracking.HeadPose; + for ( int eye = 0; eye < VRAPI_FRAME_LAYER_EYE_MAX; eye++ ) + { + ovrFramebuffer * frameBuffer = &appState.Renderer.FrameBuffer[appState.Renderer.NumBuffers == 1 ? 0 : eye]; + layer.Textures[eye].ColorSwapChain = frameBuffer->ColorTextureSwapChain; + layer.Textures[eye].SwapChainIndex = frameBuffer->TextureSwapChainIndex; - ovrSubmitFrameDescription2 frameDesc = {}; - frameDesc.Flags = 0; - frameDesc.SwapInterval = appState.SwapInterval; - frameDesc.FrameIndex = appState.FrameIndex; - frameDesc.DisplayTime = appState.DisplayTime; - frameDesc.LayerCount = 1; - frameDesc.Layers = layers; + ovrMatrix4f projectionMatrix; + float fov = getFOV(); + projectionMatrix = ovrMatrix4f_CreateProjectionFov(fov, fov, + 0.0f, 0.0f, 0.1f, 0.0f); - // Hand over the eye images to the time warp. - vrapi_SubmitFrame2(appState.Ovr, &frameDesc); + layer.Textures[eye].TexCoordsFromTanAngles = ovrMatrix4f_TanAngleMatrixFromProjection(&projectionMatrix); + + layer.Textures[eye].TextureRect.x = 0; + layer.Textures[eye].TextureRect.y = 0; + layer.Textures[eye].TextureRect.width = 1.0f; + layer.Textures[eye].TextureRect.height = 1.0f; + } + layer.Header.Flags |= VRAPI_FRAME_LAYER_FLAG_CHROMATIC_ABERRATION_CORRECTION; + + //Call the game drawing code to populate the cylinder layer texture + ovrRenderer_RenderFrame(&appState.Renderer, + &appState.Java, + &tracking, + appState.Ovr); + + // Set up the description for this frame. + const ovrLayerHeader2 *layers[] = + { + &layer.Header + }; + + ovrSubmitFrameDescription2 frameDesc = {}; + frameDesc.Flags = 0; + frameDesc.SwapInterval = appState.SwapInterval; + frameDesc.FrameIndex = appState.FrameIndex; + frameDesc.DisplayTime = appState.DisplayTime; + frameDesc.LayerCount = 1; + frameDesc.Layers = layers; + + // Hand over the eye images to the time warp. + vrapi_SubmitFrame2(appState.Ovr, &frameDesc); + + } else { + // Set-up the compositor layers for this frame. + // NOTE: Multiple independent layers are allowed, but they need to be added + // in a depth consistent order. + memset( appState.Layers, 0, sizeof( ovrLayer_Union2 ) * ovrMaxLayerCount ); + appState.LayerCount = 0; + + // Add a simple cylindrical layer + appState.Layers[appState.LayerCount++].Cylinder = + BuildCylinderLayer( &appState.Scene.CylinderRenderer, + appState.Scene.CylinderWidth, appState.Scene.CylinderHeight, &tracking, radians(playerYaw) ); + + //Call the game drawing code to populate the cylinder layer texture + ovrRenderer_RenderFrame(&appState.Scene.CylinderRenderer, + &appState.Java, + &tracking, + appState.Ovr); + + + // Compose the layers for this frame. + const ovrLayerHeader2 * layerHeaders[ovrMaxLayerCount] = { 0 }; + for ( int i = 0; i < appState.LayerCount; i++ ) + { + layerHeaders[i] = &appState.Layers[i].Header; + } + + // Set up the description for this frame. + frameDesc.Flags = 0; + frameDesc.SwapInterval = appState.SwapInterval; + frameDesc.FrameIndex = appState.FrameIndex; + frameDesc.DisplayTime = appState.DisplayTime; + frameDesc.LayerCount = appState.LayerCount; + frameDesc.Layers = layerHeaders; + + // Hand over the eye images to the time warp. + vrapi_SubmitFrame2(appState.Ovr, &frameDesc); + } } else { @@ -2538,10 +2494,11 @@ void * AppThreadFunction( void * parm ) return NULL; } -static void ovrAppThread_Create( ovrAppThread * appThread, JNIEnv * env, jobject activityObject ) +static void ovrAppThread_Create( ovrAppThread * appThread, JNIEnv * env, jobject activityObject, jclass activityClass ) { (*env)->GetJavaVM( env, &appThread->JavaVm ); appThread->ActivityObject = (*env)->NewGlobalRef( env, activityObject ); + appThread->ActivityClass = (*env)->NewGlobalRef( env, activityClass ); appThread->Thread = 0; appThread->NativeWindow = NULL; ovrMessageQueue_Create( &appThread->MessageQueue ); @@ -2557,6 +2514,7 @@ static void ovrAppThread_Destroy( ovrAppThread * appThread, JNIEnv * env ) { pthread_join( appThread->Thread, NULL ); (*env)->DeleteGlobalRef( env, appThread->ActivityObject ); + (*env)->DeleteGlobalRef( env, appThread->ActivityClass ); ovrMessageQueue_Destroy( &appThread->MessageQueue ); } @@ -2618,6 +2576,11 @@ JNIEXPORT jlong JNICALL Java_com_drbeef_quakequest_GLES3JNILib_onCreate( JNIEnv if (ss->count > 0 && ss->dval[0] > 0.0) { SS_MULTIPLIER = ss->dval[0]; + + if (SS_MULTIPLIER > 1.2F) + { + SS_MULTIPLIER = 1.2F; + } } if (cpu->count > 0 && cpu->ival[0] > 0 && cpu->ival[0] < 10) @@ -2636,9 +2599,8 @@ JNIEXPORT jlong JNICALL Java_com_drbeef_quakequest_GLES3JNILib_onCreate( JNIEnv } } - ovrAppThread * appThread = (ovrAppThread *) malloc( sizeof( ovrAppThread ) ); - ovrAppThread_Create( appThread, env, activity ); + ovrAppThread_Create( appThread, env, activity, activityClass ); ovrMessageQueue_Enable( &appThread->MessageQueue, true ); ovrMessage message; diff --git a/Projects/Android/jni/QuakeQuestSrc/VrCommon.h b/Projects/Android/jni/QuakeQuestSrc/VrCommon.h new file mode 100644 index 0000000..1d8a988 --- /dev/null +++ b/Projects/Android/jni/QuakeQuestSrc/VrCommon.h @@ -0,0 +1,38 @@ +#if !defined(vrcommon_h) +#define vrcommon_h + +//#include +#include + +#include + +#include "../darkplaces/mathlib.h" + +#define LOG_TAG "QuakeQuest" + +#ifndef NDEBUG +#define DEBUG 1 +#endif + +#define ALOGE(...) __android_log_print( ANDROID_LOG_ERROR, LOG_TAG, __VA_ARGS__ ) + +#if DEBUG +#define ALOGV(...) __android_log_print( ANDROID_LOG_VERBOSE, LOG_TAG, __VA_ARGS__ ) +#else +#define ALOGV(...) +#endif + +float playerHeight; +float playerYaw; + +float radians(float deg); +float degrees(float rad); +qboolean isMultiplayer(); +double GetTimeInMilliSeconds(); +float length(float x, float y); +float nonLinearFilter(float in); +bool between(float min, float val, float max); +void QuatToYawPitchRoll(ovrQuatf q, float pitchAdjust, vec3_t out); +qboolean useScreenLayer(); + +#endif //vrcommon_h \ No newline at end of file diff --git a/Projects/Android/jni/QuakeQuestSrc/VrCompositor.c b/Projects/Android/jni/QuakeQuestSrc/VrCompositor.c new file mode 100644 index 0000000..e92f6b6 --- /dev/null +++ b/Projects/Android/jni/QuakeQuestSrc/VrCompositor.c @@ -0,0 +1,595 @@ +/************************************************************************************ + +Filename : VrCompositor.c + +*************************************************************************************/ + +#include +#include +#include +#include +#include +#include +#include // for prctl( PR_SET_NAME ) +#include +#include // for AWINDOW_FLAG_KEEP_SCREEN_ON + +#include +#include +#include +#include + + + +#include +#include + +#include "VrCompositor.h" + + + +/* +================================================================================ + +renderState + +================================================================================ +*/ + + +void getCurrentRenderState( renderState * state) +{ + state->VertexBuffer = 0; + state->IndexBuffer = 0; + state->VertexArrayObject = 0; + state->Program = 0; + + glGetIntegerv(GL_ARRAY_BUFFER, &state->VertexBuffer ); + glGetIntegerv(GL_ELEMENT_ARRAY_BUFFER, &state->IndexBuffer ); + glGetIntegerv(GL_VERTEX_ARRAY_BINDING, &state->VertexArrayObject ); + glGetIntegerv(GL_CURRENT_PROGRAM, &state->Program ); +} + +void restoreRenderState( renderState * state ) +{ + GL( glUseProgram( state->Program ) ); + GL( glBindVertexArray( state->VertexArrayObject ) ); + GL( glBindBuffer( GL_ARRAY_BUFFER, state->VertexBuffer ) ); + GL( glBindBuffer( GL_ELEMENT_ARRAY_BUFFER, state->IndexBuffer ) ); +} + + +/* +================================================================================ + +ovrGeometry + +================================================================================ +*/ + +enum VertexAttributeLocation +{ + VERTEX_ATTRIBUTE_LOCATION_POSITION, + VERTEX_ATTRIBUTE_LOCATION_COLOR, + VERTEX_ATTRIBUTE_LOCATION_UV, +}; + +typedef struct +{ + enum VertexAttributeLocation location; + const char * name; +} ovrVertexAttribute; + +static ovrVertexAttribute ProgramVertexAttributes[] = +{ + { VERTEX_ATTRIBUTE_LOCATION_POSITION, "vertexPosition" }, + { VERTEX_ATTRIBUTE_LOCATION_COLOR, "vertexColor" }, + { VERTEX_ATTRIBUTE_LOCATION_UV, "vertexUv" }, +}; + +static void ovrGeometry_Clear( ovrGeometry * geometry ) +{ + geometry->VertexBuffer = 0; + geometry->IndexBuffer = 0; + geometry->VertexArrayObject = 0; + geometry->VertexCount = 0; + geometry->IndexCount = 0; + for ( int i = 0; i < MAX_VERTEX_ATTRIB_POINTERS; i++ ) + { + memset( &geometry->VertexAttribs[i], 0, sizeof( geometry->VertexAttribs[i] ) ); + geometry->VertexAttribs[i].Index = -1; + } +} + +static void ovrGeometry_CreateGroundPlane( ovrGeometry * geometry ) +{ + typedef struct + { + float positions[4][4]; + unsigned char colors[4][4]; + } ovrCubeVertices; + + static const ovrCubeVertices cubeVertices = + { + // positions + { + { 4.5f, -1.2f, 4.5f, 1.0f }, + { 4.5f, -1.2f, -4.5f, 1.0f }, + { -4.5f, -1.2f, -4.5f, 1.0f }, + { -4.5f, -1.2f, 4.5f, 1.0f } + }, + // colors + { + { 255, 0, 0, 255 }, + { 0, 255, 0, 255 }, + { 0, 0, 255, 255 }, + { 255, 255, 0, 255 }, + }, + }; + + static const unsigned short cubeIndices[6] = + { + 0, 1, 2, + 0, 2, 3, + }; + + geometry->VertexCount = 4; + geometry->IndexCount = 6; + + geometry->VertexAttribs[0].Index = VERTEX_ATTRIBUTE_LOCATION_POSITION; + geometry->VertexAttribs[0].Size = 4; + geometry->VertexAttribs[0].Type = GL_FLOAT; + geometry->VertexAttribs[0].Normalized = false; + geometry->VertexAttribs[0].Stride = sizeof( cubeVertices.positions[0] ); + geometry->VertexAttribs[0].Pointer = (const GLvoid *)offsetof( ovrCubeVertices, positions ); + + geometry->VertexAttribs[1].Index = VERTEX_ATTRIBUTE_LOCATION_COLOR; + geometry->VertexAttribs[1].Size = 4; + geometry->VertexAttribs[1].Type = GL_UNSIGNED_BYTE; + geometry->VertexAttribs[1].Normalized = true; + geometry->VertexAttribs[1].Stride = sizeof( cubeVertices.colors[0] ); + geometry->VertexAttribs[1].Pointer = (const GLvoid *)offsetof( ovrCubeVertices, colors ); + + renderState state; + getCurrentRenderState(&state); + + GL( glGenBuffers( 1, &geometry->VertexBuffer ) ); + GL( glBindBuffer( GL_ARRAY_BUFFER, geometry->VertexBuffer ) ); + GL( glBufferData( GL_ARRAY_BUFFER, sizeof( cubeVertices ), &cubeVertices, GL_STATIC_DRAW ) ); + + GL( glGenBuffers( 1, &geometry->IndexBuffer ) ); + GL( glBindBuffer( GL_ELEMENT_ARRAY_BUFFER, geometry->IndexBuffer ) ); + GL( glBufferData( GL_ELEMENT_ARRAY_BUFFER, sizeof( cubeIndices ), cubeIndices, GL_STATIC_DRAW ) ); + + restoreRenderState(&state); +} + +static void ovrGeometry_Destroy( ovrGeometry * geometry ) +{ + GL( glDeleteBuffers( 1, &geometry->IndexBuffer ) ); + GL( glDeleteBuffers( 1, &geometry->VertexBuffer ) ); + + ovrGeometry_Clear( geometry ); +} + +static void ovrGeometry_CreateVAO( ovrGeometry * geometry ) +{ + renderState state; + getCurrentRenderState(&state); + + GL( glGenVertexArrays( 1, &geometry->VertexArrayObject ) ); + GL( glBindVertexArray( geometry->VertexArrayObject ) ); + + GL( glBindBuffer( GL_ARRAY_BUFFER, geometry->VertexBuffer ) ); + + for ( int i = 0; i < MAX_VERTEX_ATTRIB_POINTERS; i++ ) + { + if ( geometry->VertexAttribs[i].Index != -1 ) + { + GL( glEnableVertexAttribArray( geometry->VertexAttribs[i].Index ) ); + GL( glVertexAttribPointer( geometry->VertexAttribs[i].Index, geometry->VertexAttribs[i].Size, + geometry->VertexAttribs[i].Type, geometry->VertexAttribs[i].Normalized, + geometry->VertexAttribs[i].Stride, geometry->VertexAttribs[i].Pointer ) ); + } + } + + GL( glBindBuffer( GL_ELEMENT_ARRAY_BUFFER, geometry->IndexBuffer ) ); + + restoreRenderState(&state); +} + +static void ovrGeometry_DestroyVAO( ovrGeometry * geometry ) +{ + GL( glDeleteVertexArrays( 1, &geometry->VertexArrayObject ) ); +} + +/* +================================================================================ + +ovrProgram + +================================================================================ +*/ + + +typedef struct +{ + enum + { + UNIFORM_VIEW_PROJ_MATRIX, + } index; + enum + { + UNIFORM_TYPE_VECTOR4, + UNIFORM_TYPE_MATRIX4X4, + UNIFORM_TYPE_INT, + UNIFORM_TYPE_BUFFER, + } type; + const char * name; +} ovrUniform; + +static ovrUniform ProgramUniforms[] = +{ + { UNIFORM_VIEW_PROJ_MATRIX, UNIFORM_TYPE_MATRIX4X4, "viewProjectionMatrix" }, +}; + +static void ovrProgram_Clear( ovrProgram * program ) +{ + program->Program = 0; + program->VertexShader = 0; + program->FragmentShader = 0; + memset( program->UniformLocation, 0, sizeof( program->UniformLocation ) ); + memset( program->UniformBinding, 0, sizeof( program->UniformBinding ) ); + memset( program->Textures, 0, sizeof( program->Textures ) ); +} + +static bool ovrProgram_Create( ovrProgram * program, const char * vertexSource, const char * fragmentSource ) +{ + GLint r; + + GL( program->VertexShader = glCreateShader( GL_VERTEX_SHADER ) ); + + GL( glShaderSource( program->VertexShader, 1, &vertexSource, 0 ) ); + GL( glCompileShader( program->VertexShader ) ); + GL( glGetShaderiv( program->VertexShader, GL_COMPILE_STATUS, &r ) ); + if ( r == GL_FALSE ) + { + GLchar msg[4096]; + GL( glGetShaderInfoLog( program->VertexShader, sizeof( msg ), 0, msg ) ); + ALOGE( "%s\n%s\n", vertexSource, msg ); + return false; + } + + GL( program->FragmentShader = glCreateShader( GL_FRAGMENT_SHADER ) ); + GL( glShaderSource( program->FragmentShader, 1, &fragmentSource, 0 ) ); + GL( glCompileShader( program->FragmentShader ) ); + GL( glGetShaderiv( program->FragmentShader, GL_COMPILE_STATUS, &r ) ); + if ( r == GL_FALSE ) + { + GLchar msg[4096]; + GL( glGetShaderInfoLog( program->FragmentShader, sizeof( msg ), 0, msg ) ); + ALOGE( "%s\n%s\n", fragmentSource, msg ); + return false; + } + + GL( program->Program = glCreateProgram() ); + GL( glAttachShader( program->Program, program->VertexShader ) ); + GL( glAttachShader( program->Program, program->FragmentShader ) ); + + // Bind the vertex attribute locations. + for ( int i = 0; i < sizeof( ProgramVertexAttributes ) / sizeof( ProgramVertexAttributes[0] ); i++ ) + { + GL( glBindAttribLocation( program->Program, ProgramVertexAttributes[i].location, ProgramVertexAttributes[i].name ) ); + } + + GL( glLinkProgram( program->Program ) ); + GL( glGetProgramiv( program->Program, GL_LINK_STATUS, &r ) ); + if ( r == GL_FALSE ) + { + GLchar msg[4096]; + GL( glGetProgramInfoLog( program->Program, sizeof( msg ), 0, msg ) ); + ALOGE( "Linking program failed: %s\n", msg ); + return false; + } + + int numBufferBindings = 0; + + // Get the uniform locations. + memset( program->UniformLocation, -1, sizeof( program->UniformLocation ) ); + for ( int i = 0; i < sizeof( ProgramUniforms ) / sizeof( ProgramUniforms[0] ); i++ ) + { + const int uniformIndex = ProgramUniforms[i].index; + if ( ProgramUniforms[i].type == UNIFORM_TYPE_BUFFER ) + { + GL( program->UniformLocation[uniformIndex] = glGetUniformBlockIndex( program->Program, ProgramUniforms[i].name ) ); + program->UniformBinding[uniformIndex] = numBufferBindings++; + GL( glUniformBlockBinding( program->Program, program->UniformLocation[uniformIndex], program->UniformBinding[uniformIndex] ) ); + } + else + { + GL( program->UniformLocation[uniformIndex] = glGetUniformLocation( program->Program, ProgramUniforms[i].name ) ); + program->UniformBinding[uniformIndex] = program->UniformLocation[uniformIndex]; + } + } + + renderState state; + getCurrentRenderState(&state); + + GL( glUseProgram( program->Program ) ); + + // Get the texture locations. + for ( int i = 0; i < MAX_PROGRAM_TEXTURES; i++ ) + { + char name[32]; + sprintf( name, "Texture%i", i ); + program->Textures[i] = glGetUniformLocation( program->Program, name ); + if ( program->Textures[i] != -1 ) + { + GL( glUniform1i( program->Textures[i], i ) ); + } + } + + restoreRenderState(&state); + + return true; +} + +static void ovrProgram_Destroy( ovrProgram * program ) +{ + if ( program->Program != 0 ) + { + GL( glDeleteProgram( program->Program ) ); + program->Program = 0; + } + if ( program->VertexShader != 0 ) + { + GL( glDeleteShader( program->VertexShader ) ); + program->VertexShader = 0; + } + if ( program->FragmentShader != 0 ) + { + GL( glDeleteShader( program->FragmentShader ) ); + program->FragmentShader = 0; + } +} + +static const char VERTEX_SHADER[] = + "#version 300 es\n" + "in vec3 vertexPosition;\n" + "in vec4 vertexColor;\n" + "uniform mat4 viewProjectionMatrix;\n" + "out vec4 fragmentColor;\n" + "void main()\n" + "{\n" + " gl_Position = viewProjectionMatrix * vec4( vertexPosition, 1.0 );\n" + " fragmentColor = vertexColor;\n" + "}\n"; + +static const char FRAGMENT_SHADER[] = + "#version 300 es\n" + "in lowp vec4 fragmentColor;\n" + "out lowp vec4 outColor;\n" + "void main()\n" + "{\n" + " outColor = fragmentColor;\n" + "}\n"; + +/* +================================================================================ + +ovrScene + +================================================================================ +*/ + +void ovrScene_Clear( ovrScene * scene ) +{ + scene->CreatedScene = false; + scene->CreatedVAOs = false; + ovrProgram_Clear( &scene->Program ); + ovrGeometry_Clear( &scene->GroundPlane ); + ovrRenderer_Clear( &scene->CylinderRenderer ); + + scene->CylinderWidth = 0; + scene->CylinderHeight = 0; +} + +bool ovrScene_IsCreated( ovrScene * scene ) +{ + return scene->CreatedScene; +} + +void ovrScene_CreateVAOs( ovrScene * scene ) +{ + if ( !scene->CreatedVAOs ) + { + ovrGeometry_CreateVAO( &scene->GroundPlane ); + scene->CreatedVAOs = true; + } +} + +void ovrScene_DestroyVAOs( ovrScene * scene ) +{ + if ( scene->CreatedVAOs ) + { + ovrGeometry_DestroyVAO( &scene->GroundPlane ); + scene->CreatedVAOs = false; + } +} + +void ovrScene_Create( int width, int height, ovrScene * scene, const ovrJava * java ) +{ + // Simple ground plane geometry. + { + ovrProgram_Create( &scene->Program, VERTEX_SHADER, FRAGMENT_SHADER ); + ovrGeometry_CreateGroundPlane( &scene->GroundPlane ); + ovrScene_CreateVAOs( scene ); + } + + // Create Cylinder renderer + { + scene->CylinderWidth = width; + scene->CylinderHeight = height; + + //Create cylinder renderer + ovrRenderer_Create( width, height, &scene->CylinderRenderer, java ); + } + + scene->CreatedScene = true; +} + +void ovrScene_Destroy( ovrScene * scene ) +{ + ovrScene_DestroyVAOs( scene ); + ovrProgram_Destroy( &scene->Program ); + ovrGeometry_Destroy( &scene->GroundPlane ); + ovrRenderer_Destroy( &scene->CylinderRenderer ); + + scene->CreatedScene = false; +} + +/* +================================================================================ + +ovrRenderer + +================================================================================ +*/ + +ovrLayerProjection2 ovrRenderer_RenderGroundPlaneToEyeBuffer( ovrRenderer * renderer, const ovrJava * java, + const ovrScene * scene, const ovrTracking2 * tracking ) +{ + ovrLayerProjection2 layer = vrapi_DefaultLayerProjection2(); + layer.HeadPose = tracking->HeadPose; + for ( int eye = 0; eye < VRAPI_FRAME_LAYER_EYE_MAX; eye++ ) + { + ovrFramebuffer * frameBuffer = &renderer->FrameBuffer[eye]; + layer.Textures[eye].ColorSwapChain = frameBuffer->ColorTextureSwapChain; + layer.Textures[eye].SwapChainIndex = frameBuffer->TextureSwapChainIndex; + layer.Textures[eye].TexCoordsFromTanAngles = ovrMatrix4f_TanAngleMatrixFromProjection( &tracking->Eye[eye].ProjectionMatrix ); + } + layer.Header.Flags |= VRAPI_FRAME_LAYER_FLAG_CHROMATIC_ABERRATION_CORRECTION; + + for ( int eye = 0; eye < VRAPI_FRAME_LAYER_EYE_MAX; eye++ ) + { + ovrFramebuffer * frameBuffer = &renderer->FrameBuffer[eye]; + ovrFramebuffer_SetCurrent( frameBuffer ); + + renderState state; + getCurrentRenderState(&state); + + GL( glUseProgram( scene->Program.Program ) ); + + ovrMatrix4f viewProjMatrix = ovrMatrix4f_Multiply( &tracking->Eye[eye].ProjectionMatrix, &tracking->Eye[eye].ViewMatrix ); + glUniformMatrix4fv( scene->Program.UniformLocation[UNIFORM_VIEW_PROJ_MATRIX], 1, GL_TRUE, &viewProjMatrix.M[0][0] ); + + GL( glEnable( GL_SCISSOR_TEST ) ); + GL( glDepthMask( GL_TRUE ) ); + GL( glEnable( GL_DEPTH_TEST ) ); + GL( glDepthFunc( GL_LEQUAL ) ); + GL( glEnable( GL_CULL_FACE ) ); + GL( glCullFace( GL_BACK ) ); + GL( glViewport( 0, 0, frameBuffer->Width, frameBuffer->Height ) ); + GL( glScissor( 0, 0, frameBuffer->Width, frameBuffer->Height ) ); + GL( glClearColor( 0.0f, 0.0f, 0.0f, 1.0f ) ); + GL( glClear( GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT ) ); + + //bind buffers + GL( glBindBuffer( GL_ARRAY_BUFFER, scene->GroundPlane.VertexBuffer ) ); + GL( glBindBuffer( GL_ELEMENT_ARRAY_BUFFER, scene->GroundPlane.IndexBuffer ) ); + GL( glBindVertexArray( scene->GroundPlane.VertexArrayObject ) ); + + GL( glDrawElements( GL_TRIANGLES, scene->GroundPlane.IndexCount, GL_UNSIGNED_SHORT, NULL ) ); + + restoreRenderState(&state); + + // Explicitly clear the border texels to black when GL_CLAMP_TO_BORDER is not available. + ovrFramebuffer_ClearEdgeTexels( frameBuffer ); + + ovrFramebuffer_Resolve( frameBuffer ); + ovrFramebuffer_Advance( frameBuffer ); + } + + ovrFramebuffer_SetNone(); + + return layer; +} + +// Assumes landscape cylinder shape. +static ovrMatrix4f CylinderModelMatrix( const int texWidth, const int texHeight, + const ovrVector3f translation, + const float rotateYaw, + const float rotatePitch, + const float radius, + const float density ) +{ + const ovrMatrix4f scaleMatrix = ovrMatrix4f_CreateScale( radius, radius * (float)texHeight * VRAPI_PI / density, radius ); + const ovrMatrix4f transMatrix = ovrMatrix4f_CreateTranslation( translation.x, translation.y, translation.z ); + const ovrMatrix4f rotXMatrix = ovrMatrix4f_CreateRotation( rotateYaw, 0.0f, 0.0f ); + const ovrMatrix4f rotYMatrix = ovrMatrix4f_CreateRotation( 0.0f, rotatePitch, 0.0f ); + + const ovrMatrix4f m0 = ovrMatrix4f_Multiply( &transMatrix, &scaleMatrix ); + const ovrMatrix4f m1 = ovrMatrix4f_Multiply( &rotXMatrix, &m0 ); + const ovrMatrix4f m2 = ovrMatrix4f_Multiply( &rotYMatrix, &m1 ); + + return m2; +} + +ovrLayerCylinder2 BuildCylinderLayer( ovrRenderer * cylinderRenderer, + const int textureWidth, const int textureHeight, + const ovrTracking2 * tracking, float rotatePitch ) +{ + ovrLayerCylinder2 layer = vrapi_DefaultLayerCylinder2(); + + const float fadeLevel = 1.0f; + layer.Header.ColorScale.x = + layer.Header.ColorScale.y = + layer.Header.ColorScale.z = + layer.Header.ColorScale.w = fadeLevel; + layer.Header.SrcBlend = VRAPI_FRAME_LAYER_BLEND_SRC_ALPHA; + layer.Header.DstBlend = VRAPI_FRAME_LAYER_BLEND_ONE_MINUS_SRC_ALPHA; + + //layer.Header.Flags = VRAPI_FRAME_LAYER_FLAG_CLIP_TO_TEXTURE_RECT; + + layer.HeadPose = tracking->HeadPose; + + const float density = 4500.0f; + const float rotateYaw = 0.0f; + const float radius = 2.0f; + const ovrVector3f translation = { 0.0f, 1.5f, -2.5f }; + + ovrMatrix4f cylinderTransform = + CylinderModelMatrix( textureWidth, textureHeight, translation, + rotateYaw, rotatePitch, radius, density ); + + const float circScale = density * 0.5f / textureWidth; + const float circBias = -circScale * ( 0.5f * ( 1.0f - 1.0f / circScale ) ); + + for ( int eye = 0; eye < VRAPI_FRAME_LAYER_EYE_MAX; eye++ ) + { + ovrFramebuffer * cylinderFrameBuffer = &cylinderRenderer->FrameBuffer[eye]; + + ovrMatrix4f modelViewMatrix = ovrMatrix4f_Multiply( &tracking->Eye[eye].ViewMatrix, &cylinderTransform ); + layer.Textures[eye].TexCoordsFromTanAngles = ovrMatrix4f_Inverse( &modelViewMatrix ); + layer.Textures[eye].ColorSwapChain = cylinderFrameBuffer->ColorTextureSwapChain; + layer.Textures[eye].SwapChainIndex = cylinderFrameBuffer->TextureSwapChainIndex; + + // Texcoord scale and bias is just a representation of the aspect ratio. The positioning + // of the cylinder is handled entirely by the TexCoordsFromTanAngles matrix. + + const float texScaleX = circScale; + const float texBiasX = circBias; + const float texScaleY = -0.5f; + const float texBiasY = texScaleY * ( 0.5f * ( 1.0f - ( 1.0f / texScaleY ) ) ); + + layer.Textures[eye].TextureMatrix.M[0][0] = texScaleX; + layer.Textures[eye].TextureMatrix.M[0][2] = texBiasX; + layer.Textures[eye].TextureMatrix.M[1][1] = texScaleY; + layer.Textures[eye].TextureMatrix.M[1][2] = -texBiasY; + + layer.Textures[eye].TextureRect.width = 1.0f; + layer.Textures[eye].TextureRect.height = 1.0f; + } + + return layer; +} diff --git a/Projects/Android/jni/QuakeQuestSrc/VrCompositor.h b/Projects/Android/jni/QuakeQuestSrc/VrCompositor.h new file mode 100644 index 0000000..518f937 --- /dev/null +++ b/Projects/Android/jni/QuakeQuestSrc/VrCompositor.h @@ -0,0 +1,216 @@ +/************************************************************************************ + +Filename : VrCompositor.h + +*************************************************************************************/ + +#define CHECK_GL_ERRORS +#ifdef CHECK_GL_ERRORS + +#include "VrCommon.h" + +static const char * GlErrorString( GLenum error ) +{ + switch ( error ) + { + case GL_NO_ERROR: return "GL_NO_ERROR"; + case GL_INVALID_ENUM: return "GL_INVALID_ENUM"; + case GL_INVALID_VALUE: return "GL_INVALID_VALUE"; + case GL_INVALID_OPERATION: return "GL_INVALID_OPERATION"; + case GL_INVALID_FRAMEBUFFER_OPERATION: return "GL_INVALID_FRAMEBUFFER_OPERATION"; + case GL_OUT_OF_MEMORY: return "GL_OUT_OF_MEMORY"; + default: return "unknown"; + } +} + +static void GLCheckErrors( int line ) +{ + for ( int i = 0; i < 10; i++ ) + { + const GLenum error = glGetError(); + if ( error == GL_NO_ERROR ) + { + break; + } + ALOGE( "GL error on line %d: %s", line, GlErrorString( error ) ); + } +} + +#define GL( func ) func; GLCheckErrors( __LINE__ ); + +#else // CHECK_GL_ERRORS + +#define GL( func ) func; + +#endif // CHECK_GL_ERRORS + + +/* +================================================================================ + +ovrFramebuffer + +================================================================================ +*/ + +typedef struct +{ + int Width; + int Height; + int Multisamples; + int TextureSwapChainLength; + int TextureSwapChainIndex; + ovrTextureSwapChain * ColorTextureSwapChain; + GLuint * DepthBuffers; + GLuint * FrameBuffers; +} ovrFramebuffer; + +void ovrFramebuffer_SetCurrent( ovrFramebuffer * frameBuffer ); +void ovrFramebuffer_Destroy( ovrFramebuffer * frameBuffer ); +void ovrFramebuffer_SetNone(); +void ovrFramebuffer_Resolve( ovrFramebuffer * frameBuffer ); +void ovrFramebuffer_Advance( ovrFramebuffer * frameBuffer ); +void ovrFramebuffer_ClearEdgeTexels( ovrFramebuffer * frameBuffer ); + +/* +================================================================================ + +ovrRenderer + +================================================================================ +*/ + +typedef struct +{ + ovrFramebuffer FrameBuffer[VRAPI_FRAME_LAYER_EYE_MAX]; + int NumBuffers; +} ovrRenderer; + + +void ovrRenderer_Clear( ovrRenderer * renderer ); +void ovrRenderer_Create( int width, int height, ovrRenderer * renderer, const ovrJava * java ); +void ovrRenderer_Destroy( ovrRenderer * renderer ); + + +/* +================================================================================ + +renderState + +================================================================================ +*/ + +typedef struct +{ + GLuint VertexBuffer; + GLuint IndexBuffer; + GLuint VertexArrayObject; + GLuint Program; + GLuint VertexShader; + GLuint FragmentShader; +} renderState; + +void getCurrentRenderState( renderState * state); +void restoreRenderState( renderState * state ); + +/* +================================================================================ + +ovrGeometry + +================================================================================ +*/ + +typedef struct +{ + GLuint Index; + GLint Size; + GLenum Type; + GLboolean Normalized; + GLsizei Stride; + const GLvoid * Pointer; +} ovrVertexAttribPointer; + +#define MAX_VERTEX_ATTRIB_POINTERS 3 + +typedef struct +{ + GLuint VertexBuffer; + GLuint IndexBuffer; + GLuint VertexArrayObject; + int VertexCount; + int IndexCount; + ovrVertexAttribPointer VertexAttribs[MAX_VERTEX_ATTRIB_POINTERS]; +} ovrGeometry; + +/* +================================================================================ + +ovrProgram + +================================================================================ +*/ + +#define MAX_PROGRAM_UNIFORMS 8 +#define MAX_PROGRAM_TEXTURES 8 + +typedef struct +{ + GLuint Program; + GLuint VertexShader; + GLuint FragmentShader; + // These will be -1 if not used by the program. + GLint UniformLocation[MAX_PROGRAM_UNIFORMS]; // ProgramUniforms[].name + GLint UniformBinding[MAX_PROGRAM_UNIFORMS]; // ProgramUniforms[].name + GLint Textures[MAX_PROGRAM_TEXTURES]; // Texture%i +} ovrProgram; + +/* +================================================================================ + +ovrScene + +================================================================================ +*/ + +typedef struct +{ + bool CreatedScene; + bool CreatedVAOs; + ovrProgram Program; + ovrGeometry GroundPlane; + + //Proper renderer for stereo rendering to the cylinder layer + ovrRenderer CylinderRenderer; + + int CylinderWidth; + int CylinderHeight; +} ovrScene; + +bool ovrScene_IsCreated( ovrScene * scene ); +void ovrScene_Clear( ovrScene * scene ); +void ovrScene_Create( int width, int height, ovrScene * scene, const ovrJava * java ); +void ovrScene_CreateVAOs( ovrScene * scene ); +void ovrScene_DestroyVAOs( ovrScene * scene ); +void ovrScene_Destroy( ovrScene * scene ); + +/* +================================================================================ + +ovrRenderer + +================================================================================ +*/ + +ovrLayerProjection2 ovrRenderer_RenderGroundPlaneToEyeBuffer( ovrRenderer * renderer, const ovrJava * java, + const ovrScene * scene, const ovrTracking2 * tracking ); + +ovrLayerProjection2 ovrRenderer_RenderToEyeBuffer( ovrRenderer * renderer, const ovrJava * java, + const ovrTracking2 * tracking ); + +ovrLayerCylinder2 BuildCylinderLayer( ovrRenderer * cylinderRenderer, + const int textureWidth, const int textureHeight, + const ovrTracking2 * tracking, float rotateYaw ); +; + + diff --git a/Projects/Android/jni/darkplaces/cl_input.c b/Projects/Android/jni/darkplaces/cl_input.c index ef39b64..e77cd71 100644 --- a/Projects/Android/jni/darkplaces/cl_input.c +++ b/Projects/Android/jni/darkplaces/cl_input.c @@ -430,9 +430,7 @@ float CL_KeyState (kbutton_t *key) //========================================================================== cvar_t cl_upspeed = {CVAR_SAVE, "cl_upspeed","400","vertical movement speed (while swimming or flying)"}; -cvar_t cl_forwardspeed = {CVAR_SAVE, "cl_forwardspeed","150","forward movement speed"}; -cvar_t cl_backspeed = {CVAR_SAVE, "cl_backspeed","150","backward movement speed"}; -cvar_t cl_sidespeed = {CVAR_SAVE, "cl_sidespeed","150","strafe movement speed"}; +cvar_t cl_movementspeed = {CVAR_SAVE, "cl_movementspeed","150","forward movement speed"}; cvar_t cl_movespeedkey = {CVAR_SAVE, "cl_movespeedkey","2.0","how much +speed multiplies keyboard movement speed"}; cvar_t cl_movecliptokeyboard = {0, "cl_movecliptokeyboard", "0", "if set to 1, any move is clipped to the nine keyboard states; if set to 2, only the direction is clipped, not the amount"}; @@ -446,7 +444,7 @@ cvar_t cl_yawmult = {CVAR_SAVE, "cl_yawmult","1.0","Multiplier for yaw (leave at cvar_t cl_pitchmult = {CVAR_SAVE, "cl_pitchmult","1.0","Multiplier for yaw (leave at 1.0)"}; cvar_t cl_controllerdeadzone = {0, "cl_controllerdeadzone","0.05","Amount of deadzone to prevent movement drift due to badly calibrated controller (0.0 to 1.0)"}; cvar_t cl_righthanded = {CVAR_SAVE, "cl_righthanded","1","right-handed?"}; -cvar_t vr_weaponpitchadjust = {CVAR_SAVE, "vr_weaponpitchadjust","-30.0","Weapon pitch adjustment"}; +cvar_t vr_weaponpitchadjust = {CVAR_SAVE, "vr_weaponpitchadjust","-20.0","Weapon pitch adjustment"}; cvar_t cl_trackingmode = {CVAR_SAVE, "cl_trackingmode","1","Tracking Mode:- 1 - 6DoF or 0 - 3DoF"}; @@ -578,20 +576,20 @@ void CL_Input (void) // get basic movement from keyboard if (in_strafe.state & 1) { - cl.cmd.sidemove += cl_sidespeed.value * CL_KeyState (&in_right); - cl.cmd.sidemove -= cl_sidespeed.value * CL_KeyState (&in_left); + cl.cmd.sidemove += cl_movementspeed.value * CL_KeyState (&in_right); + cl.cmd.sidemove -= cl_movementspeed.value * CL_KeyState (&in_left); } - cl.cmd.sidemove += cl_sidespeed.value * CL_KeyState (&in_moveright); - cl.cmd.sidemove -= cl_sidespeed.value * CL_KeyState (&in_moveleft); + cl.cmd.sidemove += cl_movementspeed.value * CL_KeyState (&in_moveright); + cl.cmd.sidemove -= cl_movementspeed.value * CL_KeyState (&in_moveleft); cl.cmd.upmove += cl_upspeed.value * CL_KeyState (&in_up); cl.cmd.upmove -= cl_upspeed.value * CL_KeyState (&in_down); if (! (in_klook.state & 1) ) { - cl.cmd.forwardmove += cl_forwardspeed.value * CL_KeyState (&in_forward); - cl.cmd.forwardmove -= cl_backspeed.value * CL_KeyState (&in_back); + cl.cmd.forwardmove += cl_movementspeed.value * CL_KeyState (&in_forward); + cl.cmd.forwardmove -= cl_movementspeed.value * CL_KeyState (&in_back); } // adjust for speed key @@ -777,16 +775,16 @@ void CL_Input (void) else if(cl_movecliptokeyboard.integer) { // digital direction, digital amount - if(cl.cmd.sidemove >= cl_sidespeed.value * f * 0.5) - cl.cmd.sidemove = cl_sidespeed.value * f; - else if(cl.cmd.sidemove <= -cl_sidespeed.value * f * 0.5) - cl.cmd.sidemove = -cl_sidespeed.value * f; + if(cl.cmd.sidemove >= cl_movementspeed.value * f * 0.5) + cl.cmd.sidemove = cl_movementspeed.value * f; + else if(cl.cmd.sidemove <= -cl_movementspeed.value * f * 0.5) + cl.cmd.sidemove = -cl_movementspeed.value * f; else cl.cmd.sidemove = 0; - if(cl.cmd.forwardmove >= cl_forwardspeed.value * f * 0.5) - cl.cmd.forwardmove = cl_forwardspeed.value * f; - else if(cl.cmd.forwardmove <= -cl_backspeed.value * f * 0.5) - cl.cmd.forwardmove = -cl_backspeed.value * f; + if(cl.cmd.forwardmove >= cl_movementspeed.value * f * 0.5) + cl.cmd.forwardmove = cl_movementspeed.value * f; + else if(cl.cmd.forwardmove <= -cl_movementspeed.value * f * 0.5) + cl.cmd.forwardmove = -cl_movementspeed.value * f; else cl.cmd.forwardmove = 0; } diff --git a/Projects/Android/jni/darkplaces/cl_main.c b/Projects/Android/jni/darkplaces/cl_main.c index ce2aab5..b20a8af 100644 --- a/Projects/Android/jni/darkplaces/cl_main.c +++ b/Projects/Android/jni/darkplaces/cl_main.c @@ -2469,9 +2469,7 @@ void CL_Init (void) // register our commands // Cvar_RegisterVariable (&cl_upspeed); - Cvar_RegisterVariable (&cl_forwardspeed); - Cvar_RegisterVariable (&cl_backspeed); - Cvar_RegisterVariable (&cl_sidespeed); + Cvar_RegisterVariable (&cl_movementspeed); Cvar_RegisterVariable (&cl_movespeedkey); Cvar_RegisterVariable (&cl_yawspeed); Cvar_RegisterVariable (&cl_pitchspeed); diff --git a/Projects/Android/jni/darkplaces/client.h b/Projects/Android/jni/darkplaces/client.h index 608aa0a..87c4c7e 100644 --- a/Projects/Android/jni/darkplaces/client.h +++ b/Projects/Android/jni/darkplaces/client.h @@ -1483,9 +1483,7 @@ extern cvar_t rcon_password; extern cvar_t rcon_address; extern cvar_t cl_upspeed; -extern cvar_t cl_forwardspeed; -extern cvar_t cl_backspeed; -extern cvar_t cl_sidespeed; +extern cvar_t cl_movementspeed; extern cvar_t cl_movespeedkey; diff --git a/Projects/Android/jni/darkplaces/gl_rmain.c b/Projects/Android/jni/darkplaces/gl_rmain.c index 8231e2e..aa9a2f4 100644 --- a/Projects/Android/jni/darkplaces/gl_rmain.c +++ b/Projects/Android/jni/darkplaces/gl_rmain.c @@ -51,9 +51,11 @@ static qboolean r_gpuskeletal; // 1.0 pixel unit / 1.5 inch == 0.666666 pixel units per inch cvar_t vr_worldscale = {CVAR_SAVE, "vr_worldscale", "26.2467", "VR World scale multiplier"}; +qboolean useScreenLayer(); + float GetStereoSeparation() { - return vr_worldscale.value * 0.065f; + return useScreenLayer() ? 0.0f : vr_worldscale.value * 0.065f; } diff --git a/Projects/Android/jni/darkplaces/menu.c b/Projects/Android/jni/darkplaces/menu.c index 9d9543a..125dbcb 100644 --- a/Projects/Android/jni/darkplaces/menu.c +++ b/Projects/Android/jni/darkplaces/menu.c @@ -1756,27 +1756,13 @@ static void M_Menu_Options_AdjustSliders (int dir) else if (options_cursor == optnum++) ; else if (options_cursor == optnum++) { - cl_forwardspeed.value += dir * 10; - if (cl_forwardspeed.value > 500) - cl_forwardspeed.value = 500; - if (cl_forwardspeed.value < 10) - cl_forwardspeed.value = 10; + cl_movementspeed.value += dir * 10; + if (cl_movementspeed.value > 500) + cl_movementspeed.value = 500; + if (cl_movementspeed.value < 10) + cl_movementspeed.value = 10; - cl_backspeed.value += dir * 10; - if (cl_backspeed.value > 500) - cl_backspeed.value = 500; - if (cl_backspeed.value < 10) - cl_backspeed.value = 10; - - cl_sidespeed.value += dir * 10; - if (cl_sidespeed.value > 500) - cl_sidespeed.value = 500; - if (cl_sidespeed.value < 10) - cl_sidespeed.value = 10; - - Cvar_SetValueQuick (&cl_forwardspeed, cl_forwardspeed.value); - Cvar_SetValueQuick (&cl_backspeed, cl_backspeed.value); - Cvar_SetValueQuick (&cl_sidespeed, cl_sidespeed.value); + Cvar_SetValueQuick (&cl_movementspeed, cl_movementspeed.value); } else if (options_cursor == optnum++) Cvar_SetValueQuick(&showfps, !showfps.integer); else if (options_cursor == optnum++) ; @@ -1888,7 +1874,7 @@ static void M_Options_Draw (void) M_Options_PrintCommand( " Positional Tracking: Enabled", false); - M_Options_PrintSlider( " Player Movement Speed", true, cl_forwardspeed.value, 10, 500); + M_Options_PrintSlider( " Player Movement Speed", true, cl_movementspeed.value, 10, 500); M_Options_PrintCheckbox(" Show Framerate", true, showfps.integer); M_Options_PrintCommand( " Custom Brightness", true); M_Options_PrintSlider( " Game Brightness", true, r_hdr_scenebrightness.value, 1, 4); diff --git a/Projects/Android/jni/darkplaces/sbar.c b/Projects/Android/jni/darkplaces/sbar.c index dfcce13..49f31aa 100644 --- a/Projects/Android/jni/darkplaces/sbar.c +++ b/Projects/Android/jni/darkplaces/sbar.c @@ -134,9 +134,11 @@ int Sbar_GetYOffset() return offset; } +qboolean useScreenLayer(); + int Sbar_GetXOffset() { - if (!vrMode) + if (useScreenLayer()) return 0; //This will give the status bar depth in the 3D space diff --git a/Projects/Android/jni/darkplaces/vid_shared.c b/Projects/Android/jni/darkplaces/vid_shared.c index 4b4b1e1..f6743ac 100644 --- a/Projects/Android/jni/darkplaces/vid_shared.c +++ b/Projects/Android/jni/darkplaces/vid_shared.c @@ -1382,8 +1382,8 @@ void VID_ApplyJoyState(vid_joystate_t *joystate) VID_KeyEventForButton(vid_joystate.button[j] != 0, joystate->button[j] != 0, joybuttonkey360[j][c], &vid_joybuttontimer[j]); // axes - cl.cmd.forwardmove += VID_JoyState_GetAxis(joystate, joy_x360_axisforward.integer, joy_x360_sensitivityforward.value, joy_x360_deadzoneforward.value) * cl_forwardspeed.value; - cl.cmd.sidemove += VID_JoyState_GetAxis(joystate, joy_x360_axisside.integer, joy_x360_sensitivityside.value, joy_x360_deadzoneside.value) * cl_sidespeed.value; + cl.cmd.forwardmove += VID_JoyState_GetAxis(joystate, joy_x360_axisforward.integer, joy_x360_sensitivityforward.value, joy_x360_deadzoneforward.value) * cl_movementspeed.value; + cl.cmd.sidemove += VID_JoyState_GetAxis(joystate, joy_x360_axisside.integer, joy_x360_sensitivityside.value, joy_x360_deadzoneside.value) * cl_movementspeed.value; cl.cmd.upmove += VID_JoyState_GetAxis(joystate, joy_x360_axisup.integer, joy_x360_sensitivityup.value, joy_x360_deadzoneup.value) * cl_upspeed.value; cl.viewangles[0] += VID_JoyState_GetAxis(joystate, joy_x360_axispitch.integer, joy_x360_sensitivitypitch.value, joy_x360_deadzonepitch.value) * cl.realframetime * cl_pitchspeed.value; cl.viewangles[1] += VID_JoyState_GetAxis(joystate, joy_x360_axisyaw.integer, joy_x360_sensitivityyaw.value, joy_x360_deadzoneyaw.value) * cl.realframetime * cl_yawspeed.value; @@ -1396,8 +1396,8 @@ void VID_ApplyJoyState(vid_joystate_t *joystate) VID_KeyEventForButton(vid_joystate.button[j] != 0, joystate->button[j] != 0, joybuttonkey[j][c], &vid_joybuttontimer[j]); // axes - cl.cmd.forwardmove += VID_JoyState_GetAxis(joystate, joy_axisforward.integer, joy_sensitivityforward.value, joy_deadzoneforward.value) * cl_forwardspeed.value; - cl.cmd.sidemove += VID_JoyState_GetAxis(joystate, joy_axisside.integer, joy_sensitivityside.value, joy_deadzoneside.value) * cl_sidespeed.value; + cl.cmd.forwardmove += VID_JoyState_GetAxis(joystate, joy_axisforward.integer, joy_sensitivityforward.value, joy_deadzoneforward.value) * cl_movementspeed.value; + cl.cmd.sidemove += VID_JoyState_GetAxis(joystate, joy_axisside.integer, joy_sensitivityside.value, joy_deadzoneside.value) * cl_movementspeed.value; cl.cmd.upmove += VID_JoyState_GetAxis(joystate, joy_axisup.integer, joy_sensitivityup.value, joy_deadzoneup.value) * cl_upspeed.value; cl.viewangles[0] += VID_JoyState_GetAxis(joystate, joy_axispitch.integer, joy_sensitivitypitch.value, joy_deadzonepitch.value) * cl.realframetime * cl_pitchspeed.value; cl.viewangles[1] += VID_JoyState_GetAxis(joystate, joy_axisyaw.integer, joy_sensitivityyaw.value, joy_deadzoneyaw.value) * cl.realframetime * cl_yawspeed.value; diff --git a/Projects/Android/jni/darkplaces/view.c b/Projects/Android/jni/darkplaces/view.c index 29b9ffa..ab2522c 100644 --- a/Projects/Android/jni/darkplaces/view.c +++ b/Projects/Android/jni/darkplaces/view.c @@ -203,7 +203,7 @@ void V_DriftPitch (void) // don't count small mouse motion if (cl.nodrift) { - if ( fabs(cl.cmd.forwardmove) < cl_forwardspeed.value) + if ( fabs(cl.cmd.forwardmove) < cl_movementspeed.value) cl.driftmove = 0; else cl.driftmove += cl.realframetime;