Modified source engine (2017) developed by valve and leaked in 2020. Not for commercial purporses
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1517 lines
54 KiB

5 years ago
//========= Copyright Valve Corporation, All rights reserved. ============//
//
// Purpose:
//
// $NoKeywords: $
//===========================================================================//
#include "cbase.h"
#include "client_virtualreality.h"
#include "materialsystem/itexture.h"
#include "materialsystem/materialsystem_config.h"
#include "view_shared.h"
#include "view_scene.h"
#include "VGuiMatSurface/IMatSystemSurface.h"
#include "vgui_controls/Controls.h"
#include "sourcevr/isourcevirtualreality.h"
#include "ienginevgui.h"
#include "cdll_client_int.h"
#include "vgui/IVGui.h"
#include "vgui_controls/Controls.h"
#include "tier0/vprof_telemetry.h"
#include <time.h>
#include "steam/steam_api.h"
const char *COM_GetModDirectory(); // return the mod dir (rather than the complete -game param, which can be a path)
CClientVirtualReality g_ClientVirtualReality;
EXPOSE_SINGLE_INTERFACE_GLOBALVAR( CClientVirtualReality, IClientVirtualReality,
CLIENTVIRTUALREALITY_INTERFACE_VERSION, g_ClientVirtualReality );
// --------------------------------------------------------------------
// A huge pile of VR convars
// --------------------------------------------------------------------
ConVar vr_activate_default( "vr_activate_default", "0", FCVAR_ARCHIVE, "If this is true the game will switch to VR mode once startup is complete." );
ConVar vr_moveaim_mode ( "vr_moveaim_mode", "3", FCVAR_ARCHIVE, "0=move+shoot from face. 1=move with torso. 2,3,4=shoot with face+mouse cursor. 5+ are probably not that useful." );
ConVar vr_moveaim_mode_zoom ( "vr_moveaim_mode_zoom", "3", FCVAR_ARCHIVE, "0=move+shoot from face. 1=move with torso. 2,3,4=shoot with face+mouse cursor. 5+ are probably not that useful." );
ConVar vr_moveaim_reticle_yaw_limit ( "vr_moveaim_reticle_yaw_limit", "10", FCVAR_ARCHIVE, "Beyond this number of degrees, the mouse drags the torso" );
ConVar vr_moveaim_reticle_pitch_limit ( "vr_moveaim_reticle_pitch_limit", "30", FCVAR_ARCHIVE, "Beyond this number of degrees, the mouse clamps" );
// Note these are scaled by the zoom factor.
ConVar vr_moveaim_reticle_yaw_limit_zoom ( "vr_moveaim_reticle_yaw_limit_zoom", "0", FCVAR_ARCHIVE, "Beyond this number of degrees, the mouse drags the torso" );
ConVar vr_moveaim_reticle_pitch_limit_zoom ( "vr_moveaim_reticle_pitch_limit_zoom", "-1", FCVAR_ARCHIVE, "Beyond this number of degrees, the mouse clamps" );
// This are somewhat obsolete.
ConVar vr_aim_yaw_offset( "vr_aim_yaw_offset", "90", 0, "This value is added to Yaw when returning the vehicle aim angles to Source." );
ConVar vr_stereo_swap_eyes ( "vr_stereo_swap_eyes", "0", 0, "1=swap eyes." );
// Useful for debugging wacky-projection problems, separate from multi-rendering problems.
ConVar vr_stereo_mono_set_eye ( "vr_stereo_mono_set_eye", "0", 0, "0=off, Set all eyes to 1=left, 2=right, 3=middle eye" );
// Useful for examining anims, etc.
ConVar vr_debug_remote_cam( "vr_debug_remote_cam", "0" );
ConVar vr_debug_remote_cam_pos_x( "vr_debug_remote_cam_pos_x", "150.0" );
ConVar vr_debug_remote_cam_pos_y( "vr_debug_remote_cam_pos_y", "0.0" );
ConVar vr_debug_remote_cam_pos_z( "vr_debug_remote_cam_pos_z", "0.0" );
ConVar vr_debug_remote_cam_target_x( "vr_debug_remote_cam_target_x", "0.0" );
ConVar vr_debug_remote_cam_target_y( "vr_debug_remote_cam_target_y", "0.0" );
ConVar vr_debug_remote_cam_target_z( "vr_debug_remote_cam_target_z", "-50.0" );
ConVar vr_translation_limit( "vr_translation_limit", "10.0", 0, "How far the in-game head will translate before being clamped." );
// HUD config values
ConVar vr_render_hud_in_world( "vr_render_hud_in_world", "1" );
ConVar vr_hud_max_fov( "vr_hud_max_fov", "60", FCVAR_ARCHIVE, "Max FOV of the HUD" );
ConVar vr_hud_forward( "vr_hud_forward", "500", FCVAR_ARCHIVE, "Apparent distance of the HUD in inches" );
ConVar vr_hud_display_ratio( "vr_hud_display_ratio", "0.95", FCVAR_ARCHIVE );
ConVar vr_hud_never_overlay( "vr_hud_never_overlay", "0" );
ConVar vr_hud_axis_lock_to_world( "vr_hud_axis_lock_to_world", "0", FCVAR_ARCHIVE, "Bitfield - locks HUD axes to the world - 0=pitch, 1=yaw, 2=roll" );
// Default distance clips through rocketlauncher, heavy's body, etc.
ConVar vr_projection_znear_multiplier( "vr_projection_znear_multiplier", "0.3", 0, "Allows moving the ZNear plane to deal with body clipping" );
// Should the viewmodel (weapon) translate with the HMD, or remain fixed to the in-world body (but still rotate with the head)? Purely a graphics effect - no effect on actual bullet aiming.
// Has no effect in aim modes where aiming is not controlled by the head.
ConVar vr_viewmodel_translate_with_head ( "vr_viewmodel_translate_with_head", "0", 0, "1=translate the viewmodel with the head motion." );
ConVar vr_zoom_multiplier ( "vr_zoom_multiplier", "2.0", FCVAR_ARCHIVE, "When zoomed, how big is the scope on your HUD?" );
ConVar vr_zoom_scope_scale ( "vr_zoom_scope_scale", "6.0", 0, "Something to do with the default scope HUD overlay size." ); // Horrible hack - should work out the math properly, but we need to ship.
ConVar vr_viewmodel_offset_forward( "vr_viewmodel_offset_forward", "-8", 0 );
ConVar vr_viewmodel_offset_forward_large( "vr_viewmodel_offset_forward_large", "-15", 0 );
ConVar vr_force_windowed ( "vr_force_windowed", "0", FCVAR_ARCHIVE );
ConVar vr_first_person_uses_world_model ( "vr_first_person_uses_world_model", "1", 0, "Causes the third person model to be drawn instead of the view model" );
// --------------------------------------------------------------------
// Purpose: Cycle through the aim & move modes.
// --------------------------------------------------------------------
void CC_VR_Cycle_Aim_Move_Mode ( const CCommand& args )
{
int hmmCurrentMode = vr_moveaim_mode.GetInt();
if ( g_ClientVirtualReality.CurrentlyZoomed() )
{
hmmCurrentMode = vr_moveaim_mode_zoom.GetInt();
}
hmmCurrentMode++;
if ( hmmCurrentMode >= HMM_LAST )
{
hmmCurrentMode = 0;
}
if ( g_ClientVirtualReality.CurrentlyZoomed() )
{
vr_moveaim_mode_zoom.SetValue ( hmmCurrentMode );
Warning ( "Headtrack mode (zoomed) %d\n", hmmCurrentMode );
}
else
{
vr_moveaim_mode.SetValue ( hmmCurrentMode );
Warning ( "Headtrack mode %d\n", hmmCurrentMode );
}
}
static ConCommand vr_cycle_aim_move_mode("vr_cycle_aim_move_mode", CC_VR_Cycle_Aim_Move_Mode, "Cycle through the aim & move modes." );
// --------------------------------------------------------------------
// Purpose: Switch to/from VR mode.
// --------------------------------------------------------------------
CON_COMMAND( vr_activate, "Switch to VR mode" )
{
g_ClientVirtualReality.Activate();
}
CON_COMMAND( vr_deactivate, "Switch from VR mode to normal mode" )
{
g_ClientVirtualReality.Deactivate();
}
CON_COMMAND( vr_toggle, "Toggles VR mode" )
{
if( g_pSourceVR )
{
if( g_pSourceVR->ShouldRunInVR() )
g_ClientVirtualReality.Deactivate();
else
g_ClientVirtualReality.Activate();
}
else
{
Msg( "VR Mode is not enabled.\n" );
}
}
// --------------------------------------------------------------------
// Purpose: Returns true if the matrix is orthonormal
// --------------------------------------------------------------------
bool IsOrthonormal ( VMatrix Mat, float fTolerance )
{
float LenFwd = Mat.GetForward().Length();
float LenUp = Mat.GetUp().Length();
float LenLeft = Mat.GetLeft().Length();
float DotFwdUp = Mat.GetForward().Dot ( Mat.GetUp() );
float DotUpLeft = Mat.GetUp().Dot ( Mat.GetLeft() );
float DotLeftFwd = Mat.GetLeft().Dot ( Mat.GetForward() );
if ( fabsf ( LenFwd - 1.0f ) > fTolerance )
{
return false;
}
if ( fabsf ( LenUp - 1.0f ) > fTolerance )
{
return false;
}
if ( fabsf ( LenLeft - 1.0f ) > fTolerance )
{
return false;
}
if ( fabsf ( DotFwdUp ) > fTolerance )
{
return false;
}
if ( fabsf ( DotUpLeft ) > fTolerance )
{
return false;
}
if ( fabsf ( DotLeftFwd ) > fTolerance )
{
return false;
}
return true;
}
// --------------------------------------------------------------------
// Purpose: Computes the FOV from the projection matrix
// --------------------------------------------------------------------
void CalcFovFromProjection ( float *pFov, const VMatrix &proj )
{
// The projection matrices should be of the form:
// p0 0 z1 p1
// 0 p2 z2 p3
// 0 0 z3 1
// (p0 = X fov, p1 = X offset, p2 = Y fov, p3 = Y offset )
// TODO: cope with more complex projection matrices?
float xscale = proj.m[0][0];
Assert ( proj.m[0][1] == 0.0f );
float xoffset = proj.m[0][2];
Assert ( proj.m[0][3] == 0.0f );
Assert ( proj.m[1][0] == 0.0f );
float yscale = proj.m[1][1];
float yoffset = proj.m[1][2];
Assert ( proj.m[1][3] == 0.0f );
// Row 2 determines Z-buffer values - don't care about those for now.
Assert ( proj.m[3][0] == 0.0f );
Assert ( proj.m[3][1] == 0.0f );
Assert ( proj.m[3][2] == -1.0f );
Assert ( proj.m[3][3] == 0.0f );
/*
// The math here:
// A view-space vector (x,y,z,1) is transformed by the projection matrix
// / xscale 0 xoffset 0 \
// | 0 yscale yoffset 0 |
// | ? ? ? ? |
// \ 0 0 -1 0 /
//
// Then the result is normalized (i.e. divide by w) and the result clipped to the [-1,+1] unit cube.
// (ignore Z for now, and the clipping is slightly different).
// So, we want to know what vectors produce a clip value of -1 and +1 in each direction, e.g. in the X direction:
// +-1 = ( xscale*x + xoffset*z ) / (-1*z)
// = xscale*(x/z) + xoffset (I flipped the signs of both sides)
// => (+-1 - xoffset)/xscale = x/z
// ...and x/z is tan(theta), and theta is the half-FOV.
*/
float fov_px = 2.0f * RAD2DEG ( atanf ( fabsf ( ( 1.0f - xoffset ) / xscale ) ) );
float fov_nx = 2.0f * RAD2DEG ( atanf ( fabsf ( ( -1.0f - xoffset ) / xscale ) ) );
float fov_py = 2.0f * RAD2DEG ( atanf ( fabsf ( ( 1.0f - yoffset ) / yscale ) ) );
float fov_ny = 2.0f * RAD2DEG ( atanf ( fabsf ( ( -1.0f - yoffset ) / yscale ) ) );
*pFov = Max ( Max ( fov_px, fov_nx ), Max ( fov_py, fov_ny ) );
// FIXME: hey you know, I could do the Max() series before I call all those expensive atanf()s...
}
// --------------------------------------------------------------------
// construction/destruction
// --------------------------------------------------------------------
CClientVirtualReality::CClientVirtualReality()
{
m_PlayerTorsoOrigin.Init();
m_PlayerTorsoAngle.Init();
m_WorldFromWeapon.Identity();
m_WorldFromMidEye.Identity();
m_bOverrideTorsoAngle = false;
m_OverrideTorsoOffset.Init();
// Also reset our model of the player's torso orientation
m_PlayerTorsoAngle.Init ( 0.0f, 0.0f, 0.0f );
m_WorldZoomScale = 1.0f;
m_hmmMovementActual = HMM_SHOOTFACE_MOVEFACE;
m_iAlignTorsoAndViewToWeaponCountdown = 0;
m_rtLastMotionSample = 0;
m_bMotionUpdated = false;
#if defined( USE_SDL )
m_nNonVRSDLDisplayIndex = 0;
#endif
}
CClientVirtualReality::~CClientVirtualReality()
{
}
// --------------------------------------------------------------------
// Purpose:
// --------------------------------------------------------------------
bool CClientVirtualReality::Connect( CreateInterfaceFn factory )
{
if ( !factory )
return false;
if ( !BaseClass::Connect( factory ) )
return false;
return true;
}
// --------------------------------------------------------------------
// Purpose:
// --------------------------------------------------------------------
void CClientVirtualReality::Disconnect()
{
BaseClass::Disconnect();
}
// --------------------------------------------------------------------
// Purpose:
// --------------------------------------------------------------------
void * CClientVirtualReality::QueryInterface( const char *pInterfaceName )
{
CreateInterfaceFn factory = Sys_GetFactoryThis(); // This silly construction is necessary
return factory( pInterfaceName, NULL ); // to prevent the LTCG compiler from crashing.
}
// --------------------------------------------------------------------
// Purpose:
// --------------------------------------------------------------------
InitReturnVal_t CClientVirtualReality::Init()
{
InitReturnVal_t nRetVal = BaseClass::Init();
if ( nRetVal != INIT_OK )
return nRetVal;
return INIT_OK;
}
// --------------------------------------------------------------------
// Purpose:
// --------------------------------------------------------------------
void CClientVirtualReality::Shutdown()
{
BaseClass::Shutdown();
}
// --------------------------------------------------------------------
// Purpose: Draws the main menu in Stereo
// --------------------------------------------------------------------
void CClientVirtualReality::DrawMainMenu()
{
// have to draw the UI in stereo via the render texture or it won't fuse properly
// Draw it into the render target first
ITexture *pTexture = materials->FindTexture( "_rt_gui", NULL, false );
Assert( pTexture );
if( !pTexture)
return;
CMatRenderContextPtr pRenderContext( materials );
int viewActualWidth = pTexture->GetActualWidth();
int viewActualHeight = pTexture->GetActualHeight();
int viewWidth, viewHeight;
vgui::surface()->GetScreenSize( viewWidth, viewHeight );
// clear depth in the backbuffer before we push the render target
pRenderContext->ClearBuffers( false, true, true );
// constrain where VGUI can render to the view
pRenderContext->PushRenderTargetAndViewport( pTexture, NULL, 0, 0, viewActualWidth, viewActualHeight );
pRenderContext->OverrideAlphaWriteEnable( true, true );
// clear the render target
pRenderContext->ClearColor4ub( 0, 0, 0, 0 );
pRenderContext->ClearBuffers( true, false );
tmZone( TELEMETRY_LEVEL0, TMZF_NONE, "VGui_DrawHud", __FUNCTION__ );
// Make sure the client .dll root panel is at the proper point before doing the "SolveTraverse" calls
vgui::VPANEL root = enginevgui->GetPanel( PANEL_CLIENTDLL );
if ( root != 0 )
{
vgui::ipanel()->SetSize( root, viewWidth, viewHeight );
}
// Same for client .dll tools
root = enginevgui->GetPanel( PANEL_CLIENTDLL_TOOLS );
if ( root != 0 )
{
vgui::ipanel()->SetSize( root, viewWidth, viewHeight );
}
// paint the main menu and cursor
render->VGui_Paint( (PaintMode_t) ( PAINT_UIPANELS | PAINT_CURSOR ) );
pRenderContext->OverrideAlphaWriteEnable( false, true );
pRenderContext->PopRenderTargetAndViewport();
pRenderContext->Flush();
int leftX, leftY, leftW, leftH, rightX, rightY, rightW, rightH;
g_pSourceVR->GetViewportBounds( ISourceVirtualReality::VREye_Left, &leftX, &leftY, &leftW, &leftH );
g_pSourceVR->GetViewportBounds( ISourceVirtualReality::VREye_Right, &rightX, &rightY, &rightW, &rightH );
// render the main view
CViewSetup viewEye[STEREO_EYE_MAX];
viewEye[ STEREO_EYE_MONO ].zNear = 0.1;
viewEye[ STEREO_EYE_MONO ].zFar = 10000.f;
viewEye[ STEREO_EYE_MONO ].angles.Init();
viewEye[ STEREO_EYE_MONO ].origin.Zero();
viewEye[ STEREO_EYE_MONO ].x = viewEye[ STEREO_EYE_MONO ].m_nUnscaledX = leftX;
viewEye[ STEREO_EYE_MONO ].y = viewEye[ STEREO_EYE_MONO ].m_nUnscaledY = leftY;
viewEye[ STEREO_EYE_MONO ].width = viewEye[ STEREO_EYE_MONO ].m_nUnscaledWidth = leftW;
viewEye[ STEREO_EYE_MONO ].height = viewEye[ STEREO_EYE_MONO ].m_nUnscaledHeight = leftH;
viewEye[STEREO_EYE_LEFT] = viewEye[STEREO_EYE_RIGHT] = viewEye[ STEREO_EYE_MONO ] ;
viewEye[STEREO_EYE_LEFT].m_eStereoEye = STEREO_EYE_LEFT;
viewEye[STEREO_EYE_RIGHT].x = rightX;
viewEye[STEREO_EYE_RIGHT].y = rightY;
viewEye[STEREO_EYE_RIGHT].m_eStereoEye = STEREO_EYE_RIGHT;
// let sourcevr.dll tell us where to put the cameras
ProcessCurrentTrackingState( 0 );
Vector vViewModelOrigin;
QAngle qViewModelAngles;
OverrideView( &viewEye[ STEREO_EYE_MONO ] , &vViewModelOrigin, &qViewModelAngles, HMM_NOOVERRIDE );
g_ClientVirtualReality.OverrideStereoView( &viewEye[ STEREO_EYE_MONO ] , &viewEye[STEREO_EYE_LEFT], &viewEye[STEREO_EYE_RIGHT] );
// render both eyes
for( int nView = STEREO_EYE_LEFT; nView <= STEREO_EYE_RIGHT; nView++ )
{
CMatRenderContextPtr pRenderContextMat( materials );
PIXEvent pixEvent( pRenderContextMat, nView == STEREO_EYE_LEFT ? "left eye" : "right eye" );
ITexture *pColor = g_pSourceVR->GetRenderTarget( (ISourceVirtualReality::VREye)(nView-1), ISourceVirtualReality::RT_Color );
ITexture *pDepth = g_pSourceVR->GetRenderTarget( (ISourceVirtualReality::VREye)(nView-1), ISourceVirtualReality::RT_Depth );
render->Push3DView( viewEye[nView], VIEW_CLEAR_DEPTH|VIEW_CLEAR_COLOR, pColor, NULL, pDepth );
RenderHUDQuad( false, false );
render->PopView( NULL );
PostProcessFrame( (StereoEye_t)nView );
OverlayHUDQuadWithUndistort( viewEye[nView], true, true, false );
}
}
// --------------------------------------------------------------------
// Purpose:
// Offset the incoming view appropriately.
// Set up the "middle eye" from that.
// --------------------------------------------------------------------
bool CClientVirtualReality::OverrideView ( CViewSetup *pViewMiddle, Vector *pViewModelOrigin, QAngle *pViewModelAngles, HeadtrackMovementMode_t hmmMovementOverride )
{
if( !UseVR() )
{
return false;
}
if ( hmmMovementOverride == HMM_NOOVERRIDE )
{
if ( CurrentlyZoomed() )
{
m_hmmMovementActual = static_cast<HeadtrackMovementMode_t>( vr_moveaim_mode_zoom.GetInt() );
}
else
{
m_hmmMovementActual = static_cast<HeadtrackMovementMode_t>( vr_moveaim_mode.GetInt() );
}
}
else
{
m_hmmMovementActual = hmmMovementOverride;
}
// Incoming data may or may not be useful - it is the origin and aim of the "player", i.e. where bullets come from.
// In some modes it is an independent thing, guided by the mouse & keyboard = useful.
// In other modes it's just where the HMD was pointed last frame, modified slightly by kbd+mouse.
// In those cases, we should use our internal reference (which keeps track thanks to OverridePlayerMotion)
QAngle originalMiddleAngles = pViewMiddle->angles;
Vector originalMiddleOrigin = pViewMiddle->origin;
// Figure out the in-game "torso" concept, which corresponds to the player's physical torso.
m_PlayerTorsoOrigin = pViewMiddle->origin;
// Ignore what was passed in - it's just the direction the weapon is pointing, which was determined by last frame's HMD orientation!
// Instead use our cached value.
QAngle torsoAngles = m_PlayerTorsoAngle;
VMatrix worldFromTorso;
worldFromTorso.SetupMatrixOrgAngles( m_PlayerTorsoOrigin, torsoAngles );
//// Scale translation e.g. to allow big in-game leans with only a small head movement.
//// Clamp HMD movement to a reasonable amount to avoid wallhacks, vis problems, etc.
float limit = vr_translation_limit.GetFloat();
VMatrix matMideyeZeroFromMideyeCurrent = g_pSourceVR->GetMideyePose();
Vector viewTranslation = matMideyeZeroFromMideyeCurrent.GetTranslation();
if ( viewTranslation.IsLengthGreaterThan ( limit ) )
{
viewTranslation.NormalizeInPlace();
viewTranslation *= limit;
matMideyeZeroFromMideyeCurrent.SetTranslation( viewTranslation );
}
// Now figure out the three principal matrices: m_TorsoFromMideye, m_WorldFromMidEye, m_WorldFromWeapon
// m_TorsoFromMideye is done so that OverridePlayerMotion knows what to do with WASD.
switch ( m_hmmMovementActual )
{
case HMM_SHOOTFACE_MOVEFACE:
case HMM_SHOOTFACE_MOVETORSO:
// Aim point is down your nose, i.e. same as the view angles.
m_TorsoFromMideye = matMideyeZeroFromMideyeCurrent;
m_WorldFromMidEye = worldFromTorso * matMideyeZeroFromMideyeCurrent;
m_WorldFromWeapon = m_WorldFromMidEye;
break;
case HMM_SHOOTBOUNDEDMOUSE_LOOKFACE_MOVEFACE:
case HMM_SHOOTBOUNDEDMOUSE_LOOKFACE_MOVEMOUSE:
case HMM_SHOOTMOUSE_MOVEFACE:
case HMM_SHOOTMOVEMOUSE_LOOKFACE:
// Aim point is independent of view - leave it as it was, just copy it into m_WorldFromWeapon for our use.
m_TorsoFromMideye = matMideyeZeroFromMideyeCurrent;
m_WorldFromMidEye = worldFromTorso * matMideyeZeroFromMideyeCurrent;
m_WorldFromWeapon.SetupMatrixOrgAngles( originalMiddleOrigin, originalMiddleAngles );
break;
case HMM_SHOOTMOVELOOKMOUSE:
// HMD is ignored completely, mouse does everything.
m_PlayerTorsoAngle = originalMiddleAngles;
worldFromTorso.SetupMatrixOrgAngles( m_PlayerTorsoOrigin, originalMiddleAngles );
m_TorsoFromMideye.Identity();
m_WorldFromMidEye = worldFromTorso;
m_WorldFromWeapon = worldFromTorso;
break;
case HMM_SHOOTMOVELOOKMOUSEFACE:
// mouse does everything, and then we add head tracking on top of that
worldFromTorso = worldFromTorso * matMideyeZeroFromMideyeCurrent;
m_TorsoFromMideye = matMideyeZeroFromMideyeCurrent;
m_WorldFromWeapon = worldFromTorso;
m_WorldFromMidEye = worldFromTorso;
break;
default: Assert ( false ); break;
}
// Finally convert back to origin+angles that the game understands.
pViewMiddle->origin = m_WorldFromMidEye.GetTranslation();
VectorAngles ( m_WorldFromMidEye.GetForward(), m_WorldFromMidEye.GetUp(), pViewMiddle->angles );
*pViewModelAngles = pViewMiddle->angles;
if ( vr_viewmodel_translate_with_head.GetBool() )
{
*pViewModelOrigin = pViewMiddle->origin;
}
else
{
*pViewModelOrigin = originalMiddleOrigin;
}
m_WorldFromMidEyeNoDebugCam = m_WorldFromMidEye;
if ( vr_debug_remote_cam.GetBool() )
{
Vector vOffset ( vr_debug_remote_cam_pos_x.GetFloat(), vr_debug_remote_cam_pos_y.GetFloat(), vr_debug_remote_cam_pos_z.GetFloat() );
Vector vLookat ( vr_debug_remote_cam_target_x.GetFloat(), vr_debug_remote_cam_target_y.GetFloat(), vr_debug_remote_cam_target_z.GetFloat() );
pViewMiddle->origin += vOffset;
Vector vView = vLookat - vOffset;
VectorAngles ( vView, m_WorldFromMidEye.GetUp(), pViewMiddle->angles );
m_WorldFromMidEye.SetupMatrixOrgAngles( pViewMiddle->origin, pViewMiddle->angles );
m_TorsoFromMideye.Identity();
}
// set the near clip plane so the local player clips less
pViewMiddle->zNear *= vr_projection_znear_multiplier.GetFloat();
return true;
}
// --------------------------------------------------------------------
// Purpose:
// In some aim/move modes, the HUD aim reticle lags because it's
// using slightly stale data. This will feed it the newest data.
// --------------------------------------------------------------------
bool CClientVirtualReality::OverrideWeaponHudAimVectors ( Vector *pAimOrigin, Vector *pAimDirection )
{
if( !UseVR() )
{
return false;
}
Assert ( pAimOrigin != NULL );
Assert ( pAimDirection != NULL );
// So give it some nice high-fps numbers, not the low-fps ones we get from the game.
*pAimOrigin = m_WorldFromWeapon.GetTranslation();
*pAimDirection = m_WorldFromWeapon.GetForward();
return true;
}
// --------------------------------------------------------------------
// Purpose:
// Set up the left and right eyes from the middle eye if stereo is on.
// Advise calling soonish after OverrideView().
// --------------------------------------------------------------------
bool CClientVirtualReality::OverrideStereoView( CViewSetup *pViewMiddle, CViewSetup *pViewLeft, CViewSetup *pViewRight )
{
// Everything in here is in Source coordinate space.
if( !UseVR() )
{
return false;
}
VMatrix matOffsetLeft = g_pSourceVR->GetMidEyeFromEye( ISourceVirtualReality::VREye_Left );
VMatrix matOffsetRight = g_pSourceVR->GetMidEyeFromEye( ISourceVirtualReality::VREye_Right );
// Move eyes to IPD positions.
VMatrix worldFromLeftEye = m_WorldFromMidEye * matOffsetLeft;
VMatrix worldFromRightEye = m_WorldFromMidEye * matOffsetRight;
Assert ( IsOrthonormal ( worldFromLeftEye, 0.001f ) );
Assert ( IsOrthonormal ( worldFromRightEye, 0.001f ) );
// Finally convert back to origin+angles.
MatrixAngles( worldFromLeftEye.As3x4(), pViewLeft->angles, pViewLeft->origin );
MatrixAngles( worldFromRightEye.As3x4(), pViewRight->angles, pViewRight->origin );
// Find the projection matrices.
// TODO: this isn't the fastest thing in the world. Cache them?
float headtrackFovScale = m_WorldZoomScale;
pViewLeft->m_bViewToProjectionOverride = true;
pViewRight->m_bViewToProjectionOverride = true;
g_pSourceVR->GetEyeProjectionMatrix ( &pViewLeft->m_ViewToProjection, ISourceVirtualReality::VREye_Left, pViewMiddle->zNear, pViewMiddle->zFar, 1.0f/headtrackFovScale );
g_pSourceVR->GetEyeProjectionMatrix ( &pViewRight->m_ViewToProjection, ISourceVirtualReality::VREye_Right, pViewMiddle->zNear, pViewMiddle->zFar, 1.0f/headtrackFovScale );
// And bodge together some sort of average for our cyclops friends.
pViewMiddle->m_bViewToProjectionOverride = true;
for ( int i = 0; i < 4; i++ )
{
for ( int j = 0; j < 4; j++ )
{
pViewMiddle->m_ViewToProjection.m[i][j] = (pViewLeft->m_ViewToProjection.m[i][j] + pViewRight->m_ViewToProjection.m[i][j] ) * 0.5f;
}
}
switch ( vr_stereo_mono_set_eye.GetInt() )
{
case 0:
// ... nothing.
break;
case 1:
// Override all eyes with left
*pViewMiddle = *pViewLeft;
*pViewRight = *pViewLeft;
pViewRight->m_eStereoEye = STEREO_EYE_RIGHT;
break;
case 2:
// Override all eyes with right
*pViewMiddle = *pViewRight;
*pViewLeft = *pViewRight;
pViewLeft->m_eStereoEye = STEREO_EYE_LEFT;
break;
case 3:
// Override all eyes with middle
*pViewRight = *pViewMiddle;
*pViewLeft = *pViewMiddle;
pViewLeft->m_eStereoEye = STEREO_EYE_LEFT;
pViewRight->m_eStereoEye = STEREO_EYE_RIGHT;
break;
}
// To make culling work correctly, calculate the widest FOV of each projection matrix.
CalcFovFromProjection ( &(pViewLeft ->fov), pViewLeft ->m_ViewToProjection );
CalcFovFromProjection ( &(pViewRight ->fov), pViewRight ->m_ViewToProjection );
CalcFovFromProjection ( &(pViewMiddle->fov), pViewMiddle->m_ViewToProjection );
// if we don't know the HUD FOV, figure that out now
if( m_fHudHorizontalFov == 0.f )
{
// Figure out the current HUD FOV.
m_fHudHorizontalFov = pViewLeft->fov * vr_hud_display_ratio.GetFloat();
if( m_fHudHorizontalFov > vr_hud_max_fov.GetFloat() )
{
m_fHudHorizontalFov = vr_hud_max_fov.GetFloat();
}
}
// remember the view angles so we can limit the weapon to something near those
m_PlayerViewAngle = pViewMiddle->angles;
m_PlayerViewOrigin = pViewMiddle->origin;
// Figure out the HUD vectors and frustum.
// The aspect ratio of the HMD may be something bizarre (e.g. Rift is 640x800), and the pixels may not be square, so don't use that!
static const float fAspectRatio = 4.f/3.f;
float fHFOV = m_fHudHorizontalFov;
float fVFOV = m_fHudHorizontalFov / fAspectRatio;
const float fHudForward = vr_hud_forward.GetFloat();
m_fHudHalfWidth = tan( DEG2RAD( fHFOV * 0.5f ) ) * fHudForward * m_WorldZoomScale;
m_fHudHalfHeight = tan( DEG2RAD( fVFOV * 0.5f ) ) * fHudForward * m_WorldZoomScale;
QAngle HudAngles;
switch ( m_hmmMovementActual )
{
case HMM_SHOOTFACE_MOVETORSO:
// Put the HUD in front of the player's torso.
// This helps keep you oriented about where "forwards" is, which is otherwise surprisingly tricky!
// TODO: try preserving roll and/or pitch from the view?
HudAngles = m_PlayerTorsoAngle;
break;
case HMM_SHOOTFACE_MOVEFACE:
case HMM_SHOOTMOUSE_MOVEFACE:
case HMM_SHOOTMOVEMOUSE_LOOKFACE:
case HMM_SHOOTMOVELOOKMOUSE:
case HMM_SHOOTMOVELOOKMOUSEFACE:
case HMM_SHOOTBOUNDEDMOUSE_LOOKFACE_MOVEFACE:
case HMM_SHOOTBOUNDEDMOUSE_LOOKFACE_MOVEMOUSE:
// Put the HUD in front of wherever the player is looking.
HudAngles = m_PlayerViewAngle;
break;
default: Assert ( false ); break;
}
// This is a bitfield. A set bit means lock to the world, a clear bit means don't.
int iVrHudAxisLockToWorld = vr_hud_axis_lock_to_world.GetInt();
if ( ( iVrHudAxisLockToWorld & (1<<ROLL) ) != 0 )
{
HudAngles[ROLL] = 0.0f;
}
if ( ( iVrHudAxisLockToWorld & (1<<PITCH) ) != 0 )
{
HudAngles[PITCH] = 0.0f;
}
if ( ( iVrHudAxisLockToWorld & (1<<YAW) ) != 0 )
{
// Locking the yaw to the world is not particularly helpful, so what it actually means is lock it to the weapon.
QAngle aimAngles;
MatrixAngles( m_WorldFromWeapon.As3x4(), aimAngles );
HudAngles[YAW] = aimAngles[YAW];
}
m_WorldFromHud.SetupMatrixOrgAngles( m_PlayerViewOrigin, HudAngles );
// Remember in source X forwards, Y left, Z up.
// We need to transform to a more conventional X right, Y up, Z backwards before doing the projection.
VMatrix WorldFromHudView;
WorldFromHudView./*X vector*/SetForward ( -m_WorldFromHud.GetLeft() );
WorldFromHudView./*Y vector*/SetLeft ( m_WorldFromHud.GetUp() );
WorldFromHudView./*Z vector*/SetUp ( -m_WorldFromHud.GetForward() );
WorldFromHudView.SetTranslation ( m_PlayerViewOrigin );
VMatrix HudProjection;
HudProjection.Identity();
HudProjection.m[0][0] = fHudForward / m_fHudHalfWidth;
HudProjection.m[1][1] = fHudForward / m_fHudHalfHeight;
// Z vector is not used/valid, but w is for projection.
HudProjection.m[3][2] = -1.0f;
// This will transform a world point into a homogeneous vector that
// when projected (i.e. divide by w) maps to HUD space [-1,1]
m_HudProjectionFromWorld = HudProjection * WorldFromHudView.InverseTR();
return true;
}
// --------------------------------------------------------------------
// Purpose: Updates player orientation, position and motion according
// to HMD status.
// --------------------------------------------------------------------
bool CClientVirtualReality::OverridePlayerMotion( float flInputSampleFrametime, const QAngle &oldAngles, const QAngle &curAngles, const Vector &curMotion, QAngle *pNewAngles, Vector *pNewMotion )
{
Assert ( pNewAngles != NULL );
Assert ( pNewMotion != NULL );
*pNewAngles = curAngles;
*pNewMotion = curMotion;
if ( !UseVR() )
{
return false;
}
m_bMotionUpdated = true;
// originalAngles tells us what the weapon angles were before whatever mouse, joystick, etc thing changed them - called "old"
// curAngles holds the new weapon angles after mouse, joystick, etc. applied.
// We need to compute what weapon angles WE want and return them in *pNewAngles - called "new"
VMatrix worldFromTorso;
// Whatever position is already here (set up by OverrideView) needs to be preserved.
Vector vWeaponOrigin = m_WorldFromWeapon.GetTranslation();
switch ( m_hmmMovementActual )
{
case HMM_SHOOTFACE_MOVEFACE:
case HMM_SHOOTFACE_MOVETORSO:
{
// Figure out what changes were made to the WEAPON by mouse/joystick/etc
VMatrix worldFromOldWeapon, worldFromCurWeapon;
worldFromOldWeapon.SetupMatrixAngles( oldAngles );
worldFromCurWeapon.SetupMatrixAngles( curAngles );
// We ignore mouse pitch, the mouse can't do rolls, so it's just yaw changes.
if( !m_bOverrideTorsoAngle )
{
m_PlayerTorsoAngle[YAW] += curAngles[YAW] - oldAngles[YAW];
m_PlayerTorsoAngle[ROLL] = 0.0f;
m_PlayerTorsoAngle[PITCH] = 0.0f;
}
worldFromTorso.SetupMatrixAngles( m_PlayerTorsoAngle );
// Weapon view = mideye view, so apply that to the torso to find the world view direction.
m_WorldFromWeapon = worldFromTorso * m_TorsoFromMideye;
// ...and we return this new weapon direction as the player's orientation.
MatrixAngles( m_WorldFromWeapon.As3x4(), *pNewAngles );
// Restore the translation.
m_WorldFromWeapon.SetTranslation ( vWeaponOrigin );
}
break;
case HMM_SHOOTMOVELOOKMOUSEFACE:
case HMM_SHOOTMOVEMOUSE_LOOKFACE:
case HMM_SHOOTMOVELOOKMOUSE:
{
// The mouse just controls the weapon directly.
*pNewAngles = curAngles;
*pNewMotion = curMotion;
// Move the torso by the yaw angles - torso should not have roll or pitch or you'll make folks ill.
if( !m_bOverrideTorsoAngle && m_hmmMovementActual != HMM_SHOOTMOVELOOKMOUSEFACE )
{
m_PlayerTorsoAngle[YAW] = curAngles[YAW];
m_PlayerTorsoAngle[ROLL] = 0.0f;
m_PlayerTorsoAngle[PITCH] = 0.0f;
}
// Let every other system know.
m_WorldFromWeapon.SetupMatrixOrgAngles( vWeaponOrigin, *pNewAngles );
worldFromTorso.SetupMatrixAngles( m_PlayerTorsoAngle );
}
break;
case HMM_SHOOTBOUNDEDMOUSE_LOOKFACE_MOVEFACE:
case HMM_SHOOTBOUNDEDMOUSE_LOOKFACE_MOVEMOUSE:
{
// The mouse controls the weapon directly.
*pNewAngles = curAngles;
*pNewMotion = curMotion;
float fReticleYawLimit = vr_moveaim_reticle_yaw_limit.GetFloat();
float fReticlePitchLimit = vr_moveaim_reticle_pitch_limit.GetFloat();
if ( CurrentlyZoomed() )
{
fReticleYawLimit = vr_moveaim_reticle_yaw_limit_zoom.GetFloat() * m_WorldZoomScale;
fReticlePitchLimit = vr_moveaim_reticle_pitch_limit_zoom.GetFloat() * m_WorldZoomScale;
if ( fReticleYawLimit > 180.0f )
{
fReticleYawLimit = 180.0f;
}
if ( fReticlePitchLimit > 180.0f )
{
fReticlePitchLimit = 180.0f;
}
}
if ( fReticlePitchLimit >= 0.0f )
{
// Clamp pitch to within the limits.
(*pNewAngles)[PITCH] = Clamp ( curAngles[PITCH], m_PlayerViewAngle[PITCH] - fReticlePitchLimit, m_PlayerViewAngle[PITCH] + fReticlePitchLimit );
}
// For yaw the concept here is the torso stays within a set number of degrees of the weapon in yaw.
// However, with drifty tracking systems (e.g. IMUs) the concept of "torso" is hazy.
// Really it's just a mechanism to turn the view without moving the head - its absolute
// orientation is not that useful.
// So... if the mouse is to the right greater than the chosen angle from the view, and then
// moves more right, it will drag the torso (and thus the view) right, so it stays on the edge of the view.
// But if it moves left towards the view, it does no dragging.
// Note that if the mouse does not move, but the view moves, it will NOT drag at all.
// This allows people to mouse-aim within their view, but also to flick-turn with the mouse,
// and to flick-glance with the head.
if ( fReticleYawLimit >= 0.0f )
{
float fViewToWeaponYaw = AngleDiff ( curAngles[YAW], m_PlayerViewAngle[YAW] );
float fWeaponYawMovement = AngleDiff ( curAngles[YAW], oldAngles[YAW] );
if ( fViewToWeaponYaw > fReticleYawLimit )
{
if ( fWeaponYawMovement > 0.0f )
{
m_PlayerTorsoAngle[YAW] += fWeaponYawMovement;
}
}
else if ( fViewToWeaponYaw < -fReticleYawLimit )
{
if ( fWeaponYawMovement < 0.0f )
{
m_PlayerTorsoAngle[YAW] += fWeaponYawMovement;
}
}
}
// Let every other system know.
m_WorldFromWeapon.SetupMatrixOrgAngles( vWeaponOrigin, *pNewAngles );
worldFromTorso.SetupMatrixAngles( m_PlayerTorsoAngle );
}
break;
case HMM_SHOOTMOUSE_MOVEFACE:
{
(*pNewAngles)[PITCH] = clamp( (*pNewAngles)[PITCH], m_PlayerViewAngle[PITCH]-15.f, m_PlayerViewAngle[PITCH]+15.f );
float fDiff = AngleDiff( (*pNewAngles)[YAW], m_PlayerViewAngle[YAW] );
if( fDiff > 15.f )
{
(*pNewAngles)[YAW] = AngleNormalize( m_PlayerViewAngle[YAW] + 15.f );
if( !m_bOverrideTorsoAngle )
m_PlayerTorsoAngle[ YAW ] += fDiff - 15.f;
}
else if( fDiff < -15.f )
{
(*pNewAngles)[YAW] = AngleNormalize( m_PlayerViewAngle[YAW] - 15.f );
if( !m_bOverrideTorsoAngle )
m_PlayerTorsoAngle[ YAW ] += fDiff + 15.f;
}
else
{
m_PlayerTorsoAngle[ YAW ] += AngleDiff( curAngles[YAW], oldAngles[YAW] ) /2.f;
}
m_WorldFromWeapon.SetupMatrixOrgAngles( vWeaponOrigin, *pNewAngles );
worldFromTorso.SetupMatrixAngles( m_PlayerTorsoAngle );
}
break;
default: Assert ( false ); break;
}
// Figure out player motion.
switch ( m_hmmMovementActual )
{
case HMM_SHOOTBOUNDEDMOUSE_LOOKFACE_MOVEFACE:
{
// The motion passed in is meant to be relative to the face, so jimmy it to be relative to the new weapon aim.
VMatrix mideyeFromWorld = m_WorldFromMidEye.InverseTR();
VMatrix newMidEyeFromWeapon = mideyeFromWorld * m_WorldFromWeapon;
newMidEyeFromWeapon.SetTranslation ( Vector ( 0.0f, 0.0f, 0.0f ) );
*pNewMotion = newMidEyeFromWeapon * curMotion;
}
break;
case HMM_SHOOTFACE_MOVETORSO:
{
// The motion passed in is meant to be relative to the torso, so jimmy it to be relative to the new weapon aim.
VMatrix torsoFromWorld = worldFromTorso.InverseTR();
VMatrix newTorsoFromWeapon = torsoFromWorld * m_WorldFromWeapon;
newTorsoFromWeapon.SetTranslation ( Vector ( 0.0f, 0.0f, 0.0f ) );
*pNewMotion = newTorsoFromWeapon * curMotion;
}
break;
case HMM_SHOOTBOUNDEDMOUSE_LOOKFACE_MOVEMOUSE:
case HMM_SHOOTMOVELOOKMOUSEFACE:
case HMM_SHOOTFACE_MOVEFACE:
case HMM_SHOOTMOUSE_MOVEFACE:
case HMM_SHOOTMOVEMOUSE_LOOKFACE:
case HMM_SHOOTMOVELOOKMOUSE:
// Motion is meant to be relative to the weapon, so we're fine.
*pNewMotion = curMotion;
break;
default: Assert ( false ); break;
}
// If the game told us to, recenter the torso yaw to match the weapon
if ( m_iAlignTorsoAndViewToWeaponCountdown > 0 )
{
m_iAlignTorsoAndViewToWeaponCountdown--;
// figure out the angles from the torso to the head
QAngle torsoFromHeadAngles;
MatrixAngles( m_TorsoFromMideye.As3x4(), torsoFromHeadAngles );
QAngle weaponAngles;
MatrixAngles( m_WorldFromWeapon.As3x4(), weaponAngles );
m_PlayerTorsoAngle[ YAW ] = weaponAngles[ YAW ] - torsoFromHeadAngles[ YAW ] ;
NormalizeAngles( m_PlayerTorsoAngle );
}
// remember the motion for stat tracking
m_PlayerLastMovement = *pNewMotion;
return true;
}
// --------------------------------------------------------------------
// Purpose: Returns true if the world is zoomed
// --------------------------------------------------------------------
bool CClientVirtualReality::CurrentlyZoomed()
{
return ( m_WorldZoomScale != 1.0f );
}
// --------------------------------------------------------------------
// Purpose: Tells the headtracker to keep the torso angle of the player
// fixed at this point until the game tells us something
// different.
// --------------------------------------------------------------------
void CClientVirtualReality::OverrideTorsoTransform( const Vector & position, const QAngle & angles )
{
if( m_iAlignTorsoAndViewToWeaponCountdown > 0 )
{
m_iAlignTorsoAndViewToWeaponCountdown--;
// figure out the angles from the torso to the head
QAngle torsoFromHeadAngles;
MatrixAngles( m_TorsoFromMideye.As3x4(), torsoFromHeadAngles );
// this is how far off the torso we actually set will need to be to keep the current "forward"
// vector while the torso angle is being overridden.
m_OverrideTorsoOffset[ YAW ] = -torsoFromHeadAngles[ YAW ];
}
m_bOverrideTorsoAngle = true;
m_OverrideTorsoAngle = angles + m_OverrideTorsoOffset;
// overriding pitch and roll isn't allowed to avoid making people sick
m_OverrideTorsoAngle[ PITCH ] = 0;
m_OverrideTorsoAngle[ ROLL ] = 0;
NormalizeAngles( m_OverrideTorsoAngle );
m_PlayerTorsoAngle = m_OverrideTorsoAngle;
}
// --------------------------------------------------------------------
// Purpose: Tells the headtracker to resume using its own notion of
// where the torso is pointed.
// --------------------------------------------------------------------
void CClientVirtualReality::CancelTorsoTransformOverride()
{
m_bOverrideTorsoAngle = false;
}
bool CClientVirtualReality::CanOverlayHudQuad()
{
bool bCanOverlay = true;
bCanOverlay = bCanOverlay && vr_render_hud_in_world.GetBool();
bCanOverlay = bCanOverlay && ( ! vr_hud_never_overlay.GetBool() );
bCanOverlay = bCanOverlay && ( vr_hud_axis_lock_to_world.GetInt() == 0 );
bCanOverlay = bCanOverlay && ( m_hmmMovementActual != HMM_SHOOTFACE_MOVETORSO );
return bCanOverlay;
}
// --------------------------------------------------------------------
// Purpose: Returns the bounds in world space where the game should
// position the HUD.
// --------------------------------------------------------------------
void CClientVirtualReality::GetHUDBounds( Vector *pViewer, Vector *pUL, Vector *pUR, Vector *pLL, Vector *pLR )
{
Vector vHalfWidth = m_WorldFromHud.GetLeft() * -m_fHudHalfWidth;
Vector vHalfHeight = m_WorldFromHud.GetUp() * m_fHudHalfHeight;
Vector vHUDOrigin = m_PlayerViewOrigin + m_WorldFromHud.GetForward() * vr_hud_forward.GetFloat();
*pViewer = m_PlayerViewOrigin;
*pUL = vHUDOrigin - vHalfWidth + vHalfHeight;
*pUR = vHUDOrigin + vHalfWidth + vHalfHeight;
*pLL = vHUDOrigin - vHalfWidth - vHalfHeight;
*pLR = vHUDOrigin + vHalfWidth - vHalfHeight;
}
// --------------------------------------------------------------------
// Purpose: Renders the HUD in the world.
// --------------------------------------------------------------------
void CClientVirtualReality::RenderHUDQuad( bool bBlackout, bool bTranslucent )
{
// If we can overlay the HUD directly onto the target later, we'll do that instead (higher image quality).
if ( CanOverlayHudQuad() )
return;
Vector vHead, vUL, vUR, vLL, vLR;
GetHUDBounds ( &vHead, &vUL, &vUR, &vLL, &vLR );
CMatRenderContextPtr pRenderContext( materials );
{
IMaterial *mymat = NULL;
if ( bTranslucent )
{
mymat = materials->FindMaterial( "vgui/inworldui", TEXTURE_GROUP_VGUI );
}
else
{
mymat = materials->FindMaterial( "vgui/inworldui_opaque", TEXTURE_GROUP_VGUI );
}
Assert( !mymat->IsErrorMaterial() );
IMesh *pMesh = pRenderContext->GetDynamicMesh( true, NULL, NULL, mymat );
CMeshBuilder meshBuilder;
meshBuilder.Begin( pMesh, MATERIAL_TRIANGLE_STRIP, 2 );
meshBuilder.Position3fv (vLR.Base() );
meshBuilder.TexCoord2f( 0, 1, 1 );
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 1>();
meshBuilder.Position3fv (vLL.Base());
meshBuilder.TexCoord2f( 0, 0, 1 );
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 1>();
meshBuilder.Position3fv (vUR.Base());
meshBuilder.TexCoord2f( 0, 1, 0 );
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 1>();
meshBuilder.Position3fv (vUL.Base());
meshBuilder.TexCoord2f( 0, 0, 0 );
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 1>();
meshBuilder.End();
pMesh->Draw();
}
if( bBlackout )
{
Vector vbUL, vbUR, vbLL, vbLR;
// "Reflect" the HUD bounds through the viewer to find the ones behind the head.
vbUL = 2 * vHead - vLR;
vbUR = 2 * vHead - vLL;
vbLL = 2 * vHead - vUR;
vbLR = 2 * vHead - vUL;
IMaterial *mymat = materials->FindMaterial( "vgui/black", TEXTURE_GROUP_VGUI );
IMesh *pMesh = pRenderContext->GetDynamicMesh( true, NULL, NULL, mymat );
// Tube around the outside.
CMeshBuilder meshBuilder;
meshBuilder.Begin( pMesh, MATERIAL_TRIANGLE_STRIP, 8 );
meshBuilder.Position3fv (vLR.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vbLR.Base() );
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vLL.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vbLL.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vUL.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vbUL.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vUR.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vbUR.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vLR.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vbLR.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.End();
pMesh->Draw();
// Cap behind the viewer.
meshBuilder.Begin( pMesh, MATERIAL_TRIANGLE_STRIP, 2 );
meshBuilder.Position3fv (vbUR.Base() );
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vbUL.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vbLR.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.Position3fv (vbLL.Base());
meshBuilder.AdvanceVertexF<VTX_HAVEPOS, 0>();
meshBuilder.End();
pMesh->Draw();
}
}
// --------------------------------------------------------------------
// Purpose: Gets the amount of zoom to apply
// --------------------------------------------------------------------
float CClientVirtualReality::GetZoomedModeMagnification()
{
return m_WorldZoomScale * vr_zoom_scope_scale.GetFloat();
}
// --------------------------------------------------------------------
// Purpose: Does some client-side tracking work and then tells headtrack
// to do its own work.
// --------------------------------------------------------------------
bool CClientVirtualReality::ProcessCurrentTrackingState( float fGameFOV )
{
m_WorldZoomScale = 1.0f;
if ( fGameFOV != 0.0f )
{
// To compensate for the lack of pixels on most HUDs, let's grow this a bit.
// Remember that MORE zoom equals LESS fov!
fGameFOV *= ( 1.0f / vr_zoom_multiplier.GetFloat() );
fGameFOV = Min ( fGameFOV, 170.0f );
// The game has overridden the FOV, e.g. because of a sniper scope. So we need to match this view with whatever actual FOV the HUD has.
float wantedGameTanfov = tanf ( DEG2RAD ( fGameFOV * 0.5f ) );
// OK, so now in stereo mode, we're going to also draw an overlay, but that overlay usually covers more of the screen (because in a good HMD usually our actual FOV is much wider)
float overlayActualPhysicalTanfov = tanf ( DEG2RAD ( m_fHudHorizontalFov * 0.5f ) );
// Therefore... (remembering that a zoom > 1.0 means you zoom *out*)
m_WorldZoomScale = wantedGameTanfov / overlayActualPhysicalTanfov;
}
return g_pSourceVR->SampleTrackingState( fGameFOV, 0.f /* seconds to predict */ );
}
// --------------------------------------------------------------------
// Purpose: Returns the projection matrix to use for the HUD
// --------------------------------------------------------------------
const VMatrix &CClientVirtualReality::GetHudProjectionFromWorld()
{
// This matrix will transform a world-space position into a homogenous HUD-space vector.
// So if you divide x+y by w, you will get the position on the HUD in [-1,1] space.
return m_HudProjectionFromWorld;
}
// --------------------------------------------------------------------
// Purpose: Returns the aim vector relative to the torso
// --------------------------------------------------------------------
void CClientVirtualReality::GetTorsoRelativeAim( Vector *pPosition, QAngle *pAngles )
{
MatrixAngles( m_TorsoFromMideye.As3x4(), *pAngles, *pPosition );
pAngles->y += vr_aim_yaw_offset.GetFloat();
}
// --------------------------------------------------------------------
// Purpose: Returns distance of the HUD in front of the eyes.
// --------------------------------------------------------------------
float CClientVirtualReality::GetHUDDistance()
{
return vr_hud_forward.GetFloat();
}
// --------------------------------------------------------------------
// Purpose: Returns true if the HUD should be rendered into a render
// target and then into the world on a quad.
// --------------------------------------------------------------------
bool CClientVirtualReality::ShouldRenderHUDInWorld()
{
return UseVR() && vr_render_hud_in_world.GetBool();
}
// --------------------------------------------------------------------
// Purpose: Lets headtrack tweak the view model origin and angles to match
// aim angles and handle strange viewmode FOV stuff
// --------------------------------------------------------------------
void CClientVirtualReality::OverrideViewModelTransform( Vector & vmorigin, QAngle & vmangles, bool bUseLargeOverride )
{
Vector vForward, vRight, vUp;
AngleVectors( vmangles, &vForward, &vRight, &vUp );
float fForward = bUseLargeOverride ? vr_viewmodel_offset_forward_large.GetFloat() : vr_viewmodel_offset_forward.GetFloat();
vmorigin += vForward * fForward;
MatrixAngles( m_WorldFromWeapon.As3x4(), vmangles );
}
// --------------------------------------------------------------------
// Purpose: Tells the head tracker to reset the torso position in case
// we're on a drifty tracker.
// --------------------------------------------------------------------
void CClientVirtualReality::AlignTorsoAndViewToWeapon()
{
if( !UseVR() )
return;
if( g_pSourceVR->WillDriftInYaw() )
{
m_iAlignTorsoAndViewToWeaponCountdown = 2;
}
}
// --------------------------------------------------------------------
// Purpose: Lets VR do stuff at the very end of the rendering process
// --------------------------------------------------------------------
void CClientVirtualReality::PostProcessFrame( StereoEye_t eEye )
{
if( !UseVR() )
return;
g_pSourceVR->DoDistortionProcessing( eEye == STEREO_EYE_LEFT ? ISourceVirtualReality::VREye_Left : ISourceVirtualReality::VREye_Right );
}
// --------------------------------------------------------------------
// Pastes the HUD directly onto the backbuffer / render target.
// (higher quality than the RenderHUDQuad() path but can't always be used)
// --------------------------------------------------------------------
void CClientVirtualReality::OverlayHUDQuadWithUndistort( const CViewSetup &eyeView, bool bDoUndistort, bool bBlackout, bool bTranslucent )
{
if ( ! UseVR() )
return;
// If we can't overlay the HUD, it will be handled on another path (rendered into the scene with RenderHUDQuad()).
if ( ! CanOverlayHudQuad() )
return;
// Get the position of the HUD quad in world space as used by RenderHUDQuad(). Then convert to a rectangle in normalized
// device coordinates.
Vector vHead, vUL, vUR, vLL, vLR;
GetHUDBounds ( &vHead, &vUL, &vUR, &vLL, &vLR );
VMatrix worldToView, viewToProjection, worldToProjection, worldToPixels;
render->GetMatricesForView( eyeView, &worldToView, &viewToProjection, &worldToProjection, &worldToPixels );
Vector pUL, pUR, pLL, pLR;
worldToProjection.V3Mul( vUL, pUL );
worldToProjection.V3Mul( vUR, pUR );
worldToProjection.V3Mul( vLL, pLL );
worldToProjection.V3Mul( vLR, pLR );
float ndcHudBounds[4];
ndcHudBounds[0] = Min ( Min( pUL.x, pUR.x ), Min( pLL.x, pLR.x ) );
ndcHudBounds[1] = Min ( Min( pUL.y, pUR.y ), Min( pLL.y, pLR.y ) );
ndcHudBounds[2] = Max ( Max( pUL.x, pUR.x ), Max( pLL.x, pLR.x ) );
ndcHudBounds[3] = Max ( Max( pUL.y, pUR.y ), Max( pLL.y, pLR.y ) );
ISourceVirtualReality::VREye sourceVrEye = ( eyeView.m_eStereoEye == STEREO_EYE_LEFT ) ? ISourceVirtualReality::VREye_Left : ISourceVirtualReality::VREye_Right;
g_pSourceVR->CompositeHud ( sourceVrEye, ndcHudBounds, bDoUndistort, bBlackout, bTranslucent );
}
// --------------------------------------------------------------------
// Purpose: Switches to VR mode
// --------------------------------------------------------------------
void CClientVirtualReality::Activate()
{
// we can only do this if a headtrack DLL is loaded
if( !g_pSourceVR )
return;
// These checks don't apply if we're in VR mode because Steam said so.
if ( !ShouldForceVRActive() )
{
// see if VR mode is even enabled
if ( materials->GetCurrentConfigForVideoCard().m_nVRModeAdapter == -1 )
{
Warning( "Enable VR mode in the video options before trying to use it.\n" );
return;
}
// See if we have an actual adapter
int32 nVRModeAdapter = g_pSourceVR->GetVRModeAdapter();
if ( nVRModeAdapter == -1 )
{
Warning( "Unable to get VRMode adapter from OpenVR. VR mode cannot be enabled. Try restarting and then enabling VR again.\n" );
return;
}
// we can only activate if we've got a VR device
if ( materials->GetCurrentConfigForVideoCard().m_nVRModeAdapter != nVRModeAdapter )
{
Warning( "VR Mode expects adapter %d which is different from %d which we are currently using. Try restarting and enabling VR mode again.\n",
nVRModeAdapter, materials->GetCurrentConfigForVideoCard().m_nVRModeAdapter );
engine->ExecuteClientCmd( "mat_enable_vrmode 0\n" );
return;
}
}
// can't activate twice
if( UseVR() )
return;
// remember where we were
m_bNonVRWindowed = g_pMaterialSystem->GetCurrentConfigForVideoCard().Windowed();
vgui::surface()->GetScreenSize( m_nNonVRWidth, m_nNonVRHeight );
#if defined( USE_SDL )
static ConVarRef sdl_displayindex( "sdl_displayindex" );
m_nNonVRSDLDisplayIndex = sdl_displayindex.GetInt();
#endif
if( !g_pSourceVR->Activate() )
{
// we couldn't activate, so just punt on this whole thing
return;
}
// general all-game stuff
engine->ExecuteClientCmd( "mat_reset_rendertargets\n" );
// game specific VR config
CUtlString sCmd;
sCmd.Format( "exec sourcevr_%s.cfg\n", COM_GetModDirectory() );
engine->ExecuteClientCmd( sCmd.Get() );
vgui::surface()->SetSoftwareCursor( true );
#if defined(POSIX)
ConVarRef m_rawinput( "m_rawinput" );
m_bNonVRRawInput = m_rawinput.GetBool();
m_rawinput.SetValue( 1 );
ConVarRef mat_vsync( "mat_vsync" );
mat_vsync.SetValue( 0 );
#endif
g_pMatSystemSurface->ForceScreenSizeOverride(true, 640, 480 );
int nViewportWidth, nViewportHeight;
g_pSourceVR->GetViewportBounds( ISourceVirtualReality::VREye_Left, NULL, NULL, &nViewportWidth, &nViewportHeight );
g_pMatSystemSurface->SetFullscreenViewportAndRenderTarget( 0, 0, nViewportWidth, nViewportHeight, g_pSourceVR->GetRenderTarget( ISourceVirtualReality::VREye_Left, ISourceVirtualReality::RT_Color ) );
vgui::ivgui()->SetVRMode( true );
// we can skip this extra mode change if we've always been in VR mode
if ( !ShouldForceVRActive() )
{
VRRect_t rect;
if ( g_pSourceVR->GetDisplayBounds( &rect ) )
{
// set mode
char szCmd[256];
Q_snprintf( szCmd, sizeof(szCmd), "mat_setvideomode %i %i %i\n", rect.nWidth, rect.nHeight, vr_force_windowed.GetBool() ? 1 : 0 );
engine->ClientCmd_Unrestricted( szCmd );
}
}
}
void CClientVirtualReality::Deactivate()
{
// can't deactivate when we aren't active
if( !UseVR() )
return;
g_pSourceVR->Deactivate();
g_pMatSystemSurface->ForceScreenSizeOverride(false, 0, 0 );
g_pMaterialSystem->GetRenderContext()->Viewport( 0, 0, m_nNonVRWidth, m_nNonVRHeight );
g_pMatSystemSurface->SetFullscreenViewportAndRenderTarget( 0, 0, m_nNonVRWidth, m_nNonVRHeight, NULL );
static ConVarRef cl_software_cursor( "cl_software_cursor" );
vgui::surface()->SetSoftwareCursor( cl_software_cursor.GetBool() );
#if defined( USE_SDL )
static ConVarRef sdl_displayindex( "sdl_displayindex" );
sdl_displayindex.SetValue( m_nNonVRSDLDisplayIndex );
#endif
#if defined(POSIX)
ConVarRef m_rawinput( "m_rawinput" );
m_rawinput.SetValue( m_bNonVRRawInput );
#endif
// Make sure the client .dll root panel is at the proper point before doing the "SolveTraverse" calls
vgui::VPANEL root = enginevgui->GetPanel( PANEL_CLIENTDLL );
if ( root != 0 )
{
vgui::ipanel()->SetSize( root, m_nNonVRWidth, m_nNonVRHeight );
}
// Same for client .dll tools
root = enginevgui->GetPanel( PANEL_CLIENTDLL_TOOLS );
if ( root != 0 )
{
vgui::ipanel()->SetSize( root, m_nNonVRWidth, m_nNonVRHeight );
}
int viewWidth, viewHeight;
vgui::surface()->GetScreenSize( viewWidth, viewHeight );
engine->ExecuteClientCmd( "mat_reset_rendertargets\n" );
// set mode
char szCmd[ 256 ];
Q_snprintf( szCmd, sizeof( szCmd ), "mat_setvideomode %i %i %i\n", m_nNonVRWidth, m_nNonVRHeight, m_bNonVRWindowed ? 1 : 0 );
engine->ClientCmd_Unrestricted( szCmd );
}
// Called when startup is complete
void CClientVirtualReality::StartupComplete()
{
if ( vr_activate_default.GetBool() || ShouldForceVRActive() )
Activate();
}