- Use camera.getDirection() instead of manual Euler angle calculation to properly account for camera transform hierarchy - Negate forward offsets to position objects in -Z direction (user faces -Z by design) - Replace expensive HighlightLayer hover effect with lightweight EdgesRenderer (20-50x faster) - Add comprehensive debug logging for position calculations The camera has a parent transform with 180° Y rotation, causing the user to face -Z in world space. Components now correctly position in front of the user when entering XR. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
179 lines
7.6 KiB
TypeScript
179 lines
7.6 KiB
TypeScript
import {AbstractMesh, Vector3, WebXRDefaultExperience, WebXRMotionControllerManager, WebXRState} from "@babylonjs/core";
|
|
import log from "loglevel";
|
|
import {WebController} from "../../controllers/webController";
|
|
import {Rigplatform} from "../../controllers/rigplatform";
|
|
import {DiagramManager} from "../../diagram/diagramManager";
|
|
import {Spinner} from "../../objects/spinner";
|
|
import {getAppConfig} from "../appConfig";
|
|
import {Scene} from "@babylonjs/core";
|
|
|
|
|
|
export async function groundMeshObserver(ground: AbstractMesh,
|
|
diagramManager: DiagramManager,
|
|
spinner: Spinner) {
|
|
const logger = log.getLogger('groungMeshObserver');
|
|
WebXRMotionControllerManager.PrioritizeOnlineRepository = false;
|
|
WebXRMotionControllerManager.UseOnlineRepository = true;
|
|
const xr = await WebXRDefaultExperience.CreateAsync(ground.getScene(), {
|
|
floorMeshes: [ground],
|
|
disableHandTracking: true,
|
|
disableTeleportation: true,
|
|
disableDefaultUI: true,
|
|
disableNearInteraction: true,
|
|
outputCanvasOptions: {
|
|
canvasOptions: {
|
|
|
|
framebufferScaleFactor: 1
|
|
}
|
|
},
|
|
optionalFeatures: true,
|
|
|
|
pointerSelectionOptions: {
|
|
enablePointerSelectionOnAllControllers: true
|
|
}
|
|
});
|
|
window.addEventListener('enterXr', async () => {
|
|
await xr.baseExperience.enterXRAsync('immersive-vr', 'local-floor');
|
|
logger.debug("Entering XR Experience");
|
|
})
|
|
//xr.baseExperience.featuresManager.enableFeature(WebXRFeatureName.LAYERS, "latest", { preferMultiviewOnInit: true }, true, false);
|
|
|
|
if (spinner) {
|
|
spinner.hide();
|
|
}
|
|
|
|
xr.baseExperience.sessionManager.onXRSessionInit.add((session) => {
|
|
session.addEventListener('visibilitychange', (ev) => {
|
|
logger.debug(ev);
|
|
});
|
|
});
|
|
xr.baseExperience.sessionManager.onXRSessionEnded.add(() => {
|
|
logger.debug('session ended');
|
|
window.location.reload();
|
|
});
|
|
xr.baseExperience.onStateChangedObservable.add((state) => {
|
|
logger.debug(WebXRState[state]);
|
|
switch (state) {
|
|
case WebXRState.IN_XR:
|
|
ground.getScene().audioEnabled = true;
|
|
window.addEventListener(('pa-button-state-change'), (event: any) => {
|
|
if (event.detail) {
|
|
logger.debug(event.detail);
|
|
}
|
|
});
|
|
// Position components relative to camera on XR entry
|
|
positionComponentsRelativeToCamera(ground.getScene(), diagramManager);
|
|
break;
|
|
case WebXRState.EXITING_XR:
|
|
setTimeout(() => {
|
|
logger.debug('EXITING_XR, reloading');
|
|
window.location.reload();
|
|
}, 500);
|
|
|
|
}
|
|
});
|
|
|
|
const rig = new Rigplatform(xr, diagramManager);
|
|
const config = getAppConfig();
|
|
rig.flyMode = config.flyModeEnabled;
|
|
rig.turnSnap = parseFloat(config.snapTurnSnap);
|
|
const webController = new WebController(ground.getScene(), rig, diagramManager);
|
|
|
|
}
|
|
|
|
function positionComponentsRelativeToCamera(scene: Scene, diagramManager: DiagramManager) {
|
|
const logger = log.getLogger('positionComponentsRelativeToCamera');
|
|
const platform = scene.getMeshByName('platform');
|
|
if (!platform) {
|
|
logger.warn('Platform not found, cannot position components');
|
|
return;
|
|
}
|
|
|
|
const camera = scene.activeCamera;
|
|
if (!camera) {
|
|
logger.warn('Active camera not found, cannot position components');
|
|
return;
|
|
}
|
|
|
|
// Get camera world position
|
|
const cameraWorldPos = camera.globalPosition;
|
|
|
|
// Get camera's actual forward direction in world space
|
|
// This accounts for the camera's parent transform rotation (Math.PI on Y)
|
|
const cameraForward = camera.getDirection(Vector3.Forward());
|
|
const horizontalForward = cameraForward.clone();
|
|
horizontalForward.y = 0; // Keep only horizontal component
|
|
horizontalForward.normalize(); // Ensure unit vector
|
|
|
|
// Create a left direction (perpendicular to forward, in world space)
|
|
const horizontalLeft = Vector3.Cross(Vector3.Up(), horizontalForward).normalize();
|
|
|
|
logger.info('Camera world position:', cameraWorldPos);
|
|
logger.info('Camera forward (world):', cameraForward);
|
|
logger.info('Horizontal forward:', horizontalForward);
|
|
logger.info('Horizontal left:', horizontalLeft);
|
|
logger.info('Platform world position:', platform.getAbsolutePosition());
|
|
|
|
// Position toolbox: On left side following VR best practices
|
|
// Meta guidelines: 45-60 degrees to the side, comfortable arm's reach (~0.4-0.5m)
|
|
const toolbox = diagramManager.diagramMenuManager.toolbox;
|
|
if (toolbox && toolbox.handleMesh) {
|
|
logger.info('Toolbox handleMesh BEFORE positioning:', {
|
|
position: toolbox.handleMesh.position.clone(),
|
|
absolutePosition: toolbox.handleMesh.getAbsolutePosition().clone(),
|
|
rotation: toolbox.handleMesh.rotation.clone()
|
|
});
|
|
|
|
// Position at 45 degrees to the left, 0.45m away, slightly below eye level
|
|
// NOTE: User faces -Z direction by design, so negate forward offset
|
|
const forwardOffset = horizontalForward.scale(-0.3);
|
|
const leftOffset = horizontalLeft.scale(0.35);
|
|
const toolboxWorldPos = cameraWorldPos.add(forwardOffset).add(leftOffset);
|
|
toolboxWorldPos.y = cameraWorldPos.y - 0.3; // Below eye level
|
|
|
|
logger.info('Calculated toolbox world position:', toolboxWorldPos);
|
|
logger.info('Forward offset:', forwardOffset);
|
|
logger.info('Left offset:', leftOffset);
|
|
|
|
const toolboxLocalPos = Vector3.TransformCoordinates(toolboxWorldPos, platform.getWorldMatrix().invert());
|
|
logger.info('Calculated toolbox local position:', toolboxLocalPos);
|
|
|
|
toolbox.handleMesh.position = toolboxLocalPos;
|
|
|
|
// Orient toolbox to face the user
|
|
const toolboxToCamera = cameraWorldPos.subtract(toolboxWorldPos).normalize();
|
|
const toolboxYaw = Math.atan2(toolboxToCamera.x, toolboxToCamera.z);
|
|
toolbox.handleMesh.rotation.y = toolboxYaw;
|
|
|
|
logger.info('Toolbox handleMesh AFTER positioning:', {
|
|
position: toolbox.handleMesh.position.clone(),
|
|
absolutePosition: toolbox.handleMesh.getAbsolutePosition().clone(),
|
|
rotation: toolbox.handleMesh.rotation.clone()
|
|
});
|
|
}
|
|
|
|
// Position input text view: Centered in front, slightly below eye level
|
|
const inputTextView = diagramManager.diagramMenuManager['_inputTextView'];
|
|
if (inputTextView && inputTextView.handleMesh) {
|
|
logger.info('InputTextView handleMesh BEFORE positioning:', {
|
|
position: inputTextView.handleMesh.position.clone(),
|
|
absolutePosition: inputTextView.handleMesh.getAbsolutePosition().clone()
|
|
});
|
|
|
|
// NOTE: User faces -Z direction by design, so negate forward offset
|
|
const inputWorldPos = cameraWorldPos.add(horizontalForward.scale(-0.5));
|
|
inputWorldPos.y = cameraWorldPos.y - 0.4; // Below eye level
|
|
|
|
logger.info('Calculated input world position:', inputWorldPos);
|
|
|
|
const inputLocalPos = Vector3.TransformCoordinates(inputWorldPos, platform.getWorldMatrix().invert());
|
|
logger.info('Calculated input local position:', inputLocalPos);
|
|
|
|
inputTextView.handleMesh.position = inputLocalPos;
|
|
|
|
logger.info('InputTextView handleMesh AFTER positioning:', {
|
|
position: inputTextView.handleMesh.position.clone(),
|
|
absolutePosition: inputTextView.handleMesh.getAbsolutePosition().clone()
|
|
});
|
|
}
|
|
} |