mirror of
https://github.com/godotengine/godot.git
synced 2024-11-22 12:12:28 +00:00
daa4704c44
Introduced in 08ffa5d89e
.
3545 lines
117 KiB
C++
3545 lines
117 KiB
C++
/**************************************************************************/
|
|
/* openxr_api.cpp */
|
|
/**************************************************************************/
|
|
/* This file is part of: */
|
|
/* GODOT ENGINE */
|
|
/* https://godotengine.org */
|
|
/**************************************************************************/
|
|
/* Copyright (c) 2014-present Godot Engine contributors (see AUTHORS.md). */
|
|
/* Copyright (c) 2007-2014 Juan Linietsky, Ariel Manzur. */
|
|
/* */
|
|
/* Permission is hereby granted, free of charge, to any person obtaining */
|
|
/* a copy of this software and associated documentation files (the */
|
|
/* "Software"), to deal in the Software without restriction, including */
|
|
/* without limitation the rights to use, copy, modify, merge, publish, */
|
|
/* distribute, sublicense, and/or sell copies of the Software, and to */
|
|
/* permit persons to whom the Software is furnished to do so, subject to */
|
|
/* the following conditions: */
|
|
/* */
|
|
/* The above copyright notice and this permission notice shall be */
|
|
/* included in all copies or substantial portions of the Software. */
|
|
/* */
|
|
/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
|
|
/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
|
|
/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. */
|
|
/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
|
|
/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
|
|
/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
|
|
/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
|
|
/**************************************************************************/
|
|
|
|
#include "openxr_api.h"
|
|
|
|
#include "extensions/openxr_extension_wrapper_extension.h"
|
|
#include "openxr_interface.h"
|
|
#include "openxr_util.h"
|
|
|
|
#include "core/config/engine.h"
|
|
#include "core/config/project_settings.h"
|
|
#include "core/os/memory.h"
|
|
#include "core/version.h"
|
|
|
|
#ifdef TOOLS_ENABLED
|
|
#include "editor/editor_settings.h"
|
|
#endif
|
|
|
|
#include "openxr_platform_inc.h"
|
|
|
|
#ifdef VULKAN_ENABLED
|
|
#include "extensions/platform/openxr_vulkan_extension.h"
|
|
#endif
|
|
|
|
#if defined(GLES3_ENABLED) && !defined(MACOS_ENABLED)
|
|
#include "extensions/platform/openxr_opengl_extension.h"
|
|
#endif
|
|
|
|
#include "extensions/openxr_composition_layer_depth_extension.h"
|
|
#include "extensions/openxr_debug_utils_extension.h"
|
|
#include "extensions/openxr_eye_gaze_interaction.h"
|
|
#include "extensions/openxr_fb_display_refresh_rate_extension.h"
|
|
#include "extensions/openxr_fb_foveation_extension.h"
|
|
#include "extensions/openxr_fb_update_swapchain_extension.h"
|
|
#include "extensions/openxr_hand_tracking_extension.h"
|
|
|
|
#ifdef ANDROID_ENABLED
|
|
#define OPENXR_LOADER_NAME "libopenxr_loader.so"
|
|
#endif
|
|
|
|
////////////////////////////////////
|
|
// OpenXRAPI::OpenXRSwapChainInfo
|
|
|
|
Vector<OpenXRAPI::OpenXRSwapChainInfo> OpenXRAPI::OpenXRSwapChainInfo::free_queue;
|
|
|
|
bool OpenXRAPI::OpenXRSwapChainInfo::create(XrSwapchainCreateFlags p_create_flags, XrSwapchainUsageFlags p_usage_flags, int64_t p_swapchain_format, uint32_t p_width, uint32_t p_height, uint32_t p_sample_count, uint32_t p_array_size) {
|
|
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
|
|
ERR_FAIL_NULL_V(openxr_api, false);
|
|
|
|
XrSession xr_session = openxr_api->get_session();
|
|
ERR_FAIL_COND_V(xr_session == XR_NULL_HANDLE, false);
|
|
|
|
OpenXRGraphicsExtensionWrapper *xr_graphics_extension = openxr_api->get_graphics_extension();
|
|
ERR_FAIL_NULL_V(xr_graphics_extension, false);
|
|
|
|
// We already have a swapchain?
|
|
ERR_FAIL_COND_V(swapchain != XR_NULL_HANDLE, false);
|
|
|
|
XrResult result;
|
|
|
|
void *next_pointer = nullptr;
|
|
for (OpenXRExtensionWrapper *wrapper : openxr_api->get_registered_extension_wrappers()) {
|
|
void *np = wrapper->set_swapchain_create_info_and_get_next_pointer(next_pointer);
|
|
if (np != nullptr) {
|
|
next_pointer = np;
|
|
}
|
|
}
|
|
|
|
XrSwapchainCreateInfo swapchain_create_info = {
|
|
XR_TYPE_SWAPCHAIN_CREATE_INFO, // type
|
|
next_pointer, // next
|
|
p_create_flags, // createFlags
|
|
p_usage_flags, // usageFlags
|
|
p_swapchain_format, // format
|
|
p_sample_count, // sampleCount
|
|
p_width, // width
|
|
p_height, // height
|
|
1, // faceCount
|
|
p_array_size, // arraySize
|
|
1 // mipCount
|
|
};
|
|
|
|
XrSwapchain new_swapchain;
|
|
result = openxr_api->xrCreateSwapchain(xr_session, &swapchain_create_info, &new_swapchain);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to get swapchain [", openxr_api->get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
if (!xr_graphics_extension->get_swapchain_image_data(new_swapchain, p_swapchain_format, p_width, p_height, p_sample_count, p_array_size, &swapchain_graphics_data)) {
|
|
openxr_api->xrDestroySwapchain(new_swapchain);
|
|
return false;
|
|
}
|
|
|
|
swapchain = new_swapchain;
|
|
|
|
return true;
|
|
}
|
|
|
|
void OpenXRAPI::OpenXRSwapChainInfo::queue_free() {
|
|
if (image_acquired) {
|
|
release();
|
|
}
|
|
|
|
if (swapchain != XR_NULL_HANDLE) {
|
|
free_queue.push_back(*this);
|
|
|
|
swapchain_graphics_data = nullptr;
|
|
swapchain = XR_NULL_HANDLE;
|
|
}
|
|
}
|
|
|
|
void OpenXRAPI::OpenXRSwapChainInfo::free_queued() {
|
|
for (OpenXRAPI::OpenXRSwapChainInfo &swapchain_info : free_queue) {
|
|
swapchain_info.free();
|
|
}
|
|
free_queue.clear();
|
|
}
|
|
|
|
void OpenXRAPI::OpenXRSwapChainInfo::free() {
|
|
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
|
|
ERR_FAIL_NULL(openxr_api);
|
|
|
|
if (image_acquired) {
|
|
release();
|
|
}
|
|
|
|
if (openxr_api->get_graphics_extension() && swapchain_graphics_data != nullptr) {
|
|
openxr_api->get_graphics_extension()->cleanup_swapchain_graphics_data(&swapchain_graphics_data);
|
|
}
|
|
|
|
if (swapchain != XR_NULL_HANDLE) {
|
|
openxr_api->xrDestroySwapchain(swapchain);
|
|
swapchain = XR_NULL_HANDLE;
|
|
}
|
|
}
|
|
|
|
bool OpenXRAPI::OpenXRSwapChainInfo::acquire(bool &p_should_render) {
|
|
ERR_FAIL_COND_V(image_acquired, true); // This was not released when it should be, error out and reuse...
|
|
|
|
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
|
|
ERR_FAIL_NULL_V(openxr_api, false);
|
|
|
|
XrResult result;
|
|
|
|
if (!skip_acquire_swapchain) {
|
|
XrSwapchainImageAcquireInfo swapchain_image_acquire_info = {
|
|
XR_TYPE_SWAPCHAIN_IMAGE_ACQUIRE_INFO, // type
|
|
nullptr // next
|
|
};
|
|
|
|
result = openxr_api->xrAcquireSwapchainImage(swapchain, &swapchain_image_acquire_info, &image_index);
|
|
if (!XR_UNQUALIFIED_SUCCESS(result)) {
|
|
// Make sure end_frame knows we need to submit an empty frame
|
|
p_should_render = false;
|
|
|
|
if (XR_FAILED(result)) {
|
|
// Unexpected failure, log this!
|
|
print_line("OpenXR: failed to acquire swapchain image [", openxr_api->get_error_string(result), "]");
|
|
return false;
|
|
} else {
|
|
// In this scenario we silently fail, the XR runtime is simply not ready yet to acquire the swapchain.
|
|
return false;
|
|
}
|
|
}
|
|
}
|
|
|
|
XrSwapchainImageWaitInfo swapchain_image_wait_info = {
|
|
XR_TYPE_SWAPCHAIN_IMAGE_WAIT_INFO, // type
|
|
nullptr, // next
|
|
1000000000 // 1s timeout in nanoseconds
|
|
};
|
|
|
|
// Wait for a maximum of 10 seconds before calling it a critical failure...
|
|
for (int retry = 0; retry < 10; retry++) {
|
|
result = openxr_api->xrWaitSwapchainImage(swapchain, &swapchain_image_wait_info);
|
|
if (result != XR_TIMEOUT_EXPIRED) {
|
|
break;
|
|
}
|
|
WARN_PRINT("OpenXR: timed out waiting for swapchain image.");
|
|
}
|
|
|
|
if (!XR_UNQUALIFIED_SUCCESS(result)) {
|
|
// Make sure end_frame knows we need to submit an empty frame
|
|
p_should_render = false;
|
|
|
|
if (XR_FAILED(result)) {
|
|
// Unexpected failure, log this!
|
|
print_line("OpenXR: failed to wait for swapchain image [", openxr_api->get_error_string(result), "]");
|
|
return false;
|
|
} else {
|
|
WARN_PRINT("OpenXR: couldn't to wait for swapchain but not a complete error [" + openxr_api->get_error_string(result) + "]");
|
|
|
|
// Make sure to skip trying to acquire the swapchain image in the next frame
|
|
skip_acquire_swapchain = true;
|
|
return false;
|
|
}
|
|
} else {
|
|
skip_acquire_swapchain = false;
|
|
}
|
|
|
|
image_acquired = true;
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::OpenXRSwapChainInfo::release() {
|
|
if (!image_acquired) {
|
|
// Already released or never acquired.
|
|
return true;
|
|
}
|
|
|
|
image_acquired = false; // Regardless if we succeed or not, consider this released.
|
|
|
|
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
|
|
ERR_FAIL_NULL_V(openxr_api, false);
|
|
|
|
XrSwapchainImageReleaseInfo swapchain_image_release_info = {
|
|
XR_TYPE_SWAPCHAIN_IMAGE_RELEASE_INFO, // type
|
|
nullptr // next
|
|
};
|
|
XrResult result = openxr_api->xrReleaseSwapchainImage(swapchain, &swapchain_image_release_info);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to release swapchain image! [", openxr_api->get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
RID OpenXRAPI::OpenXRSwapChainInfo::get_image() {
|
|
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
|
|
|
|
if (image_acquired && openxr_api && openxr_api->get_graphics_extension()) {
|
|
return OpenXRAPI::get_singleton()->get_graphics_extension()->get_texture(swapchain_graphics_data, image_index);
|
|
} else {
|
|
return RID();
|
|
}
|
|
}
|
|
|
|
////////////////////////////////////
|
|
// OpenXRAPI
|
|
|
|
OpenXRAPI *OpenXRAPI::singleton = nullptr;
|
|
Vector<OpenXRExtensionWrapper *> OpenXRAPI::registered_extension_wrappers;
|
|
|
|
bool OpenXRAPI::openxr_is_enabled(bool p_check_run_in_editor) {
|
|
if (XRServer::get_xr_mode() == XRServer::XRMODE_DEFAULT) {
|
|
if (Engine::get_singleton()->is_editor_hint() && p_check_run_in_editor) {
|
|
// For now, don't start OpenXR when the editor starts up. In the future, this may change
|
|
// if we want to integrate more XR features into the editor experience.
|
|
return false;
|
|
} else {
|
|
return GLOBAL_GET("xr/openxr/enabled");
|
|
}
|
|
} else {
|
|
return XRServer::get_xr_mode() == XRServer::XRMODE_ON;
|
|
}
|
|
}
|
|
|
|
String OpenXRAPI::get_default_action_map_resource_name() {
|
|
String name = GLOBAL_GET("xr/openxr/default_action_map");
|
|
|
|
return name;
|
|
}
|
|
|
|
String OpenXRAPI::get_error_string(XrResult result) const {
|
|
if (XR_SUCCEEDED(result)) {
|
|
return String("Succeeded");
|
|
}
|
|
|
|
if (instance == XR_NULL_HANDLE) {
|
|
Array args;
|
|
args.push_back(Variant(result));
|
|
return String("Error code {0}").format(args);
|
|
}
|
|
|
|
char resultString[XR_MAX_RESULT_STRING_SIZE];
|
|
xrResultToString(instance, result, resultString);
|
|
|
|
return String(resultString);
|
|
}
|
|
|
|
String OpenXRAPI::get_swapchain_format_name(int64_t p_swapchain_format) const {
|
|
// This is rendering engine dependent...
|
|
if (graphics_extension) {
|
|
return graphics_extension->get_swapchain_format_name(p_swapchain_format);
|
|
}
|
|
|
|
return String("Swapchain format ") + String::num_int64(int64_t(p_swapchain_format));
|
|
}
|
|
|
|
void OpenXRAPI::set_object_name(XrObjectType p_object_type, uint64_t p_object_handle, const String &p_object_name) {
|
|
OpenXRDebugUtilsExtension *debug_utils = OpenXRDebugUtilsExtension::get_singleton();
|
|
if (!debug_utils || !debug_utils->get_active()) {
|
|
// Not enabled/active? Ignore.
|
|
return;
|
|
}
|
|
|
|
debug_utils->set_object_name(p_object_type, p_object_handle, p_object_name.utf8().get_data());
|
|
}
|
|
|
|
void OpenXRAPI::begin_debug_label_region(const String &p_label_name) {
|
|
OpenXRDebugUtilsExtension *debug_utils = OpenXRDebugUtilsExtension::get_singleton();
|
|
if (!debug_utils || !debug_utils->get_active()) {
|
|
// Not enabled/active? Ignore.
|
|
return;
|
|
}
|
|
|
|
debug_utils->begin_debug_label_region(p_label_name.utf8().get_data());
|
|
}
|
|
|
|
void OpenXRAPI::end_debug_label_region() {
|
|
OpenXRDebugUtilsExtension *debug_utils = OpenXRDebugUtilsExtension::get_singleton();
|
|
if (!debug_utils || !debug_utils->get_active()) {
|
|
// Not enabled/active? Ignore.
|
|
return;
|
|
}
|
|
|
|
debug_utils->end_debug_label_region();
|
|
}
|
|
|
|
void OpenXRAPI::insert_debug_label(const String &p_label_name) {
|
|
OpenXRDebugUtilsExtension *debug_utils = OpenXRDebugUtilsExtension::get_singleton();
|
|
if (!debug_utils || !debug_utils->get_active()) {
|
|
// Not enabled/active? Ignore.
|
|
return;
|
|
}
|
|
|
|
debug_utils->insert_debug_label(p_label_name.utf8().get_data());
|
|
}
|
|
|
|
bool OpenXRAPI::load_layer_properties() {
|
|
// This queries additional layers that are available and can be initialized when we create our OpenXR instance
|
|
if (layer_properties != nullptr) {
|
|
// already retrieved this
|
|
return true;
|
|
}
|
|
|
|
// Note, instance is not yet setup so we can't use get_error_string to retrieve our error
|
|
XrResult result = xrEnumerateApiLayerProperties(0, &num_layer_properties, nullptr);
|
|
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate number of api layer properties");
|
|
|
|
layer_properties = (XrApiLayerProperties *)memalloc(sizeof(XrApiLayerProperties) * num_layer_properties);
|
|
ERR_FAIL_NULL_V(layer_properties, false);
|
|
for (uint32_t i = 0; i < num_layer_properties; i++) {
|
|
layer_properties[i].type = XR_TYPE_API_LAYER_PROPERTIES;
|
|
layer_properties[i].next = nullptr;
|
|
}
|
|
|
|
result = xrEnumerateApiLayerProperties(num_layer_properties, &num_layer_properties, layer_properties);
|
|
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate api layer properties");
|
|
|
|
for (uint32_t i = 0; i < num_layer_properties; i++) {
|
|
print_verbose(String("OpenXR: Found OpenXR layer ") + layer_properties[i].layerName);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::load_supported_extensions() {
|
|
// This queries supported extensions that are available and can be initialized when we create our OpenXR instance
|
|
|
|
if (supported_extensions != nullptr) {
|
|
// already retrieved this
|
|
return true;
|
|
}
|
|
|
|
// Note, instance is not yet setup so we can't use get_error_string to retrieve our error
|
|
XrResult result = xrEnumerateInstanceExtensionProperties(nullptr, 0, &num_supported_extensions, nullptr);
|
|
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate number of extension properties");
|
|
|
|
supported_extensions = (XrExtensionProperties *)memalloc(sizeof(XrExtensionProperties) * num_supported_extensions);
|
|
ERR_FAIL_NULL_V(supported_extensions, false);
|
|
|
|
// set our types
|
|
for (uint32_t i = 0; i < num_supported_extensions; i++) {
|
|
supported_extensions[i].type = XR_TYPE_EXTENSION_PROPERTIES;
|
|
supported_extensions[i].next = nullptr;
|
|
}
|
|
result = xrEnumerateInstanceExtensionProperties(nullptr, num_supported_extensions, &num_supported_extensions, supported_extensions);
|
|
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate extension properties");
|
|
|
|
for (uint32_t i = 0; i < num_supported_extensions; i++) {
|
|
print_verbose(String("OpenXR: Found OpenXR extension ") + supported_extensions[i].extensionName);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::is_extension_supported(const String &p_extension) const {
|
|
for (uint32_t i = 0; i < num_supported_extensions; i++) {
|
|
if (supported_extensions[i].extensionName == p_extension) {
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool OpenXRAPI::is_extension_enabled(const String &p_extension) const {
|
|
CharString extension = p_extension.ascii();
|
|
|
|
for (int i = 0; i < enabled_extensions.size(); i++) {
|
|
if (strcmp(enabled_extensions[i].ptr(), extension.ptr()) == 0) {
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool OpenXRAPI::is_top_level_path_supported(const String &p_toplevel_path) {
|
|
String required_extension = OpenXRInteractionProfileMetadata::get_singleton()->get_top_level_extension(p_toplevel_path);
|
|
|
|
// If unsupported is returned we likely have a misspelled interaction profile path in our action map. Always output that as an error.
|
|
ERR_FAIL_COND_V_MSG(required_extension == XR_PATH_UNSUPPORTED_NAME, false, "OpenXR: Unsupported toplevel path " + p_toplevel_path);
|
|
|
|
if (required_extension == "") {
|
|
// no extension needed, core top level are always "supported", they just won't be used if not really supported
|
|
return true;
|
|
}
|
|
|
|
if (!is_extension_enabled(required_extension)) {
|
|
// It is very likely we have top level paths for which the extension is not available so don't flood the logs with unnecessary spam.
|
|
print_verbose("OpenXR: Top level path " + p_toplevel_path + " requires extension " + required_extension);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::is_interaction_profile_supported(const String &p_ip_path) {
|
|
String required_extension = OpenXRInteractionProfileMetadata::get_singleton()->get_interaction_profile_extension(p_ip_path);
|
|
|
|
// If unsupported is returned we likely have a misspelled interaction profile path in our action map. Always output that as an error.
|
|
ERR_FAIL_COND_V_MSG(required_extension == XR_PATH_UNSUPPORTED_NAME, false, "OpenXR: Unsupported interaction profile " + p_ip_path);
|
|
|
|
if (required_extension == "") {
|
|
// no extension needed, core interaction profiles are always "supported", they just won't be used if not really supported
|
|
return true;
|
|
}
|
|
|
|
if (!is_extension_enabled(required_extension)) {
|
|
// It is very likely we have interaction profiles for which the extension is not available so don't flood the logs with unnecessary spam.
|
|
print_verbose("OpenXR: Interaction profile " + p_ip_path + " requires extension " + required_extension);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::interaction_profile_supports_io_path(const String &p_ip_path, const String &p_io_path) {
|
|
if (!is_interaction_profile_supported(p_ip_path)) {
|
|
return false;
|
|
}
|
|
|
|
const OpenXRInteractionProfileMetadata::IOPath *io_path = OpenXRInteractionProfileMetadata::get_singleton()->get_io_path(p_ip_path, p_io_path);
|
|
|
|
// If the io_path is not part of our metadata we've likely got a misspelled name or a bad action map, report
|
|
ERR_FAIL_NULL_V_MSG(io_path, false, "OpenXR: Unsupported io path " + String(p_ip_path) + String(p_io_path));
|
|
|
|
if (io_path->openxr_extension_name == "") {
|
|
// no extension needed, core io paths are always "supported", they just won't be used if not really supported
|
|
return true;
|
|
}
|
|
|
|
if (!is_extension_enabled(io_path->openxr_extension_name)) {
|
|
// It is very likely we have io paths for which the extension is not available so don't flood the logs with unnecessary spam.
|
|
print_verbose("OpenXR: IO path " + String(p_ip_path) + String(p_io_path) + " requires extension " + io_path->openxr_extension_name);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void OpenXRAPI::copy_string_to_char_buffer(const String p_string, char *p_buffer, int p_buffer_len) {
|
|
CharString char_string = p_string.utf8();
|
|
int len = char_string.length();
|
|
if (len < p_buffer_len - 1) {
|
|
// was having weird CI issues with strcpy so....
|
|
memcpy(p_buffer, char_string.get_data(), len);
|
|
p_buffer[len] = '\0';
|
|
} else {
|
|
memcpy(p_buffer, char_string.get_data(), p_buffer_len - 1);
|
|
p_buffer[p_buffer_len - 1] = '\0';
|
|
}
|
|
}
|
|
|
|
bool OpenXRAPI::create_instance() {
|
|
// Create our OpenXR instance, this will query any registered extension wrappers for extensions we need to enable.
|
|
|
|
// Append the extensions requested by the registered extension wrappers.
|
|
HashMap<String, bool *> requested_extensions;
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
const HashMap<String, bool *> &wrapper_request_extensions = wrapper->get_requested_extensions();
|
|
|
|
for (const KeyValue<String, bool *> &requested_extension : wrapper_request_extensions) {
|
|
requested_extensions[requested_extension.key] = requested_extension.value;
|
|
}
|
|
}
|
|
|
|
// Check which extensions are supported.
|
|
enabled_extensions.clear();
|
|
|
|
for (KeyValue<String, bool *> &requested_extension : requested_extensions) {
|
|
if (!is_extension_supported(requested_extension.key)) {
|
|
if (requested_extension.value == nullptr) {
|
|
// Null means this is a mandatory extension so we fail.
|
|
ERR_FAIL_V_MSG(false, String("OpenXR: OpenXR Runtime does not support ") + requested_extension.key + String(" extension!"));
|
|
} else {
|
|
// Set this extension as not supported.
|
|
*requested_extension.value = false;
|
|
}
|
|
} else if (requested_extension.value != nullptr) {
|
|
// Set this extension as supported.
|
|
*requested_extension.value = true;
|
|
|
|
// And record that we want to enable it.
|
|
enabled_extensions.push_back(requested_extension.key.ascii());
|
|
} else {
|
|
// Record that we want to enable this.
|
|
enabled_extensions.push_back(requested_extension.key.ascii());
|
|
}
|
|
}
|
|
|
|
Vector<const char *> extension_ptrs;
|
|
for (int i = 0; i < enabled_extensions.size(); i++) {
|
|
print_verbose(String("OpenXR: Enabling extension ") + String(enabled_extensions[i]));
|
|
extension_ptrs.push_back(enabled_extensions[i].get_data());
|
|
}
|
|
|
|
// Create our OpenXR instance
|
|
XrApplicationInfo application_info{
|
|
"Godot Engine", // applicationName, if we're running a game we'll update this down below.
|
|
1, // applicationVersion, we don't currently have this
|
|
"Godot Engine", // engineName
|
|
VERSION_MAJOR * 10000 + VERSION_MINOR * 100 + VERSION_PATCH, // engineVersion 4.0 -> 40000, 4.0.1 -> 40001, 4.1 -> 40100, etc.
|
|
XR_API_VERSION_1_0 // apiVersion
|
|
};
|
|
|
|
void *next_pointer = nullptr;
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
void *np = wrapper->set_instance_create_info_and_get_next_pointer(next_pointer);
|
|
if (np != nullptr) {
|
|
next_pointer = np;
|
|
}
|
|
}
|
|
|
|
XrInstanceCreateInfo instance_create_info = {
|
|
XR_TYPE_INSTANCE_CREATE_INFO, // type
|
|
next_pointer, // next
|
|
0, // createFlags
|
|
application_info, // applicationInfo
|
|
0, // enabledApiLayerCount, need to find out if we need support for this?
|
|
nullptr, // enabledApiLayerNames
|
|
uint32_t(extension_ptrs.size()), // enabledExtensionCount
|
|
extension_ptrs.ptr() // enabledExtensionNames
|
|
};
|
|
|
|
// Get our project name
|
|
String project_name = GLOBAL_GET("application/config/name");
|
|
if (!project_name.is_empty()) {
|
|
copy_string_to_char_buffer(project_name, instance_create_info.applicationInfo.applicationName, XR_MAX_APPLICATION_NAME_SIZE);
|
|
}
|
|
|
|
XrResult result = xrCreateInstance(&instance_create_info, &instance);
|
|
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "Failed to create XR instance.");
|
|
|
|
// from this point on we can use get_error_string to get more info about our errors...
|
|
|
|
XrInstanceProperties instanceProps = {
|
|
XR_TYPE_INSTANCE_PROPERTIES, // type;
|
|
nullptr, // next
|
|
0, // runtimeVersion, from here will be set by our get call
|
|
"" // runtimeName
|
|
};
|
|
|
|
OPENXR_API_INIT_XR_FUNC_V(xrGetInstanceProperties);
|
|
|
|
result = xrGetInstanceProperties(instance, &instanceProps);
|
|
if (XR_FAILED(result)) {
|
|
// not fatal probably
|
|
print_line("OpenXR: Failed to get XR instance properties [", get_error_string(result), "]");
|
|
|
|
runtime_name = "";
|
|
runtime_version = "";
|
|
} else {
|
|
runtime_name = instanceProps.runtimeName;
|
|
runtime_version = OpenXRUtil::make_xr_version_string(instanceProps.runtimeVersion);
|
|
print_line("OpenXR: Running on OpenXR runtime: ", runtime_name, " ", runtime_version);
|
|
}
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_instance_created(instance);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::get_system_info() {
|
|
// Retrieve basic OpenXR system info based on the form factor we desire
|
|
|
|
// Retrieve the system for our form factor, fails if form factor is not available
|
|
XrSystemGetInfo system_get_info = {
|
|
XR_TYPE_SYSTEM_GET_INFO, // type;
|
|
nullptr, // next
|
|
form_factor // formFactor
|
|
};
|
|
|
|
XrResult result = xrGetSystem(instance, &system_get_info, &system_id);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to get system for our form factor [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
// obtain info about our system, writing this out completely to make CI on Linux happy..
|
|
void *next_pointer = nullptr;
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
void *np = wrapper->set_system_properties_and_get_next_pointer(next_pointer);
|
|
if (np != nullptr) {
|
|
next_pointer = np;
|
|
}
|
|
}
|
|
|
|
XrSystemProperties system_properties = {
|
|
XR_TYPE_SYSTEM_PROPERTIES, // type
|
|
next_pointer, // next
|
|
0, // systemId, from here will be set by our get call
|
|
0, // vendorId
|
|
"", // systemName
|
|
{
|
|
0, // maxSwapchainImageHeight
|
|
0, // maxSwapchainImageWidth
|
|
0, // maxLayerCount
|
|
}, // graphicsProperties
|
|
{
|
|
false, // orientationTracking
|
|
false // positionTracking
|
|
} // trackingProperties
|
|
};
|
|
|
|
result = xrGetSystemProperties(instance, system_id, &system_properties);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to get System properties [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
// remember this state, we'll use it later
|
|
system_name = String(system_properties.systemName);
|
|
vendor_id = system_properties.vendorId;
|
|
graphics_properties = system_properties.graphicsProperties;
|
|
tracking_properties = system_properties.trackingProperties;
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::load_supported_view_configuration_types() {
|
|
// This queries the supported configuration types, likely there will only be one choosing between Mono (phone AR) and Stereo (HMDs)
|
|
|
|
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
|
|
|
|
if (supported_view_configuration_types != nullptr) {
|
|
// free previous results
|
|
memfree(supported_view_configuration_types);
|
|
supported_view_configuration_types = nullptr;
|
|
}
|
|
|
|
XrResult result = xrEnumerateViewConfigurations(instance, system_id, 0, &num_view_configuration_types, nullptr);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to get view configuration count [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
supported_view_configuration_types = (XrViewConfigurationType *)memalloc(sizeof(XrViewConfigurationType) * num_view_configuration_types);
|
|
ERR_FAIL_NULL_V(supported_view_configuration_types, false);
|
|
|
|
result = xrEnumerateViewConfigurations(instance, system_id, num_view_configuration_types, &num_view_configuration_types, supported_view_configuration_types);
|
|
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerateview configurations");
|
|
ERR_FAIL_COND_V_MSG(num_view_configuration_types == 0, false, "OpenXR: Failed to enumerateview configurations"); // JIC there should be at least 1!
|
|
|
|
for (uint32_t i = 0; i < num_view_configuration_types; i++) {
|
|
print_verbose(String("OpenXR: Found supported view configuration ") + OpenXRUtil::get_view_configuration_name(supported_view_configuration_types[i]));
|
|
}
|
|
|
|
// Check value we loaded at startup...
|
|
if (!is_view_configuration_supported(view_configuration)) {
|
|
print_verbose(String("OpenXR: ") + OpenXRUtil::get_view_configuration_name(view_configuration) + String(" isn't supported, defaulting to ") + OpenXRUtil::get_view_configuration_name(supported_view_configuration_types[0]));
|
|
|
|
view_configuration = supported_view_configuration_types[0];
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::load_supported_environmental_blend_modes() {
|
|
// This queries the supported environmental blend modes.
|
|
|
|
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
|
|
|
|
if (supported_environment_blend_modes != nullptr) {
|
|
// free previous results
|
|
memfree(supported_environment_blend_modes);
|
|
supported_environment_blend_modes = nullptr;
|
|
num_supported_environment_blend_modes = 0;
|
|
}
|
|
|
|
XrResult result = xrEnumerateEnvironmentBlendModes(instance, system_id, view_configuration, 0, &num_supported_environment_blend_modes, nullptr);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to get supported environmental blend mode count [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
supported_environment_blend_modes = (XrEnvironmentBlendMode *)memalloc(sizeof(XrEnvironmentBlendMode) * num_supported_environment_blend_modes);
|
|
ERR_FAIL_NULL_V(supported_environment_blend_modes, false);
|
|
|
|
result = xrEnumerateEnvironmentBlendModes(instance, system_id, view_configuration, num_supported_environment_blend_modes, &num_supported_environment_blend_modes, supported_environment_blend_modes);
|
|
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate environmental blend modes");
|
|
ERR_FAIL_COND_V_MSG(num_supported_environment_blend_modes == 0, false, "OpenXR: Failed to enumerate environmental blend modes"); // JIC there should be at least 1!
|
|
|
|
for (uint32_t i = 0; i < num_supported_environment_blend_modes; i++) {
|
|
print_verbose(String("OpenXR: Found environmental blend mode ") + OpenXRUtil::get_environment_blend_mode_name(supported_environment_blend_modes[i]));
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::is_view_configuration_supported(XrViewConfigurationType p_configuration_type) const {
|
|
ERR_FAIL_NULL_V(supported_view_configuration_types, false);
|
|
|
|
for (uint32_t i = 0; i < num_view_configuration_types; i++) {
|
|
if (supported_view_configuration_types[i] == p_configuration_type) {
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool OpenXRAPI::load_supported_view_configuration_views(XrViewConfigurationType p_configuration_type) {
|
|
// This loads our view configuration for each view so for a stereo HMD, we'll get two entries (that are likely identical)
|
|
// The returned data supplies us with the recommended render target size
|
|
|
|
if (!is_view_configuration_supported(p_configuration_type)) {
|
|
print_line("OpenXR: View configuration ", OpenXRUtil::get_view_configuration_name(view_configuration), " is not supported.");
|
|
return false;
|
|
}
|
|
|
|
if (view_configuration_views != nullptr) {
|
|
// free previous results
|
|
memfree(view_configuration_views);
|
|
view_configuration_views = nullptr;
|
|
}
|
|
|
|
XrResult result = xrEnumerateViewConfigurationViews(instance, system_id, p_configuration_type, 0, &view_count, nullptr);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to get view configuration count [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
view_configuration_views = (XrViewConfigurationView *)memalloc(sizeof(XrViewConfigurationView) * view_count);
|
|
ERR_FAIL_NULL_V(view_configuration_views, false);
|
|
|
|
for (uint32_t i = 0; i < view_count; i++) {
|
|
view_configuration_views[i].type = XR_TYPE_VIEW_CONFIGURATION_VIEW;
|
|
view_configuration_views[i].next = nullptr;
|
|
}
|
|
|
|
result = xrEnumerateViewConfigurationViews(instance, system_id, p_configuration_type, view_count, &view_count, view_configuration_views);
|
|
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate view configurations");
|
|
|
|
for (uint32_t i = 0; i < view_count; i++) {
|
|
print_verbose("OpenXR: Found supported view configuration view");
|
|
print_verbose(String(" - width: ") + itos(view_configuration_views[i].maxImageRectWidth));
|
|
print_verbose(String(" - height: ") + itos(view_configuration_views[i].maxImageRectHeight));
|
|
print_verbose(String(" - sample count: ") + itos(view_configuration_views[i].maxSwapchainSampleCount));
|
|
print_verbose(String(" - recommended render width: ") + itos(view_configuration_views[i].recommendedImageRectWidth));
|
|
print_verbose(String(" - recommended render height: ") + itos(view_configuration_views[i].recommendedImageRectHeight));
|
|
print_verbose(String(" - recommended render sample count: ") + itos(view_configuration_views[i].recommendedSwapchainSampleCount));
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void OpenXRAPI::destroy_instance() {
|
|
if (view_configuration_views != nullptr) {
|
|
memfree(view_configuration_views);
|
|
view_configuration_views = nullptr;
|
|
}
|
|
|
|
if (supported_view_configuration_types != nullptr) {
|
|
memfree(supported_view_configuration_types);
|
|
supported_view_configuration_types = nullptr;
|
|
}
|
|
|
|
if (supported_environment_blend_modes != nullptr) {
|
|
memfree(supported_environment_blend_modes);
|
|
supported_environment_blend_modes = nullptr;
|
|
num_supported_environment_blend_modes = 0;
|
|
}
|
|
|
|
if (instance != XR_NULL_HANDLE) {
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_instance_destroyed();
|
|
}
|
|
|
|
xrDestroyInstance(instance);
|
|
instance = XR_NULL_HANDLE;
|
|
}
|
|
enabled_extensions.clear();
|
|
|
|
if (graphics_extension != nullptr) {
|
|
unregister_extension_wrapper(graphics_extension);
|
|
memdelete(graphics_extension);
|
|
graphics_extension = nullptr;
|
|
}
|
|
}
|
|
|
|
bool OpenXRAPI::create_session() {
|
|
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
|
|
ERR_FAIL_COND_V(session != XR_NULL_HANDLE, false);
|
|
|
|
void *next_pointer = nullptr;
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
void *np = wrapper->set_session_create_and_get_next_pointer(next_pointer);
|
|
if (np != nullptr) {
|
|
next_pointer = np;
|
|
}
|
|
}
|
|
|
|
XrSessionCreateInfo session_create_info = {
|
|
XR_TYPE_SESSION_CREATE_INFO, // type
|
|
next_pointer, // next
|
|
0, // createFlags
|
|
system_id // systemId
|
|
};
|
|
|
|
XrResult result = xrCreateSession(instance, &session_create_info, &session);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to create session [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
set_object_name(XR_OBJECT_TYPE_SESSION, uint64_t(session), "Main Godot OpenXR Session");
|
|
|
|
begin_debug_label_region("Godot session active");
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_session_created(session);
|
|
}
|
|
|
|
// Check our environment blend mode. This needs to happen after we call `on_session_created()`
|
|
// on the extension wrappers, so they can emulate alpha blend mode.
|
|
if (!set_environment_blend_mode(environment_blend_mode)) {
|
|
print_verbose(String("OpenXR: ") + OpenXRUtil::get_environment_blend_mode_name(environment_blend_mode) + String(" isn't supported, defaulting to ") + OpenXRUtil::get_environment_blend_mode_name(supported_environment_blend_modes[0]));
|
|
set_environment_blend_mode(supported_environment_blend_modes[0]);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::load_supported_reference_spaces() {
|
|
// loads the supported reference spaces for our OpenXR session
|
|
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
|
|
|
|
if (supported_reference_spaces != nullptr) {
|
|
// free previous results
|
|
memfree(supported_reference_spaces);
|
|
supported_reference_spaces = nullptr;
|
|
}
|
|
|
|
XrResult result = xrEnumerateReferenceSpaces(session, 0, &num_reference_spaces, nullptr);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to get reference space count [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
supported_reference_spaces = (XrReferenceSpaceType *)memalloc(sizeof(XrReferenceSpaceType) * num_reference_spaces);
|
|
ERR_FAIL_NULL_V(supported_reference_spaces, false);
|
|
|
|
result = xrEnumerateReferenceSpaces(session, num_reference_spaces, &num_reference_spaces, supported_reference_spaces);
|
|
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate reference spaces");
|
|
ERR_FAIL_COND_V_MSG(num_reference_spaces == 0, false, "OpenXR: Failed to enumerate reference spaces");
|
|
|
|
for (uint32_t i = 0; i < num_reference_spaces; i++) {
|
|
print_verbose(String("OpenXR: Found supported reference space ") + OpenXRUtil::get_reference_space_name(supported_reference_spaces[i]));
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::is_reference_space_supported(XrReferenceSpaceType p_reference_space) {
|
|
ERR_FAIL_NULL_V(supported_reference_spaces, false);
|
|
|
|
for (uint32_t i = 0; i < num_reference_spaces; i++) {
|
|
if (supported_reference_spaces[i] == p_reference_space) {
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool OpenXRAPI::setup_play_space() {
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
|
|
|
|
XrPosef identityPose = {
|
|
{ 0.0, 0.0, 0.0, 1.0 },
|
|
{ 0.0, 0.0, 0.0 }
|
|
};
|
|
|
|
XrReferenceSpaceType new_reference_space;
|
|
XrSpace new_play_space = XR_NULL_HANDLE;
|
|
bool will_emulate_local_floor = false;
|
|
|
|
if (is_reference_space_supported(requested_reference_space)) {
|
|
new_reference_space = requested_reference_space;
|
|
} else if (requested_reference_space == XR_REFERENCE_SPACE_TYPE_LOCAL_FLOOR_EXT && is_reference_space_supported(XR_REFERENCE_SPACE_TYPE_STAGE)) {
|
|
print_verbose("OpenXR: LOCAL_FLOOR space isn't supported, emulating using STAGE and LOCAL spaces.");
|
|
|
|
new_reference_space = XR_REFERENCE_SPACE_TYPE_LOCAL;
|
|
will_emulate_local_floor = true;
|
|
|
|
if (local_floor_emulation.local_space == XR_NULL_HANDLE) {
|
|
XrReferenceSpaceCreateInfo create_info = {
|
|
XR_TYPE_REFERENCE_SPACE_CREATE_INFO, // type
|
|
nullptr, // next
|
|
XR_REFERENCE_SPACE_TYPE_LOCAL, // referenceSpaceType
|
|
identityPose, // poseInReferenceSpace
|
|
};
|
|
|
|
XrResult result = xrCreateReferenceSpace(session, &create_info, &local_floor_emulation.local_space);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to create LOCAL space in order to emulate LOCAL_FLOOR [", get_error_string(result), "]");
|
|
will_emulate_local_floor = false;
|
|
}
|
|
|
|
set_object_name(XR_OBJECT_TYPE_SPACE, uint64_t(local_floor_emulation.local_space), "Emulation local space");
|
|
}
|
|
|
|
if (local_floor_emulation.stage_space == XR_NULL_HANDLE) {
|
|
XrReferenceSpaceCreateInfo create_info = {
|
|
XR_TYPE_REFERENCE_SPACE_CREATE_INFO, // type
|
|
nullptr, // next
|
|
XR_REFERENCE_SPACE_TYPE_STAGE, // referenceSpaceType
|
|
identityPose, // poseInReferenceSpace
|
|
};
|
|
|
|
XrResult result = xrCreateReferenceSpace(session, &create_info, &local_floor_emulation.stage_space);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to create STAGE space in order to emulate LOCAL_FLOOR [", get_error_string(result), "]");
|
|
will_emulate_local_floor = false;
|
|
}
|
|
|
|
set_object_name(XR_OBJECT_TYPE_SPACE, uint64_t(local_floor_emulation.stage_space), "Emulation stage space");
|
|
}
|
|
|
|
if (!will_emulate_local_floor) {
|
|
if (local_floor_emulation.local_space != XR_NULL_HANDLE) {
|
|
xrDestroySpace(local_floor_emulation.local_space);
|
|
local_floor_emulation.local_space = XR_NULL_HANDLE;
|
|
}
|
|
if (local_floor_emulation.stage_space != XR_NULL_HANDLE) {
|
|
xrDestroySpace(local_floor_emulation.stage_space);
|
|
local_floor_emulation.stage_space = XR_NULL_HANDLE;
|
|
}
|
|
}
|
|
} else {
|
|
// Fallback on LOCAL, which all OpenXR runtimes are required to support.
|
|
print_verbose(String("OpenXR: ") + OpenXRUtil::get_reference_space_name(requested_reference_space) + String(" isn't supported, defaulting to LOCAL space."));
|
|
new_reference_space = XR_REFERENCE_SPACE_TYPE_LOCAL;
|
|
}
|
|
|
|
XrReferenceSpaceCreateInfo play_space_create_info = {
|
|
XR_TYPE_REFERENCE_SPACE_CREATE_INFO, // type
|
|
nullptr, // next
|
|
new_reference_space, // referenceSpaceType
|
|
identityPose, // poseInReferenceSpace
|
|
};
|
|
|
|
XrResult result = xrCreateReferenceSpace(session, &play_space_create_info, &new_play_space);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to create play space [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
// If we've previously created a play space, clean it up first.
|
|
if (play_space != XR_NULL_HANDLE) {
|
|
// TODO Investigate if destroying our play space here is safe,
|
|
// it may still be used in the rendering thread.
|
|
|
|
xrDestroySpace(play_space);
|
|
}
|
|
play_space = new_play_space;
|
|
reference_space = new_reference_space;
|
|
|
|
set_object_name(XR_OBJECT_TYPE_SPACE, uint64_t(play_space), "Play space");
|
|
|
|
local_floor_emulation.enabled = will_emulate_local_floor;
|
|
local_floor_emulation.should_reset_floor_height = will_emulate_local_floor;
|
|
|
|
// Update render state so this play space is used rendering the upcoming frame.
|
|
set_render_play_space(play_space);
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::setup_view_space() {
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
|
|
|
|
if (!is_reference_space_supported(XR_REFERENCE_SPACE_TYPE_VIEW)) {
|
|
print_line("OpenXR: reference space XR_REFERENCE_SPACE_TYPE_VIEW is not supported.");
|
|
return false;
|
|
}
|
|
|
|
XrPosef identityPose = {
|
|
{ 0.0, 0.0, 0.0, 1.0 },
|
|
{ 0.0, 0.0, 0.0 }
|
|
};
|
|
|
|
XrReferenceSpaceCreateInfo view_space_create_info = {
|
|
XR_TYPE_REFERENCE_SPACE_CREATE_INFO, // type
|
|
nullptr, // next
|
|
XR_REFERENCE_SPACE_TYPE_VIEW, // referenceSpaceType
|
|
identityPose // poseInReferenceSpace
|
|
};
|
|
|
|
XrResult result = xrCreateReferenceSpace(session, &view_space_create_info, &view_space);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to create view space [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
set_object_name(XR_OBJECT_TYPE_SPACE, uint64_t(view_space), "View space");
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::reset_emulated_floor_height() {
|
|
ERR_FAIL_COND_V(!local_floor_emulation.enabled, false);
|
|
ERR_FAIL_COND_V(local_floor_emulation.local_space == XR_NULL_HANDLE, false);
|
|
ERR_FAIL_COND_V(local_floor_emulation.stage_space == XR_NULL_HANDLE, false);
|
|
|
|
XrResult result;
|
|
|
|
XrSpaceLocation stage_location = {
|
|
XR_TYPE_SPACE_LOCATION, // type
|
|
nullptr, // next
|
|
0, // locationFlags
|
|
{ { 0.0, 0.0, 0.0, 1.0 }, { 0.0, 0.0, 0.0 } }, // pose
|
|
};
|
|
|
|
result = xrLocateSpace(local_floor_emulation.stage_space, local_floor_emulation.local_space, get_predicted_display_time(), &stage_location);
|
|
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to locate STAGE space in LOCAL space, in order to emulate LOCAL_FLOOR [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
XrPosef pose = {
|
|
{ 0.0, 0.0, 0.0, 1.0 },
|
|
{ 0.0, stage_location.pose.position.y, 0.0 }
|
|
};
|
|
|
|
XrReferenceSpaceCreateInfo create_info = {
|
|
XR_TYPE_REFERENCE_SPACE_CREATE_INFO, // type
|
|
nullptr, // next
|
|
XR_REFERENCE_SPACE_TYPE_LOCAL, // referenceSpaceType
|
|
pose, // poseInReferenceSpace
|
|
};
|
|
|
|
XrSpace new_play_space;
|
|
result = xrCreateReferenceSpace(session, &create_info, &new_play_space);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to recreate emulated LOCAL_FLOOR play space with latest floor estimate [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
xrDestroySpace(play_space);
|
|
play_space = new_play_space;
|
|
|
|
// If we've made it this far, it means we can properly emulate LOCAL_FLOOR, so we'll
|
|
// report that as the reference space to the outside world.
|
|
reference_space = XR_REFERENCE_SPACE_TYPE_LOCAL_FLOOR_EXT;
|
|
|
|
// Update render state so this play space is used rendering the upcoming frame.
|
|
set_render_play_space(play_space);
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::load_supported_swapchain_formats() {
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
|
|
|
|
if (supported_swapchain_formats != nullptr) {
|
|
// free previous results
|
|
memfree(supported_swapchain_formats);
|
|
supported_swapchain_formats = nullptr;
|
|
}
|
|
|
|
XrResult result = xrEnumerateSwapchainFormats(session, 0, &num_swapchain_formats, nullptr);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to get swapchain format count [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
supported_swapchain_formats = (int64_t *)memalloc(sizeof(int64_t) * num_swapchain_formats);
|
|
ERR_FAIL_NULL_V(supported_swapchain_formats, false);
|
|
|
|
result = xrEnumerateSwapchainFormats(session, num_swapchain_formats, &num_swapchain_formats, supported_swapchain_formats);
|
|
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate swapchain formats");
|
|
|
|
for (uint32_t i = 0; i < num_swapchain_formats; i++) {
|
|
print_verbose(String("OpenXR: Found supported swapchain format ") + get_swapchain_format_name(supported_swapchain_formats[i]));
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::is_swapchain_format_supported(int64_t p_swapchain_format) {
|
|
ERR_FAIL_NULL_V(supported_swapchain_formats, false);
|
|
|
|
for (uint32_t i = 0; i < num_swapchain_formats; i++) {
|
|
if (supported_swapchain_formats[i] == p_swapchain_format) {
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool OpenXRAPI::obtain_swapchain_formats() {
|
|
ERR_FAIL_NULL_V(graphics_extension, false);
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
|
|
|
|
{
|
|
// Build a vector with swapchain formats we want to use, from best fit to worst
|
|
Vector<int64_t> usable_swapchain_formats;
|
|
color_swapchain_format = 0;
|
|
|
|
graphics_extension->get_usable_swapchain_formats(usable_swapchain_formats);
|
|
|
|
// now find out which one is supported
|
|
for (int i = 0; i < usable_swapchain_formats.size() && color_swapchain_format == 0; i++) {
|
|
if (is_swapchain_format_supported(usable_swapchain_formats[i])) {
|
|
color_swapchain_format = usable_swapchain_formats[i];
|
|
}
|
|
}
|
|
|
|
if (color_swapchain_format == 0) {
|
|
color_swapchain_format = usable_swapchain_formats[0]; // just use the first one and hope for the best...
|
|
print_line("Couldn't find usable color swap chain format, using", get_swapchain_format_name(color_swapchain_format), "instead.");
|
|
} else {
|
|
print_verbose(String("Using color swap chain format:") + get_swapchain_format_name(color_swapchain_format));
|
|
}
|
|
}
|
|
|
|
{
|
|
// Build a vector with swapchain formats we want to use, from best fit to worst
|
|
Vector<int64_t> usable_swapchain_formats;
|
|
depth_swapchain_format = 0;
|
|
|
|
graphics_extension->get_usable_depth_formats(usable_swapchain_formats);
|
|
|
|
// now find out which one is supported
|
|
for (int i = 0; i < usable_swapchain_formats.size() && depth_swapchain_format == 0; i++) {
|
|
if (is_swapchain_format_supported(usable_swapchain_formats[i])) {
|
|
depth_swapchain_format = usable_swapchain_formats[i];
|
|
}
|
|
}
|
|
|
|
if (depth_swapchain_format == 0) {
|
|
WARN_PRINT_ONCE("Couldn't find usable depth swap chain format, depth buffer will not be submitted if requested.");
|
|
} else {
|
|
print_verbose(String("Using depth swap chain format:") + get_swapchain_format_name(depth_swapchain_format));
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::create_main_swapchains(Size2i p_size) {
|
|
ERR_NOT_ON_RENDER_THREAD_V(false);
|
|
ERR_FAIL_NULL_V(graphics_extension, false);
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
|
|
|
|
/*
|
|
TODO: We need to improve on this, for now we're taking our old approach of creating our main swapchains and substituting
|
|
those for the ones Godot normally creates.
|
|
This however means we can only use swapchains for our main XR view.
|
|
|
|
It would have been nicer if we could override the swapchain creation in Godot with ours but we have a timing issue here.
|
|
We can't create XR swapchains until after our XR session is fully instantiated, yet Godot creates its swapchain much earlier.
|
|
|
|
We only creates a swapchain for the main output here.
|
|
Additional swapchains may be created through our composition layer extension.
|
|
|
|
Finally an area we need to expand upon is that Foveated rendering is only enabled for the swap chain we create,
|
|
as we render 3D content into internal buffers that are copied into the swapchain, we do now have (basic) VRS support
|
|
*/
|
|
|
|
render_state.main_swapchain_size = p_size;
|
|
uint32_t sample_count = 1;
|
|
|
|
// We start with our color swapchain...
|
|
if (color_swapchain_format != 0) {
|
|
if (!render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].create(0, XR_SWAPCHAIN_USAGE_SAMPLED_BIT | XR_SWAPCHAIN_USAGE_COLOR_ATTACHMENT_BIT | XR_SWAPCHAIN_USAGE_MUTABLE_FORMAT_BIT, color_swapchain_format, render_state.main_swapchain_size.width, render_state.main_swapchain_size.height, sample_count, view_count)) {
|
|
return false;
|
|
}
|
|
|
|
set_object_name(XR_OBJECT_TYPE_SWAPCHAIN, uint64_t(render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].get_swapchain()), "Main color swapchain");
|
|
}
|
|
|
|
// We create our depth swapchain if:
|
|
// - we've enabled submitting depth buffer
|
|
// - we support our depth layer extension
|
|
// - we have our spacewarp extension (not yet implemented)
|
|
if (depth_swapchain_format != 0 && submit_depth_buffer && OpenXRCompositionLayerDepthExtension::get_singleton()->is_available()) {
|
|
if (!render_state.main_swapchains[OPENXR_SWAPCHAIN_DEPTH].create(0, XR_SWAPCHAIN_USAGE_SAMPLED_BIT | XR_SWAPCHAIN_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT, depth_swapchain_format, render_state.main_swapchain_size.width, render_state.main_swapchain_size.height, sample_count, view_count)) {
|
|
return false;
|
|
}
|
|
|
|
set_object_name(XR_OBJECT_TYPE_SWAPCHAIN, uint64_t(render_state.main_swapchains[OPENXR_SWAPCHAIN_DEPTH].get_swapchain()), "Main depth swapchain");
|
|
}
|
|
|
|
// We create our velocity swapchain if:
|
|
// - we have our spacewarp extension (not yet implemented)
|
|
{
|
|
// TBD
|
|
}
|
|
|
|
for (uint32_t i = 0; i < render_state.view_count; i++) {
|
|
render_state.views[i].type = XR_TYPE_VIEW;
|
|
render_state.views[i].next = nullptr;
|
|
|
|
render_state.projection_views[i].type = XR_TYPE_COMPOSITION_LAYER_PROJECTION_VIEW;
|
|
render_state.projection_views[i].next = nullptr;
|
|
render_state.projection_views[i].subImage.swapchain = render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].get_swapchain();
|
|
render_state.projection_views[i].subImage.imageArrayIndex = i;
|
|
render_state.projection_views[i].subImage.imageRect.offset.x = 0;
|
|
render_state.projection_views[i].subImage.imageRect.offset.y = 0;
|
|
render_state.projection_views[i].subImage.imageRect.extent.width = render_state.main_swapchain_size.width;
|
|
render_state.projection_views[i].subImage.imageRect.extent.height = render_state.main_swapchain_size.height;
|
|
|
|
if (render_state.submit_depth_buffer && OpenXRCompositionLayerDepthExtension::get_singleton()->is_available() && render_state.depth_views) {
|
|
render_state.projection_views[i].next = &render_state.depth_views[i];
|
|
|
|
render_state.depth_views[i].type = XR_TYPE_COMPOSITION_LAYER_DEPTH_INFO_KHR;
|
|
render_state.depth_views[i].next = nullptr;
|
|
render_state.depth_views[i].subImage.swapchain = render_state.main_swapchains[OPENXR_SWAPCHAIN_DEPTH].get_swapchain();
|
|
render_state.depth_views[i].subImage.imageArrayIndex = i;
|
|
render_state.depth_views[i].subImage.imageRect.offset.x = 0;
|
|
render_state.depth_views[i].subImage.imageRect.offset.y = 0;
|
|
render_state.depth_views[i].subImage.imageRect.extent.width = render_state.main_swapchain_size.width;
|
|
render_state.depth_views[i].subImage.imageRect.extent.height = render_state.main_swapchain_size.height;
|
|
// OpenXR spec says that: minDepth < maxDepth.
|
|
render_state.depth_views[i].minDepth = 0.0;
|
|
render_state.depth_views[i].maxDepth = 1.0;
|
|
// But we can reverse near and far for reverse-Z.
|
|
render_state.depth_views[i].nearZ = 100.0; // Near and far Z will be set to the correct values in fill_projection_matrix
|
|
render_state.depth_views[i].farZ = 0.01;
|
|
}
|
|
};
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_main_swapchains_created();
|
|
}
|
|
|
|
return true;
|
|
};
|
|
|
|
void OpenXRAPI::destroy_session() {
|
|
// TODO need to figure out if we're still rendering our current frame
|
|
// in a separate rendering thread and if so,
|
|
// if we need to wait for completion.
|
|
// We could be pulling the rug from underneath rendering...
|
|
|
|
if (running) {
|
|
if (session != XR_NULL_HANDLE) {
|
|
xrEndSession(session);
|
|
}
|
|
|
|
running = false;
|
|
render_state.running = false;
|
|
}
|
|
|
|
if (render_state.views != nullptr) {
|
|
memfree(render_state.views);
|
|
render_state.views = nullptr;
|
|
}
|
|
|
|
if (render_state.projection_views != nullptr) {
|
|
memfree(render_state.projection_views);
|
|
render_state.projection_views = nullptr;
|
|
}
|
|
|
|
if (render_state.depth_views != nullptr) {
|
|
memfree(render_state.depth_views);
|
|
render_state.depth_views = nullptr;
|
|
}
|
|
|
|
free_main_swapchains();
|
|
OpenXRSwapChainInfo::free_queued();
|
|
|
|
if (supported_swapchain_formats != nullptr) {
|
|
memfree(supported_swapchain_formats);
|
|
supported_swapchain_formats = nullptr;
|
|
}
|
|
|
|
// destroy our spaces
|
|
if (play_space != XR_NULL_HANDLE) {
|
|
xrDestroySpace(play_space);
|
|
play_space = XR_NULL_HANDLE;
|
|
render_state.play_space = XR_NULL_HANDLE;
|
|
}
|
|
if (view_space != XR_NULL_HANDLE) {
|
|
xrDestroySpace(view_space);
|
|
view_space = XR_NULL_HANDLE;
|
|
}
|
|
if (local_floor_emulation.local_space != XR_NULL_HANDLE) {
|
|
xrDestroySpace(local_floor_emulation.local_space);
|
|
local_floor_emulation.local_space = XR_NULL_HANDLE;
|
|
}
|
|
if (local_floor_emulation.stage_space != XR_NULL_HANDLE) {
|
|
xrDestroySpace(local_floor_emulation.stage_space);
|
|
local_floor_emulation.stage_space = XR_NULL_HANDLE;
|
|
}
|
|
local_floor_emulation.enabled = false;
|
|
local_floor_emulation.should_reset_floor_height = false;
|
|
|
|
if (supported_reference_spaces != nullptr) {
|
|
// free previous results
|
|
memfree(supported_reference_spaces);
|
|
supported_reference_spaces = nullptr;
|
|
}
|
|
|
|
if (session != XR_NULL_HANDLE) {
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_session_destroyed();
|
|
}
|
|
|
|
end_debug_label_region();
|
|
|
|
xrDestroySession(session);
|
|
session = XR_NULL_HANDLE;
|
|
}
|
|
}
|
|
|
|
bool OpenXRAPI::on_state_idle() {
|
|
print_verbose("On state idle");
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_state_idle();
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::on_state_ready() {
|
|
print_verbose("On state ready");
|
|
|
|
// begin session
|
|
XrSessionBeginInfo session_begin_info = {
|
|
XR_TYPE_SESSION_BEGIN_INFO, // type
|
|
nullptr, // next
|
|
view_configuration // primaryViewConfigurationType
|
|
};
|
|
|
|
XrResult result = xrBeginSession(session, &session_begin_info);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to begin session [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
// we're running
|
|
running = true;
|
|
set_render_session_running(true);
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_state_ready();
|
|
}
|
|
|
|
if (xr_interface) {
|
|
xr_interface->on_state_ready();
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::on_state_synchronized() {
|
|
print_verbose("On state synchronized");
|
|
|
|
// Just in case, see if we already have active trackers...
|
|
List<RID> trackers;
|
|
tracker_owner.get_owned_list(&trackers);
|
|
for (const RID &tracker : trackers) {
|
|
tracker_check_profile(tracker);
|
|
}
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_state_synchronized();
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::on_state_visible() {
|
|
print_verbose("On state visible");
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_state_visible();
|
|
}
|
|
|
|
if (xr_interface) {
|
|
xr_interface->on_state_visible();
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::on_state_focused() {
|
|
print_verbose("On state focused");
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_state_focused();
|
|
}
|
|
|
|
if (xr_interface) {
|
|
xr_interface->on_state_focused();
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::on_state_stopping() {
|
|
print_verbose("On state stopping");
|
|
|
|
if (xr_interface) {
|
|
xr_interface->on_state_stopping();
|
|
}
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_state_stopping();
|
|
}
|
|
|
|
if (running) {
|
|
XrResult result = xrEndSession(session);
|
|
if (XR_FAILED(result)) {
|
|
// we only report this..
|
|
print_line("OpenXR: Failed to end session [", get_error_string(result), "]");
|
|
}
|
|
|
|
running = false;
|
|
set_render_session_running(false);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::on_state_loss_pending() {
|
|
print_verbose("On state loss pending");
|
|
|
|
if (xr_interface) {
|
|
xr_interface->on_state_loss_pending();
|
|
}
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_state_loss_pending();
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::on_state_exiting() {
|
|
print_verbose("On state existing");
|
|
|
|
if (xr_interface) {
|
|
xr_interface->on_state_exiting();
|
|
}
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_state_exiting();
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void OpenXRAPI::set_form_factor(XrFormFactor p_form_factor) {
|
|
ERR_FAIL_COND(is_initialized());
|
|
|
|
form_factor = p_form_factor;
|
|
}
|
|
|
|
uint32_t OpenXRAPI::get_view_count() {
|
|
return view_count;
|
|
}
|
|
|
|
void OpenXRAPI::set_view_configuration(XrViewConfigurationType p_view_configuration) {
|
|
ERR_FAIL_COND(is_initialized());
|
|
|
|
view_configuration = p_view_configuration;
|
|
}
|
|
|
|
bool OpenXRAPI::set_requested_reference_space(XrReferenceSpaceType p_requested_reference_space) {
|
|
requested_reference_space = p_requested_reference_space;
|
|
play_space_is_dirty = true;
|
|
|
|
return true;
|
|
}
|
|
|
|
void OpenXRAPI::set_submit_depth_buffer(bool p_submit_depth_buffer) {
|
|
ERR_FAIL_COND(is_initialized());
|
|
|
|
submit_depth_buffer = p_submit_depth_buffer;
|
|
}
|
|
|
|
bool OpenXRAPI::is_initialized() {
|
|
return (instance != XR_NULL_HANDLE);
|
|
}
|
|
|
|
bool OpenXRAPI::is_running() {
|
|
if (instance == XR_NULL_HANDLE) {
|
|
return false;
|
|
}
|
|
if (session == XR_NULL_HANDLE) {
|
|
return false;
|
|
}
|
|
|
|
return running;
|
|
}
|
|
|
|
bool OpenXRAPI::openxr_loader_init() {
|
|
#ifdef ANDROID_ENABLED
|
|
ERR_FAIL_COND_V_MSG(openxr_loader_library_handle != nullptr, false, "OpenXR Loader library is already loaded.");
|
|
|
|
{
|
|
Error error_code = OS::get_singleton()->open_dynamic_library(OPENXR_LOADER_NAME, openxr_loader_library_handle);
|
|
ERR_FAIL_COND_V_MSG(error_code != OK, false, "OpenXR loader not found.");
|
|
}
|
|
|
|
{
|
|
Error error_code = OS::get_singleton()->get_dynamic_library_symbol_handle(openxr_loader_library_handle, "xrGetInstanceProcAddr", (void *&)xrGetInstanceProcAddr);
|
|
ERR_FAIL_COND_V_MSG(error_code != OK, false, "Symbol xrGetInstanceProcAddr not found in OpenXR Loader library.");
|
|
}
|
|
#endif
|
|
|
|
// Resolve the symbols that don't require an instance
|
|
OPENXR_API_INIT_XR_FUNC_V(xrCreateInstance);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateApiLayerProperties);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateInstanceExtensionProperties);
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::resolve_instance_openxr_symbols() {
|
|
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
|
|
|
|
OPENXR_API_INIT_XR_FUNC_V(xrAcquireSwapchainImage);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrApplyHapticFeedback);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrAttachSessionActionSets);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrBeginFrame);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrBeginSession);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrCreateAction);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrCreateActionSet);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrCreateActionSpace);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrCreateReferenceSpace);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrCreateSession);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrCreateSwapchain);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrDestroyAction);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrDestroyActionSet);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrDestroyInstance);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrDestroySession);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrDestroySpace);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrDestroySwapchain);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrEndFrame);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrEndSession);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateEnvironmentBlendModes);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateReferenceSpaces);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateSwapchainFormats);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateViewConfigurations);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateViewConfigurationViews);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrGetActionStateBoolean);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrGetActionStateFloat);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrGetActionStateVector2f);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrGetCurrentInteractionProfile);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrGetReferenceSpaceBoundsRect);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrGetSystem);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrGetSystemProperties);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrLocateViews);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrLocateSpace);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrPathToString);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrPollEvent);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrReleaseSwapchainImage);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrResultToString);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrStringToPath);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrSuggestInteractionProfileBindings);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrSyncActions);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrWaitFrame);
|
|
OPENXR_API_INIT_XR_FUNC_V(xrWaitSwapchainImage);
|
|
|
|
return true;
|
|
}
|
|
|
|
XrResult OpenXRAPI::try_get_instance_proc_addr(const char *p_name, PFN_xrVoidFunction *p_addr) {
|
|
return xrGetInstanceProcAddr(instance, p_name, p_addr);
|
|
}
|
|
|
|
XrResult OpenXRAPI::get_instance_proc_addr(const char *p_name, PFN_xrVoidFunction *p_addr) {
|
|
XrResult result = try_get_instance_proc_addr(p_name, p_addr);
|
|
|
|
if (result != XR_SUCCESS) {
|
|
String error_message = String("Symbol ") + p_name + " not found in OpenXR instance.";
|
|
ERR_FAIL_V_MSG(result, error_message.utf8().get_data());
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
bool OpenXRAPI::initialize(const String &p_rendering_driver) {
|
|
ERR_FAIL_COND_V_MSG(instance != XR_NULL_HANDLE, false, "OpenXR instance was already created");
|
|
|
|
if (!openxr_loader_init()) {
|
|
return false;
|
|
}
|
|
|
|
if (p_rendering_driver == "vulkan") {
|
|
#ifdef VULKAN_ENABLED
|
|
graphics_extension = memnew(OpenXRVulkanExtension);
|
|
register_extension_wrapper(graphics_extension);
|
|
#else
|
|
// shouldn't be possible...
|
|
ERR_FAIL_V(false);
|
|
#endif
|
|
} else if (p_rendering_driver == "opengl3") {
|
|
#if defined(GLES3_ENABLED) && !defined(MACOS_ENABLED)
|
|
graphics_extension = memnew(OpenXROpenGLExtension);
|
|
register_extension_wrapper(graphics_extension);
|
|
#else
|
|
// shouldn't be possible...
|
|
ERR_FAIL_V(false);
|
|
#endif
|
|
} else {
|
|
ERR_FAIL_V_MSG(false, "OpenXR: Unsupported rendering device.");
|
|
}
|
|
|
|
// Also register our rendering extensions
|
|
register_extension_wrapper(memnew(OpenXRFBUpdateSwapchainExtension(p_rendering_driver)));
|
|
register_extension_wrapper(memnew(OpenXRFBFoveationExtension(p_rendering_driver)));
|
|
|
|
// initialize
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_before_instance_created();
|
|
}
|
|
|
|
if (!load_layer_properties()) {
|
|
destroy_instance();
|
|
return false;
|
|
}
|
|
|
|
if (!load_supported_extensions()) {
|
|
destroy_instance();
|
|
return false;
|
|
}
|
|
|
|
if (!create_instance()) {
|
|
destroy_instance();
|
|
return false;
|
|
}
|
|
|
|
if (!resolve_instance_openxr_symbols()) {
|
|
destroy_instance();
|
|
return false;
|
|
}
|
|
|
|
if (!get_system_info()) {
|
|
destroy_instance();
|
|
return false;
|
|
}
|
|
|
|
if (!load_supported_view_configuration_types()) {
|
|
destroy_instance();
|
|
return false;
|
|
}
|
|
|
|
if (!load_supported_view_configuration_views(view_configuration)) {
|
|
destroy_instance();
|
|
return false;
|
|
}
|
|
|
|
if (!load_supported_environmental_blend_modes()) {
|
|
destroy_instance();
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::initialize_session() {
|
|
if (!create_session()) {
|
|
destroy_session();
|
|
return false;
|
|
}
|
|
|
|
if (!load_supported_reference_spaces()) {
|
|
destroy_session();
|
|
return false;
|
|
}
|
|
|
|
if (!setup_view_space()) {
|
|
destroy_session();
|
|
return false;
|
|
}
|
|
|
|
if (!load_supported_swapchain_formats()) {
|
|
destroy_session();
|
|
return false;
|
|
}
|
|
|
|
if (!obtain_swapchain_formats()) {
|
|
destroy_session();
|
|
return false;
|
|
}
|
|
|
|
allocate_view_buffers(view_count, submit_depth_buffer);
|
|
|
|
return true;
|
|
}
|
|
|
|
void OpenXRAPI::finish() {
|
|
destroy_session();
|
|
|
|
destroy_instance();
|
|
}
|
|
|
|
void OpenXRAPI::set_xr_interface(OpenXRInterface *p_xr_interface) {
|
|
xr_interface = p_xr_interface;
|
|
}
|
|
|
|
void OpenXRAPI::register_extension_wrapper(OpenXRExtensionWrapper *p_extension_wrapper) {
|
|
registered_extension_wrappers.push_back(p_extension_wrapper);
|
|
}
|
|
|
|
void OpenXRAPI::unregister_extension_wrapper(OpenXRExtensionWrapper *p_extension_wrapper) {
|
|
registered_extension_wrappers.erase(p_extension_wrapper);
|
|
}
|
|
|
|
const Vector<OpenXRExtensionWrapper *> &OpenXRAPI::get_registered_extension_wrappers() {
|
|
return registered_extension_wrappers;
|
|
}
|
|
|
|
void OpenXRAPI::register_extension_metadata() {
|
|
for (OpenXRExtensionWrapper *extension_wrapper : registered_extension_wrappers) {
|
|
extension_wrapper->on_register_metadata();
|
|
}
|
|
}
|
|
|
|
void OpenXRAPI::cleanup_extension_wrappers() {
|
|
for (OpenXRExtensionWrapper *extension_wrapper : registered_extension_wrappers) {
|
|
// Fix crash when the extension wrapper comes from GDExtension.
|
|
OpenXRExtensionWrapperExtension *gdextension_extension_wrapper = dynamic_cast<OpenXRExtensionWrapperExtension *>(extension_wrapper);
|
|
if (gdextension_extension_wrapper) {
|
|
memdelete(gdextension_extension_wrapper);
|
|
} else {
|
|
memdelete(extension_wrapper);
|
|
}
|
|
}
|
|
registered_extension_wrappers.clear();
|
|
}
|
|
|
|
XrHandTrackerEXT OpenXRAPI::get_hand_tracker(int p_hand_index) {
|
|
ERR_FAIL_INDEX_V(p_hand_index, OpenXRHandTrackingExtension::HandTrackedHands::OPENXR_MAX_TRACKED_HANDS, XR_NULL_HANDLE);
|
|
|
|
OpenXRHandTrackingExtension *hand_tracking = OpenXRHandTrackingExtension::get_singleton();
|
|
ERR_FAIL_NULL_V(hand_tracking, XR_NULL_HANDLE);
|
|
|
|
OpenXRHandTrackingExtension::HandTrackedHands hand = static_cast<OpenXRHandTrackingExtension::HandTrackedHands>(p_hand_index);
|
|
return hand_tracking->get_hand_tracker(hand)->hand_tracker;
|
|
}
|
|
|
|
Size2 OpenXRAPI::get_recommended_target_size() {
|
|
RenderingServer *rendering_server = RenderingServer::get_singleton();
|
|
ERR_FAIL_NULL_V(view_configuration_views, Size2());
|
|
|
|
Size2 target_size;
|
|
|
|
if (rendering_server && rendering_server->is_on_render_thread()) {
|
|
target_size.width = view_configuration_views[0].recommendedImageRectWidth * render_state.render_target_size_multiplier;
|
|
target_size.height = view_configuration_views[0].recommendedImageRectHeight * render_state.render_target_size_multiplier;
|
|
} else {
|
|
target_size.width = view_configuration_views[0].recommendedImageRectWidth * render_target_size_multiplier;
|
|
target_size.height = view_configuration_views[0].recommendedImageRectHeight * render_target_size_multiplier;
|
|
}
|
|
|
|
return target_size;
|
|
}
|
|
|
|
XRPose::TrackingConfidence OpenXRAPI::get_head_center(Transform3D &r_transform, Vector3 &r_linear_velocity, Vector3 &r_angular_velocity) {
|
|
XrResult result;
|
|
|
|
if (!running) {
|
|
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
|
|
}
|
|
|
|
// Get display time
|
|
XrTime display_time = get_predicted_display_time();
|
|
if (display_time == 0) {
|
|
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
|
|
}
|
|
|
|
XrSpaceVelocity velocity = {
|
|
XR_TYPE_SPACE_VELOCITY, // type
|
|
nullptr, // next
|
|
0, // velocityFlags
|
|
{ 0.0, 0.0, 0.0 }, // linearVelocity
|
|
{ 0.0, 0.0, 0.0 } // angularVelocity
|
|
};
|
|
|
|
XrSpaceLocation location = {
|
|
XR_TYPE_SPACE_LOCATION, // type
|
|
&velocity, // next
|
|
0, // locationFlags
|
|
{
|
|
{ 0.0, 0.0, 0.0, 0.0 }, // orientation
|
|
{ 0.0, 0.0, 0.0 } // position
|
|
} // pose
|
|
};
|
|
|
|
result = xrLocateSpace(view_space, play_space, display_time, &location);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to locate view space in play space [", get_error_string(result), "]");
|
|
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
|
|
}
|
|
|
|
XRPose::TrackingConfidence confidence = transform_from_location(location, r_transform);
|
|
parse_velocities(velocity, r_linear_velocity, r_angular_velocity);
|
|
|
|
if (head_pose_confidence != confidence) {
|
|
// prevent error spam
|
|
head_pose_confidence = confidence;
|
|
if (head_pose_confidence == XRPose::XR_TRACKING_CONFIDENCE_NONE) {
|
|
print_line("OpenXR head space location not valid (check tracking?)");
|
|
} else if (head_pose_confidence == XRPose::XR_TRACKING_CONFIDENCE_LOW) {
|
|
print_verbose("OpenVR Head pose now tracking with low confidence");
|
|
} else {
|
|
print_verbose("OpenVR Head pose now tracking with high confidence");
|
|
}
|
|
}
|
|
|
|
return confidence;
|
|
}
|
|
|
|
bool OpenXRAPI::get_view_transform(uint32_t p_view, Transform3D &r_transform) {
|
|
ERR_NOT_ON_RENDER_THREAD_V(false);
|
|
|
|
if (!render_state.running) {
|
|
return false;
|
|
}
|
|
|
|
// we don't have valid view info
|
|
if (render_state.views == nullptr || !render_state.view_pose_valid) {
|
|
return false;
|
|
}
|
|
|
|
// Note, the timing of this is set right before rendering, which is what we need here.
|
|
r_transform = transform_from_pose(render_state.views[p_view].pose);
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::get_view_projection(uint32_t p_view, double p_z_near, double p_z_far, Projection &p_camera_matrix) {
|
|
ERR_NOT_ON_RENDER_THREAD_V(false);
|
|
ERR_FAIL_NULL_V(graphics_extension, false);
|
|
|
|
if (!render_state.running) {
|
|
return false;
|
|
}
|
|
|
|
// we don't have valid view info
|
|
if (render_state.views == nullptr || !render_state.view_pose_valid) {
|
|
return false;
|
|
}
|
|
|
|
// if we're using depth views, make sure we update our near and far there...
|
|
if (render_state.depth_views != nullptr) {
|
|
for (uint32_t i = 0; i < render_state.view_count; i++) {
|
|
// As we are using reverse-Z these need to be flipped.
|
|
render_state.depth_views[i].nearZ = p_z_far;
|
|
render_state.depth_views[i].farZ = p_z_near;
|
|
}
|
|
}
|
|
|
|
// now update our projection
|
|
return graphics_extension->create_projection_fov(render_state.views[p_view].fov, p_z_near, p_z_far, p_camera_matrix);
|
|
}
|
|
|
|
Vector2 OpenXRAPI::get_eye_focus(uint32_t p_view, float p_aspect) {
|
|
ERR_FAIL_NULL_V(graphics_extension, Vector2());
|
|
|
|
if (!render_state.running) {
|
|
return Vector2();
|
|
}
|
|
|
|
// xrWaitFrame not run yet
|
|
if (render_state.predicted_display_time == 0) {
|
|
return Vector2();
|
|
}
|
|
|
|
// we don't have valid view info
|
|
if (render_state.views == nullptr || !render_state.view_pose_valid) {
|
|
return Vector2();
|
|
}
|
|
|
|
Projection cm;
|
|
if (!graphics_extension->create_projection_fov(render_state.views[p_view].fov, 0.1, 1000.0, cm)) {
|
|
return Vector2();
|
|
}
|
|
|
|
// Default focus to center...
|
|
Vector3 focus = cm.xform(Vector3(0.0, 0.0, 999.9));
|
|
|
|
// Lets check for eye tracking...
|
|
OpenXREyeGazeInteractionExtension *eye_gaze_interaction = OpenXREyeGazeInteractionExtension::get_singleton();
|
|
if (eye_gaze_interaction && eye_gaze_interaction->supports_eye_gaze_interaction()) {
|
|
Vector3 eye_gaze_pose;
|
|
if (eye_gaze_interaction->get_eye_gaze_pose(1.0, eye_gaze_pose)) {
|
|
Transform3D view_transform = transform_from_pose(render_state.views[p_view].pose);
|
|
|
|
eye_gaze_pose = view_transform.xform_inv(eye_gaze_pose);
|
|
focus = cm.xform(eye_gaze_pose);
|
|
}
|
|
}
|
|
|
|
return Vector2(focus.x, focus.y);
|
|
}
|
|
|
|
bool OpenXRAPI::poll_events() {
|
|
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
|
|
|
|
XrEventDataBuffer runtimeEvent;
|
|
runtimeEvent.type = XR_TYPE_EVENT_DATA_BUFFER;
|
|
runtimeEvent.next = nullptr;
|
|
// runtimeEvent.varying = ...
|
|
|
|
XrResult pollResult = xrPollEvent(instance, &runtimeEvent);
|
|
while (pollResult == XR_SUCCESS) {
|
|
bool handled = false;
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
handled |= wrapper->on_event_polled(runtimeEvent);
|
|
}
|
|
switch (runtimeEvent.type) {
|
|
case XR_TYPE_EVENT_DATA_EVENTS_LOST: {
|
|
XrEventDataEventsLost *event = (XrEventDataEventsLost *)&runtimeEvent;
|
|
|
|
// We probably didn't poll fast enough, just output warning
|
|
WARN_PRINT("OpenXR EVENT: " + itos(event->lostEventCount) + " event data lost!");
|
|
} break;
|
|
case XR_TYPE_EVENT_DATA_INSTANCE_LOSS_PENDING: {
|
|
XrEventDataInstanceLossPending *event = (XrEventDataInstanceLossPending *)&runtimeEvent;
|
|
|
|
// TODO We get this event if we're about to loose our OpenXR instance.
|
|
// We should queue exiting Godot at this point.
|
|
|
|
print_verbose(String("OpenXR EVENT: instance loss pending at ") + itos(event->lossTime));
|
|
return false;
|
|
} break;
|
|
case XR_TYPE_EVENT_DATA_SESSION_STATE_CHANGED: {
|
|
XrEventDataSessionStateChanged *event = (XrEventDataSessionStateChanged *)&runtimeEvent;
|
|
|
|
session_state = event->state;
|
|
if (session_state >= XR_SESSION_STATE_MAX_ENUM) {
|
|
print_verbose(String("OpenXR EVENT: session state changed to UNKNOWN - ") + itos(session_state));
|
|
} else {
|
|
print_verbose(String("OpenXR EVENT: session state changed to ") + OpenXRUtil::get_session_state_name(session_state));
|
|
|
|
switch (session_state) {
|
|
case XR_SESSION_STATE_IDLE:
|
|
on_state_idle();
|
|
break;
|
|
case XR_SESSION_STATE_READY:
|
|
on_state_ready();
|
|
break;
|
|
case XR_SESSION_STATE_SYNCHRONIZED:
|
|
on_state_synchronized();
|
|
break;
|
|
case XR_SESSION_STATE_VISIBLE:
|
|
on_state_visible();
|
|
break;
|
|
case XR_SESSION_STATE_FOCUSED:
|
|
on_state_focused();
|
|
break;
|
|
case XR_SESSION_STATE_STOPPING:
|
|
on_state_stopping();
|
|
break;
|
|
case XR_SESSION_STATE_LOSS_PENDING:
|
|
on_state_loss_pending();
|
|
break;
|
|
case XR_SESSION_STATE_EXITING:
|
|
on_state_exiting();
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
} break;
|
|
case XR_TYPE_EVENT_DATA_REFERENCE_SPACE_CHANGE_PENDING: {
|
|
XrEventDataReferenceSpaceChangePending *event = (XrEventDataReferenceSpaceChangePending *)&runtimeEvent;
|
|
|
|
print_verbose(String("OpenXR EVENT: reference space type ") + OpenXRUtil::get_reference_space_name(event->referenceSpaceType) + " change pending!");
|
|
if (local_floor_emulation.enabled) {
|
|
local_floor_emulation.should_reset_floor_height = true;
|
|
}
|
|
if (event->poseValid && xr_interface) {
|
|
xr_interface->on_pose_recentered();
|
|
}
|
|
} break;
|
|
case XR_TYPE_EVENT_DATA_INTERACTION_PROFILE_CHANGED: {
|
|
print_verbose("OpenXR EVENT: interaction profile changed!");
|
|
|
|
XrEventDataInteractionProfileChanged *event = (XrEventDataInteractionProfileChanged *)&runtimeEvent;
|
|
|
|
List<RID> trackers;
|
|
tracker_owner.get_owned_list(&trackers);
|
|
for (const RID &tracker : trackers) {
|
|
tracker_check_profile(tracker, event->session);
|
|
}
|
|
|
|
} break;
|
|
default:
|
|
if (!handled) {
|
|
print_verbose(String("OpenXR Unhandled event type ") + OpenXRUtil::get_structure_type_name(runtimeEvent.type));
|
|
}
|
|
break;
|
|
}
|
|
|
|
runtimeEvent.type = XR_TYPE_EVENT_DATA_BUFFER;
|
|
pollResult = xrPollEvent(instance, &runtimeEvent);
|
|
}
|
|
|
|
if (pollResult == XR_EVENT_UNAVAILABLE) {
|
|
// processed all events in the queue
|
|
return true;
|
|
} else {
|
|
ERR_FAIL_V_MSG(false, "OpenXR: Failed to poll events!");
|
|
}
|
|
}
|
|
|
|
void OpenXRAPI::_allocate_view_buffers(uint32_t p_view_count, bool p_submit_depth_buffer) {
|
|
// Must be called from rendering thread!
|
|
ERR_NOT_ON_RENDER_THREAD;
|
|
|
|
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
|
|
ERR_FAIL_NULL(openxr_api);
|
|
|
|
openxr_api->render_state.view_count = p_view_count;
|
|
openxr_api->render_state.submit_depth_buffer = p_submit_depth_buffer;
|
|
|
|
// Allocate buffers we'll be populating with view information.
|
|
openxr_api->render_state.views = (XrView *)memalloc(sizeof(XrView) * p_view_count);
|
|
ERR_FAIL_NULL_MSG(openxr_api->render_state.views, "OpenXR Couldn't allocate memory for views");
|
|
memset(openxr_api->render_state.views, 0, sizeof(XrView) * p_view_count);
|
|
|
|
openxr_api->render_state.projection_views = (XrCompositionLayerProjectionView *)memalloc(sizeof(XrCompositionLayerProjectionView) * p_view_count);
|
|
ERR_FAIL_NULL_MSG(openxr_api->render_state.projection_views, "OpenXR Couldn't allocate memory for projection views");
|
|
memset(openxr_api->render_state.projection_views, 0, sizeof(XrCompositionLayerProjectionView) * p_view_count);
|
|
|
|
if (p_submit_depth_buffer && OpenXRCompositionLayerDepthExtension::get_singleton()->is_available()) {
|
|
openxr_api->render_state.depth_views = (XrCompositionLayerDepthInfoKHR *)memalloc(sizeof(XrCompositionLayerDepthInfoKHR) * p_view_count);
|
|
ERR_FAIL_NULL_MSG(openxr_api->render_state.depth_views, "OpenXR Couldn't allocate memory for depth views");
|
|
memset(openxr_api->render_state.depth_views, 0, sizeof(XrCompositionLayerDepthInfoKHR) * p_view_count);
|
|
}
|
|
}
|
|
|
|
void OpenXRAPI::_set_render_session_running(bool p_is_running) {
|
|
// Must be called from rendering thread!
|
|
ERR_NOT_ON_RENDER_THREAD;
|
|
|
|
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
|
|
ERR_FAIL_NULL(openxr_api);
|
|
openxr_api->render_state.running = p_is_running;
|
|
}
|
|
|
|
void OpenXRAPI::_set_render_display_info(XrTime p_predicted_display_time, bool p_should_render) {
|
|
// Must be called from rendering thread!
|
|
ERR_NOT_ON_RENDER_THREAD;
|
|
|
|
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
|
|
ERR_FAIL_NULL(openxr_api);
|
|
openxr_api->render_state.predicted_display_time = p_predicted_display_time;
|
|
openxr_api->render_state.should_render = p_should_render;
|
|
}
|
|
|
|
void OpenXRAPI::_set_render_play_space(uint64_t p_play_space) {
|
|
// Must be called from rendering thread!
|
|
ERR_NOT_ON_RENDER_THREAD;
|
|
|
|
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
|
|
ERR_FAIL_NULL(openxr_api);
|
|
openxr_api->render_state.play_space = XrSpace(p_play_space);
|
|
}
|
|
|
|
void OpenXRAPI::_set_render_state_multiplier(double p_render_target_size_multiplier) {
|
|
// Must be called from rendering thread!
|
|
ERR_NOT_ON_RENDER_THREAD;
|
|
|
|
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
|
|
ERR_FAIL_NULL(openxr_api);
|
|
openxr_api->render_state.render_target_size_multiplier = p_render_target_size_multiplier;
|
|
}
|
|
|
|
bool OpenXRAPI::process() {
|
|
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
|
|
|
|
if (!poll_events()) {
|
|
return false;
|
|
}
|
|
|
|
if (!running) {
|
|
return false;
|
|
}
|
|
|
|
// We call xrWaitFrame as early as possible, this will allow OpenXR to get
|
|
// proper timing info between this point, and when we're ready to start rendering.
|
|
// As the name suggests, OpenXR can pause the thread to minimize the time between
|
|
// retrieving tracking data and using that tracking data to render.
|
|
// OpenXR thus works best if rendering is performed on a separate thread.
|
|
XrFrameWaitInfo frame_wait_info = { XR_TYPE_FRAME_WAIT_INFO, nullptr };
|
|
frame_state.predictedDisplayTime = 0;
|
|
frame_state.predictedDisplayPeriod = 0;
|
|
frame_state.shouldRender = false;
|
|
|
|
XrResult result = xrWaitFrame(session, &frame_wait_info, &frame_state);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: xrWaitFrame() was not successful [", get_error_string(result), "]");
|
|
|
|
// reset just in case
|
|
frame_state.predictedDisplayTime = 0;
|
|
frame_state.predictedDisplayPeriod = 0;
|
|
frame_state.shouldRender = false;
|
|
|
|
set_render_display_info(0, false);
|
|
|
|
return false;
|
|
}
|
|
|
|
if (frame_state.predictedDisplayPeriod > 500000000) {
|
|
// display period more then 0.5 seconds? must be wrong data
|
|
print_verbose(String("OpenXR resetting invalid display period ") + rtos(frame_state.predictedDisplayPeriod));
|
|
frame_state.predictedDisplayPeriod = 0;
|
|
}
|
|
|
|
set_render_display_info(frame_state.predictedDisplayTime, frame_state.shouldRender);
|
|
|
|
// This is before setup_play_space() to ensure that it happens on the frame after
|
|
// the play space has been created.
|
|
if (unlikely(local_floor_emulation.should_reset_floor_height && !play_space_is_dirty)) {
|
|
reset_emulated_floor_height();
|
|
local_floor_emulation.should_reset_floor_height = false;
|
|
}
|
|
|
|
if (unlikely(play_space_is_dirty)) {
|
|
setup_play_space();
|
|
play_space_is_dirty = false;
|
|
}
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_process();
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void OpenXRAPI::free_main_swapchains() {
|
|
for (int i = 0; i < OPENXR_SWAPCHAIN_MAX; i++) {
|
|
render_state.main_swapchains[i].queue_free();
|
|
}
|
|
}
|
|
|
|
void OpenXRAPI::pre_render() {
|
|
ERR_FAIL_COND(session == XR_NULL_HANDLE);
|
|
|
|
// Must be called from rendering thread!
|
|
ERR_NOT_ON_RENDER_THREAD;
|
|
|
|
if (!render_state.running) {
|
|
return;
|
|
}
|
|
|
|
// Process any swapchains that were queued to be freed
|
|
OpenXRSwapChainInfo::free_queued();
|
|
|
|
Size2i swapchain_size = get_recommended_target_size();
|
|
if (swapchain_size != render_state.main_swapchain_size) {
|
|
// Out with the old.
|
|
free_main_swapchains();
|
|
|
|
// In with the new.
|
|
create_main_swapchains(swapchain_size);
|
|
}
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_pre_render();
|
|
}
|
|
|
|
// Get our view info for the frame we're about to render, note from the OpenXR manual:
|
|
// "Repeatedly calling xrLocateViews with the same time may not necessarily return the same result. Instead the prediction gets increasingly accurate as the function is called closer to the given time for which a prediction is made"
|
|
|
|
// We're calling this "relatively" early, the positioning we're obtaining here will be used to do our frustum culling,
|
|
// occlusion culling, etc. There is however a technique that we can investigate in the future where after our entire
|
|
// Vulkan command buffer is build, but right before vkSubmitQueue is called, we call xrLocateViews one more time and
|
|
// update the view and projection matrix once more with a slightly more accurate predication and then submit the
|
|
// command queues.
|
|
|
|
// That is not possible yet but worth investigating in the future.
|
|
|
|
XrViewLocateInfo view_locate_info = {
|
|
XR_TYPE_VIEW_LOCATE_INFO, // type
|
|
nullptr, // next
|
|
view_configuration, // viewConfigurationType
|
|
render_state.predicted_display_time, // displayTime
|
|
render_state.play_space // space
|
|
};
|
|
XrViewState view_state = {
|
|
XR_TYPE_VIEW_STATE, // type
|
|
nullptr, // next
|
|
0 // viewStateFlags
|
|
};
|
|
uint32_t view_count_output;
|
|
XrResult result = xrLocateViews(session, &view_locate_info, &view_state, render_state.view_count, &view_count_output, render_state.views);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Couldn't locate views [", get_error_string(result), "]");
|
|
return;
|
|
}
|
|
|
|
bool pose_valid = true;
|
|
for (uint64_t i = 0; i < view_count_output; i++) {
|
|
if ((view_state.viewStateFlags & XR_VIEW_STATE_ORIENTATION_VALID_BIT) == 0 ||
|
|
(view_state.viewStateFlags & XR_VIEW_STATE_POSITION_VALID_BIT) == 0) {
|
|
pose_valid = false;
|
|
}
|
|
}
|
|
if (render_state.view_pose_valid != pose_valid) {
|
|
render_state.view_pose_valid = pose_valid;
|
|
if (!render_state.view_pose_valid) {
|
|
print_verbose("OpenXR View pose became invalid");
|
|
} else {
|
|
print_verbose("OpenXR View pose became valid");
|
|
}
|
|
}
|
|
|
|
// We should get our frame no from the rendering server, but this will do.
|
|
begin_debug_label_region(String("Session Frame ") + String::num_uint64(++render_state.frame));
|
|
|
|
// let's start our frame..
|
|
XrFrameBeginInfo frame_begin_info = {
|
|
XR_TYPE_FRAME_BEGIN_INFO, // type
|
|
nullptr // next
|
|
};
|
|
result = xrBeginFrame(session, &frame_begin_info);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to being frame [", get_error_string(result), "]");
|
|
return;
|
|
}
|
|
|
|
// Reset this, we haven't found a viewport for output yet
|
|
render_state.has_xr_viewport = false;
|
|
}
|
|
|
|
bool OpenXRAPI::pre_draw_viewport(RID p_render_target) {
|
|
// Must be called from rendering thread!
|
|
ERR_NOT_ON_RENDER_THREAD_V(false);
|
|
|
|
// We found an XR viewport!
|
|
render_state.has_xr_viewport = true;
|
|
|
|
if (instance == XR_NULL_HANDLE || session == XR_NULL_HANDLE || !render_state.running || !render_state.view_pose_valid || !render_state.should_render) {
|
|
return false;
|
|
}
|
|
|
|
// Acquire our images
|
|
for (int i = 0; i < OPENXR_SWAPCHAIN_MAX; i++) {
|
|
if (!render_state.main_swapchains[i].is_image_acquired() && render_state.main_swapchains[i].get_swapchain() != XR_NULL_HANDLE) {
|
|
if (!render_state.main_swapchains[i].acquire(render_state.should_render)) {
|
|
return false;
|
|
}
|
|
}
|
|
}
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_pre_draw_viewport(p_render_target);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
XrSwapchain OpenXRAPI::get_color_swapchain() {
|
|
ERR_NOT_ON_RENDER_THREAD_V(XR_NULL_HANDLE);
|
|
|
|
return render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].get_swapchain();
|
|
}
|
|
|
|
RID OpenXRAPI::get_color_texture() {
|
|
ERR_NOT_ON_RENDER_THREAD_V(RID());
|
|
|
|
return render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].get_image();
|
|
}
|
|
|
|
RID OpenXRAPI::get_depth_texture() {
|
|
ERR_NOT_ON_RENDER_THREAD_V(RID());
|
|
|
|
// Note, image will not be acquired if we didn't have a suitable swap chain format.
|
|
if (render_state.submit_depth_buffer && render_state.main_swapchains[OPENXR_SWAPCHAIN_DEPTH].is_image_acquired()) {
|
|
return render_state.main_swapchains[OPENXR_SWAPCHAIN_DEPTH].get_image();
|
|
} else {
|
|
return RID();
|
|
}
|
|
}
|
|
|
|
void OpenXRAPI::post_draw_viewport(RID p_render_target) {
|
|
// Must be called from rendering thread!
|
|
ERR_NOT_ON_RENDER_THREAD;
|
|
|
|
if (instance == XR_NULL_HANDLE || session == XR_NULL_HANDLE || !render_state.running || !render_state.view_pose_valid || !render_state.should_render) {
|
|
return;
|
|
}
|
|
|
|
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
|
|
wrapper->on_post_draw_viewport(p_render_target);
|
|
}
|
|
};
|
|
|
|
void OpenXRAPI::end_frame() {
|
|
XrResult result;
|
|
|
|
ERR_FAIL_COND(session == XR_NULL_HANDLE);
|
|
|
|
// Must be called from rendering thread!
|
|
ERR_NOT_ON_RENDER_THREAD;
|
|
|
|
if (!render_state.running) {
|
|
return;
|
|
}
|
|
|
|
if (render_state.should_render && render_state.view_pose_valid) {
|
|
if (!render_state.has_xr_viewport) {
|
|
print_line("OpenXR: No viewport was marked with use_xr, there is no rendered output!");
|
|
} else if (!render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].is_image_acquired()) {
|
|
print_line("OpenXR: No swapchain could be acquired to render to!");
|
|
}
|
|
}
|
|
|
|
// must have:
|
|
// - should_render set to true
|
|
// - a valid view pose for projection_views[eye].pose to submit layer
|
|
// - an image to render
|
|
if (!render_state.should_render || !render_state.view_pose_valid || !render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].is_image_acquired()) {
|
|
// submit 0 layers when we shouldn't render
|
|
XrFrameEndInfo frame_end_info = {
|
|
XR_TYPE_FRAME_END_INFO, // type
|
|
nullptr, // next
|
|
render_state.predicted_display_time, // displayTime
|
|
environment_blend_mode, // environmentBlendMode
|
|
0, // layerCount
|
|
nullptr // layers
|
|
};
|
|
result = xrEndFrame(session, &frame_end_info);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: rendering skipped and failed to end frame! [", get_error_string(result), "]");
|
|
return;
|
|
}
|
|
|
|
end_debug_label_region(); // Session frame #
|
|
|
|
// neither eye is rendered
|
|
return;
|
|
}
|
|
|
|
// release our swapchain image if we acquired it
|
|
for (int i = 0; i < OPENXR_SWAPCHAIN_MAX; i++) {
|
|
if (render_state.main_swapchains[i].is_image_acquired()) {
|
|
render_state.main_swapchains[i].release();
|
|
}
|
|
}
|
|
|
|
for (uint32_t eye = 0; eye < render_state.view_count; eye++) {
|
|
render_state.projection_views[eye].fov = render_state.views[eye].fov;
|
|
render_state.projection_views[eye].pose = render_state.views[eye].pose;
|
|
}
|
|
|
|
Vector<OrderedCompositionLayer> ordered_layers_list;
|
|
bool projection_layer_is_first = true;
|
|
|
|
// Add composition layers from providers
|
|
for (OpenXRCompositionLayerProvider *provider : composition_layer_providers) {
|
|
for (int i = 0; i < provider->get_composition_layer_count(); i++) {
|
|
OrderedCompositionLayer layer = {
|
|
provider->get_composition_layer(i),
|
|
provider->get_composition_layer_order(i),
|
|
};
|
|
if (layer.composition_layer) {
|
|
ordered_layers_list.push_back(layer);
|
|
if (layer.sort_order == 0) {
|
|
WARN_PRINT_ONCE_ED("Composition layer returned sort order 0, it may be overwritten by projection layer.");
|
|
} else if (layer.sort_order < 0) {
|
|
projection_layer_is_first = false;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
XrCompositionLayerFlags layer_flags = XR_COMPOSITION_LAYER_CORRECT_CHROMATIC_ABERRATION_BIT;
|
|
if (!projection_layer_is_first || environment_blend_mode != XR_ENVIRONMENT_BLEND_MODE_OPAQUE) {
|
|
layer_flags |= XR_COMPOSITION_LAYER_BLEND_TEXTURE_SOURCE_ALPHA_BIT;
|
|
}
|
|
|
|
XrCompositionLayerProjection projection_layer = {
|
|
XR_TYPE_COMPOSITION_LAYER_PROJECTION, // type
|
|
nullptr, // next
|
|
layer_flags, // layerFlags
|
|
render_state.play_space, // space
|
|
render_state.view_count, // viewCount
|
|
render_state.projection_views, // views
|
|
};
|
|
ordered_layers_list.push_back({ (const XrCompositionLayerBaseHeader *)&projection_layer, 0 });
|
|
|
|
// Sort our layers.
|
|
ordered_layers_list.sort_custom<OrderedCompositionLayer>();
|
|
|
|
// Now make a list we can pass on to OpenXR.
|
|
Vector<const XrCompositionLayerBaseHeader *> layers_list;
|
|
for (OrderedCompositionLayer &ordered_layer : ordered_layers_list) {
|
|
layers_list.push_back(ordered_layer.composition_layer);
|
|
}
|
|
|
|
XrFrameEndInfo frame_end_info = {
|
|
XR_TYPE_FRAME_END_INFO, // type
|
|
nullptr, // next
|
|
render_state.predicted_display_time, // displayTime
|
|
environment_blend_mode, // environmentBlendMode
|
|
static_cast<uint32_t>(layers_list.size()), // layerCount
|
|
layers_list.ptr() // layers
|
|
};
|
|
result = xrEndFrame(session, &frame_end_info);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to end frame! [", get_error_string(result), "]");
|
|
return;
|
|
}
|
|
|
|
end_debug_label_region(); // Session frame #
|
|
}
|
|
|
|
float OpenXRAPI::get_display_refresh_rate() const {
|
|
OpenXRDisplayRefreshRateExtension *drrext = OpenXRDisplayRefreshRateExtension::get_singleton();
|
|
if (drrext) {
|
|
return drrext->get_refresh_rate();
|
|
}
|
|
|
|
return 0.0;
|
|
}
|
|
|
|
void OpenXRAPI::set_display_refresh_rate(float p_refresh_rate) {
|
|
OpenXRDisplayRefreshRateExtension *drrext = OpenXRDisplayRefreshRateExtension::get_singleton();
|
|
if (drrext != nullptr) {
|
|
drrext->set_refresh_rate(p_refresh_rate);
|
|
}
|
|
}
|
|
|
|
Array OpenXRAPI::get_available_display_refresh_rates() const {
|
|
OpenXRDisplayRefreshRateExtension *drrext = OpenXRDisplayRefreshRateExtension::get_singleton();
|
|
if (drrext != nullptr) {
|
|
return drrext->get_available_refresh_rates();
|
|
}
|
|
|
|
return Array();
|
|
}
|
|
|
|
double OpenXRAPI::get_render_target_size_multiplier() const {
|
|
return render_target_size_multiplier;
|
|
}
|
|
|
|
void OpenXRAPI::set_render_target_size_multiplier(double multiplier) {
|
|
render_target_size_multiplier = multiplier;
|
|
set_render_state_multiplier(multiplier);
|
|
}
|
|
|
|
bool OpenXRAPI::is_foveation_supported() const {
|
|
OpenXRFBFoveationExtension *fov_ext = OpenXRFBFoveationExtension::get_singleton();
|
|
return fov_ext != nullptr && fov_ext->is_enabled();
|
|
}
|
|
|
|
int OpenXRAPI::get_foveation_level() const {
|
|
OpenXRFBFoveationExtension *fov_ext = OpenXRFBFoveationExtension::get_singleton();
|
|
if (fov_ext != nullptr && fov_ext->is_enabled()) {
|
|
switch (fov_ext->get_foveation_level()) {
|
|
case XR_FOVEATION_LEVEL_NONE_FB:
|
|
return 0;
|
|
case XR_FOVEATION_LEVEL_LOW_FB:
|
|
return 1;
|
|
case XR_FOVEATION_LEVEL_MEDIUM_FB:
|
|
return 2;
|
|
case XR_FOVEATION_LEVEL_HIGH_FB:
|
|
return 3;
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void OpenXRAPI::set_foveation_level(int p_foveation_level) {
|
|
ERR_FAIL_UNSIGNED_INDEX(p_foveation_level, 4);
|
|
OpenXRFBFoveationExtension *fov_ext = OpenXRFBFoveationExtension::get_singleton();
|
|
if (fov_ext != nullptr && fov_ext->is_enabled()) {
|
|
XrFoveationLevelFB levels[] = { XR_FOVEATION_LEVEL_NONE_FB, XR_FOVEATION_LEVEL_LOW_FB, XR_FOVEATION_LEVEL_MEDIUM_FB, XR_FOVEATION_LEVEL_HIGH_FB };
|
|
fov_ext->set_foveation_level(levels[p_foveation_level]);
|
|
}
|
|
}
|
|
|
|
bool OpenXRAPI::get_foveation_dynamic() const {
|
|
OpenXRFBFoveationExtension *fov_ext = OpenXRFBFoveationExtension::get_singleton();
|
|
if (fov_ext != nullptr && fov_ext->is_enabled()) {
|
|
return fov_ext->get_foveation_dynamic() == XR_FOVEATION_DYNAMIC_LEVEL_ENABLED_FB;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
void OpenXRAPI::set_foveation_dynamic(bool p_foveation_dynamic) {
|
|
OpenXRFBFoveationExtension *fov_ext = OpenXRFBFoveationExtension::get_singleton();
|
|
if (fov_ext != nullptr && fov_ext->is_enabled()) {
|
|
fov_ext->set_foveation_dynamic(p_foveation_dynamic ? XR_FOVEATION_DYNAMIC_LEVEL_ENABLED_FB : XR_FOVEATION_DYNAMIC_DISABLED_FB);
|
|
}
|
|
}
|
|
|
|
Size2 OpenXRAPI::get_play_space_bounds() const {
|
|
Size2 ret;
|
|
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, Size2());
|
|
|
|
XrExtent2Df extents;
|
|
|
|
XrResult result = xrGetReferenceSpaceBoundsRect(session, reference_space, &extents);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to get play space bounds! [", get_error_string(result), "]");
|
|
return ret;
|
|
}
|
|
|
|
ret.width = extents.width;
|
|
ret.height = extents.height;
|
|
|
|
return ret;
|
|
}
|
|
|
|
OpenXRAPI::OpenXRAPI() {
|
|
// OpenXRAPI is only constructed if OpenXR is enabled.
|
|
singleton = this;
|
|
|
|
if (Engine::get_singleton()->is_editor_hint()) {
|
|
// Enabled OpenXR in the editor? Adjust our settings for the editor
|
|
} else {
|
|
// Load settings from project settings
|
|
int form_factor_setting = GLOBAL_GET("xr/openxr/form_factor");
|
|
switch (form_factor_setting) {
|
|
case 0: {
|
|
form_factor = XR_FORM_FACTOR_HEAD_MOUNTED_DISPLAY;
|
|
} break;
|
|
case 1: {
|
|
form_factor = XR_FORM_FACTOR_HANDHELD_DISPLAY;
|
|
} break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
int view_configuration_setting = GLOBAL_GET("xr/openxr/view_configuration");
|
|
switch (view_configuration_setting) {
|
|
case 0: {
|
|
view_configuration = XR_VIEW_CONFIGURATION_TYPE_PRIMARY_MONO;
|
|
} break;
|
|
case 1: {
|
|
view_configuration = XR_VIEW_CONFIGURATION_TYPE_PRIMARY_STEREO;
|
|
} break;
|
|
/* we don't support quad and observer configurations (yet)
|
|
case 2: {
|
|
view_configuration = XR_VIEW_CONFIGURATION_TYPE_PRIMARY_QUAD_VARJO;
|
|
} break;
|
|
case 3: {
|
|
view_configuration = XR_VIEW_CONFIGURATION_TYPE_SECONDARY_MONO_FIRST_PERSON_OBSERVER_MSFT;
|
|
} break;
|
|
*/
|
|
default:
|
|
break;
|
|
}
|
|
|
|
int reference_space_setting = GLOBAL_GET("xr/openxr/reference_space");
|
|
switch (reference_space_setting) {
|
|
case 0: {
|
|
requested_reference_space = XR_REFERENCE_SPACE_TYPE_LOCAL;
|
|
} break;
|
|
case 1: {
|
|
requested_reference_space = XR_REFERENCE_SPACE_TYPE_STAGE;
|
|
} break;
|
|
case 2: {
|
|
requested_reference_space = XR_REFERENCE_SPACE_TYPE_LOCAL_FLOOR_EXT;
|
|
} break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
int environment_blend_mode_setting = GLOBAL_GET("xr/openxr/environment_blend_mode");
|
|
switch (environment_blend_mode_setting) {
|
|
case 0: {
|
|
environment_blend_mode = XR_ENVIRONMENT_BLEND_MODE_OPAQUE;
|
|
} break;
|
|
case 1: {
|
|
environment_blend_mode = XR_ENVIRONMENT_BLEND_MODE_ADDITIVE;
|
|
} break;
|
|
case 2: {
|
|
environment_blend_mode = XR_ENVIRONMENT_BLEND_MODE_ALPHA_BLEND;
|
|
} break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
submit_depth_buffer = GLOBAL_GET("xr/openxr/submit_depth_buffer");
|
|
}
|
|
}
|
|
|
|
OpenXRAPI::~OpenXRAPI() {
|
|
// cleanup our composition layer providers
|
|
for (OpenXRCompositionLayerProvider *provider : composition_layer_providers) {
|
|
memdelete(provider);
|
|
}
|
|
composition_layer_providers.clear();
|
|
|
|
if (supported_extensions != nullptr) {
|
|
memfree(supported_extensions);
|
|
supported_extensions = nullptr;
|
|
}
|
|
|
|
if (layer_properties != nullptr) {
|
|
memfree(layer_properties);
|
|
layer_properties = nullptr;
|
|
}
|
|
|
|
#ifdef ANDROID_ENABLED
|
|
if (openxr_loader_library_handle) {
|
|
OS::get_singleton()->close_dynamic_library(openxr_loader_library_handle);
|
|
openxr_loader_library_handle = nullptr;
|
|
}
|
|
#endif
|
|
|
|
singleton = nullptr;
|
|
}
|
|
|
|
Transform3D OpenXRAPI::transform_from_pose(const XrPosef &p_pose) {
|
|
Quaternion q(p_pose.orientation.x, p_pose.orientation.y, p_pose.orientation.z, p_pose.orientation.w);
|
|
Basis basis(q);
|
|
Vector3 origin(p_pose.position.x, p_pose.position.y, p_pose.position.z);
|
|
|
|
return Transform3D(basis, origin);
|
|
}
|
|
|
|
template <typename T>
|
|
XRPose::TrackingConfidence _transform_from_location(const T &p_location, Transform3D &r_transform) {
|
|
Basis basis;
|
|
Vector3 origin;
|
|
XRPose::TrackingConfidence confidence = XRPose::XR_TRACKING_CONFIDENCE_NONE;
|
|
const XrPosef &pose = p_location.pose;
|
|
|
|
// Check orientation
|
|
if (p_location.locationFlags & XR_SPACE_LOCATION_ORIENTATION_VALID_BIT) {
|
|
Quaternion q(pose.orientation.x, pose.orientation.y, pose.orientation.z, pose.orientation.w);
|
|
r_transform.basis = Basis(q);
|
|
|
|
if (p_location.locationFlags & XR_SPACE_LOCATION_ORIENTATION_TRACKED_BIT) {
|
|
// Fully valid orientation, so either 3DOF or 6DOF tracking with high confidence so default to HIGH_TRACKING
|
|
confidence = XRPose::XR_TRACKING_CONFIDENCE_HIGH;
|
|
} else {
|
|
// Orientation is being tracked but we're using old/predicted data, so low tracking confidence
|
|
confidence = XRPose::XR_TRACKING_CONFIDENCE_LOW;
|
|
}
|
|
} else {
|
|
r_transform.basis = Basis();
|
|
}
|
|
|
|
// Check location
|
|
if (p_location.locationFlags & XR_SPACE_LOCATION_POSITION_VALID_BIT) {
|
|
r_transform.origin = Vector3(pose.position.x, pose.position.y, pose.position.z);
|
|
|
|
if (!(p_location.locationFlags & XR_SPACE_LOCATION_ORIENTATION_TRACKED_BIT)) {
|
|
// Location is being tracked but we're using old/predicted data, so low tracking confidence
|
|
confidence = XRPose::XR_TRACKING_CONFIDENCE_LOW;
|
|
} else if (confidence == XRPose::XR_TRACKING_CONFIDENCE_NONE) {
|
|
// Position tracking without orientation tracking?
|
|
confidence = XRPose::XR_TRACKING_CONFIDENCE_HIGH;
|
|
}
|
|
} else {
|
|
// No tracking or 3DOF I guess..
|
|
r_transform.origin = Vector3();
|
|
}
|
|
|
|
return confidence;
|
|
}
|
|
|
|
XRPose::TrackingConfidence OpenXRAPI::transform_from_location(const XrSpaceLocation &p_location, Transform3D &r_transform) {
|
|
return _transform_from_location(p_location, r_transform);
|
|
}
|
|
|
|
XRPose::TrackingConfidence OpenXRAPI::transform_from_location(const XrHandJointLocationEXT &p_location, Transform3D &r_transform) {
|
|
return _transform_from_location(p_location, r_transform);
|
|
}
|
|
|
|
void OpenXRAPI::parse_velocities(const XrSpaceVelocity &p_velocity, Vector3 &r_linear_velocity, Vector3 &r_angular_velocity) {
|
|
if (p_velocity.velocityFlags & XR_SPACE_VELOCITY_LINEAR_VALID_BIT) {
|
|
XrVector3f linear_velocity = p_velocity.linearVelocity;
|
|
r_linear_velocity = Vector3(linear_velocity.x, linear_velocity.y, linear_velocity.z);
|
|
} else {
|
|
r_linear_velocity = Vector3();
|
|
}
|
|
if (p_velocity.velocityFlags & XR_SPACE_VELOCITY_ANGULAR_VALID_BIT) {
|
|
XrVector3f angular_velocity = p_velocity.angularVelocity;
|
|
r_angular_velocity = Vector3(angular_velocity.x, angular_velocity.y, angular_velocity.z);
|
|
} else {
|
|
r_angular_velocity = Vector3();
|
|
}
|
|
}
|
|
|
|
bool OpenXRAPI::xr_result(XrResult result, const char *format, Array args) const {
|
|
if (XR_SUCCEEDED(result))
|
|
return true;
|
|
|
|
char resultString[XR_MAX_RESULT_STRING_SIZE];
|
|
xrResultToString(instance, result, resultString);
|
|
|
|
print_error(String("OpenXR ") + String(format).format(args) + String(" [") + String(resultString) + String("]"));
|
|
|
|
return false;
|
|
}
|
|
|
|
RID OpenXRAPI::get_tracker_rid(XrPath p_path) {
|
|
List<RID> current;
|
|
tracker_owner.get_owned_list(¤t);
|
|
for (const RID &E : current) {
|
|
Tracker *tracker = tracker_owner.get_or_null(E);
|
|
if (tracker && tracker->toplevel_path == p_path) {
|
|
return E;
|
|
}
|
|
}
|
|
|
|
return RID();
|
|
}
|
|
|
|
RID OpenXRAPI::find_tracker(const String &p_name) {
|
|
List<RID> current;
|
|
tracker_owner.get_owned_list(¤t);
|
|
for (const RID &E : current) {
|
|
Tracker *tracker = tracker_owner.get_or_null(E);
|
|
if (tracker && tracker->name == p_name) {
|
|
return E;
|
|
}
|
|
}
|
|
|
|
return RID();
|
|
}
|
|
|
|
RID OpenXRAPI::tracker_create(const String p_name) {
|
|
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, RID());
|
|
|
|
Tracker new_tracker;
|
|
new_tracker.name = p_name;
|
|
new_tracker.toplevel_path = XR_NULL_PATH;
|
|
new_tracker.active_profile_rid = RID();
|
|
|
|
XrResult result = xrStringToPath(instance, p_name.utf8().get_data(), &new_tracker.toplevel_path);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to get path for ", p_name, "! [", get_error_string(result), "]");
|
|
return RID();
|
|
}
|
|
|
|
return tracker_owner.make_rid(new_tracker);
|
|
}
|
|
|
|
String OpenXRAPI::tracker_get_name(RID p_tracker) {
|
|
if (p_tracker.is_null()) {
|
|
return String("None");
|
|
}
|
|
|
|
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
|
|
ERR_FAIL_NULL_V(tracker, String());
|
|
|
|
return tracker->name;
|
|
}
|
|
|
|
void OpenXRAPI::tracker_check_profile(RID p_tracker, XrSession p_session) {
|
|
if (p_session == XR_NULL_HANDLE) {
|
|
p_session = session;
|
|
}
|
|
|
|
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
|
|
ERR_FAIL_NULL(tracker);
|
|
|
|
if (tracker->toplevel_path == XR_NULL_PATH) {
|
|
// no path, how was this even created?
|
|
return;
|
|
}
|
|
|
|
XrInteractionProfileState profile_state = {
|
|
XR_TYPE_INTERACTION_PROFILE_STATE, // type
|
|
nullptr, // next
|
|
XR_NULL_PATH // interactionProfile
|
|
};
|
|
|
|
XrResult result = xrGetCurrentInteractionProfile(p_session, tracker->toplevel_path, &profile_state);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: Failed to get interaction profile for", itos(tracker->toplevel_path), "[", get_error_string(result), "]");
|
|
return;
|
|
}
|
|
|
|
XrPath new_profile = profile_state.interactionProfile;
|
|
XrPath was_profile = get_interaction_profile_path(tracker->active_profile_rid);
|
|
if (was_profile != new_profile) {
|
|
tracker->active_profile_rid = get_interaction_profile_rid(new_profile);
|
|
|
|
if (xr_interface) {
|
|
xr_interface->tracker_profile_changed(p_tracker, tracker->active_profile_rid);
|
|
}
|
|
}
|
|
}
|
|
|
|
void OpenXRAPI::tracker_free(RID p_tracker) {
|
|
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
|
|
ERR_FAIL_NULL(tracker);
|
|
|
|
// there is nothing to free here
|
|
|
|
tracker_owner.free(p_tracker);
|
|
}
|
|
|
|
RID OpenXRAPI::action_set_create(const String p_name, const String p_localized_name, const int p_priority) {
|
|
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, RID());
|
|
ActionSet action_set;
|
|
|
|
action_set.name = p_name;
|
|
action_set.is_attached = false;
|
|
|
|
// create our action set...
|
|
XrActionSetCreateInfo action_set_info = {
|
|
XR_TYPE_ACTION_SET_CREATE_INFO, // type
|
|
nullptr, // next
|
|
"", // actionSetName
|
|
"", // localizedActionSetName
|
|
uint32_t(p_priority) // priority
|
|
};
|
|
|
|
copy_string_to_char_buffer(p_name, action_set_info.actionSetName, XR_MAX_ACTION_SET_NAME_SIZE);
|
|
copy_string_to_char_buffer(p_localized_name, action_set_info.localizedActionSetName, XR_MAX_LOCALIZED_ACTION_SET_NAME_SIZE);
|
|
|
|
XrResult result = xrCreateActionSet(instance, &action_set_info, &action_set.handle);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to create action set ", p_name, "! [", get_error_string(result), "]");
|
|
return RID();
|
|
}
|
|
|
|
set_object_name(XR_OBJECT_TYPE_ACTION_SET, uint64_t(action_set.handle), p_name);
|
|
|
|
return action_set_owner.make_rid(action_set);
|
|
}
|
|
|
|
String OpenXRAPI::action_set_get_name(RID p_action_set) {
|
|
if (p_action_set.is_null()) {
|
|
return String("None");
|
|
}
|
|
|
|
ActionSet *action_set = action_set_owner.get_or_null(p_action_set);
|
|
ERR_FAIL_NULL_V(action_set, String());
|
|
|
|
return action_set->name;
|
|
}
|
|
|
|
bool OpenXRAPI::attach_action_sets(const Vector<RID> &p_action_sets) {
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
|
|
|
|
Vector<XrActionSet> action_handles;
|
|
action_handles.resize(p_action_sets.size());
|
|
for (int i = 0; i < p_action_sets.size(); i++) {
|
|
ActionSet *action_set = action_set_owner.get_or_null(p_action_sets[i]);
|
|
ERR_FAIL_NULL_V(action_set, false);
|
|
|
|
if (action_set->is_attached) {
|
|
return false;
|
|
}
|
|
|
|
action_handles.set(i, action_set->handle);
|
|
}
|
|
|
|
// So according to the docs, once we attach our action set to our session it becomes read only..
|
|
// https://www.khronos.org/registry/OpenXR/specs/1.0/man/html/xrAttachSessionActionSets.html
|
|
XrSessionActionSetsAttachInfo attach_info = {
|
|
XR_TYPE_SESSION_ACTION_SETS_ATTACH_INFO, // type
|
|
nullptr, // next
|
|
(uint32_t)p_action_sets.size(), // countActionSets,
|
|
action_handles.ptr() // actionSets
|
|
};
|
|
|
|
XrResult result = xrAttachSessionActionSets(session, &attach_info);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to attach action sets! [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
for (int i = 0; i < p_action_sets.size(); i++) {
|
|
ActionSet *action_set = action_set_owner.get_or_null(p_action_sets[i]);
|
|
ERR_FAIL_NULL_V(action_set, false);
|
|
action_set->is_attached = true;
|
|
}
|
|
|
|
/* For debugging:
|
|
print_verbose("Attached set " + action_set->name);
|
|
List<RID> action_rids;
|
|
action_owner.get_owned_list(&action_rids);
|
|
for (int i = 0; i < action_rids.size(); i++) {
|
|
Action * action = action_owner.get_or_null(action_rids[i]);
|
|
if (action && action->action_set_rid == p_action_set) {
|
|
print_verbose(" - Action " + action->name + ": " + OpenXRUtil::get_action_type_name(action->action_type));
|
|
for (int j = 0; j < action->trackers.size(); j++) {
|
|
Tracker * tracker = tracker_owner.get_or_null(action->trackers[j].tracker_rid);
|
|
if (tracker) {
|
|
print_verbose(" - " + tracker->name);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
*/
|
|
|
|
return true;
|
|
}
|
|
|
|
void OpenXRAPI::action_set_free(RID p_action_set) {
|
|
ActionSet *action_set = action_set_owner.get_or_null(p_action_set);
|
|
ERR_FAIL_NULL(action_set);
|
|
|
|
if (action_set->handle != XR_NULL_HANDLE) {
|
|
xrDestroyActionSet(action_set->handle);
|
|
}
|
|
|
|
action_set_owner.free(p_action_set);
|
|
}
|
|
|
|
RID OpenXRAPI::get_action_rid(XrAction p_action) {
|
|
List<RID> current;
|
|
action_owner.get_owned_list(¤t);
|
|
for (const RID &E : current) {
|
|
Action *action = action_owner.get_or_null(E);
|
|
if (action && action->handle == p_action) {
|
|
return E;
|
|
}
|
|
}
|
|
|
|
return RID();
|
|
}
|
|
|
|
RID OpenXRAPI::find_action(const String &p_name) {
|
|
List<RID> current;
|
|
action_owner.get_owned_list(¤t);
|
|
for (const RID &E : current) {
|
|
Action *action = action_owner.get_or_null(E);
|
|
if (action && action->name == p_name) {
|
|
return E;
|
|
}
|
|
}
|
|
|
|
return RID();
|
|
}
|
|
|
|
RID OpenXRAPI::action_create(RID p_action_set, const String p_name, const String p_localized_name, OpenXRAction::ActionType p_action_type, const Vector<RID> &p_trackers) {
|
|
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, RID());
|
|
|
|
Action action;
|
|
action.name = p_name;
|
|
|
|
ActionSet *action_set = action_set_owner.get_or_null(p_action_set);
|
|
ERR_FAIL_NULL_V(action_set, RID());
|
|
ERR_FAIL_COND_V(action_set->handle == XR_NULL_HANDLE, RID());
|
|
action.action_set_rid = p_action_set;
|
|
|
|
switch (p_action_type) {
|
|
case OpenXRAction::OPENXR_ACTION_BOOL:
|
|
action.action_type = XR_ACTION_TYPE_BOOLEAN_INPUT;
|
|
break;
|
|
case OpenXRAction::OPENXR_ACTION_FLOAT:
|
|
action.action_type = XR_ACTION_TYPE_FLOAT_INPUT;
|
|
break;
|
|
case OpenXRAction::OPENXR_ACTION_VECTOR2:
|
|
action.action_type = XR_ACTION_TYPE_VECTOR2F_INPUT;
|
|
break;
|
|
case OpenXRAction::OPENXR_ACTION_POSE:
|
|
action.action_type = XR_ACTION_TYPE_POSE_INPUT;
|
|
break;
|
|
case OpenXRAction::OPENXR_ACTION_HAPTIC:
|
|
action.action_type = XR_ACTION_TYPE_VIBRATION_OUTPUT;
|
|
break;
|
|
default:
|
|
ERR_FAIL_V(RID());
|
|
break;
|
|
}
|
|
|
|
Vector<XrPath> toplevel_paths;
|
|
for (int i = 0; i < p_trackers.size(); i++) {
|
|
Tracker *tracker = tracker_owner.get_or_null(p_trackers[i]);
|
|
if (tracker != nullptr && tracker->toplevel_path != XR_NULL_PATH) {
|
|
ActionTracker action_tracker = {
|
|
p_trackers[i], // tracker
|
|
XR_NULL_HANDLE, // space
|
|
false // was_location_valid
|
|
};
|
|
action.trackers.push_back(action_tracker);
|
|
|
|
toplevel_paths.push_back(tracker->toplevel_path);
|
|
}
|
|
}
|
|
|
|
XrActionCreateInfo action_info = {
|
|
XR_TYPE_ACTION_CREATE_INFO, // type
|
|
nullptr, // next
|
|
"", // actionName
|
|
action.action_type, // actionType
|
|
uint32_t(toplevel_paths.size()), // countSubactionPaths
|
|
toplevel_paths.ptr(), // subactionPaths
|
|
"" // localizedActionName
|
|
};
|
|
|
|
copy_string_to_char_buffer(p_name, action_info.actionName, XR_MAX_ACTION_NAME_SIZE);
|
|
copy_string_to_char_buffer(p_localized_name, action_info.localizedActionName, XR_MAX_LOCALIZED_ACTION_NAME_SIZE);
|
|
|
|
XrResult result = xrCreateAction(action_set->handle, &action_info, &action.handle);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to create action ", p_name, "! [", get_error_string(result), "]");
|
|
return RID();
|
|
}
|
|
|
|
set_object_name(XR_OBJECT_TYPE_ACTION, uint64_t(action.handle), p_name);
|
|
|
|
return action_owner.make_rid(action);
|
|
}
|
|
|
|
String OpenXRAPI::action_get_name(RID p_action) {
|
|
if (p_action.is_null()) {
|
|
return String("None");
|
|
}
|
|
|
|
Action *action = action_owner.get_or_null(p_action);
|
|
ERR_FAIL_NULL_V(action, String());
|
|
|
|
return action->name;
|
|
}
|
|
|
|
void OpenXRAPI::action_free(RID p_action) {
|
|
Action *action = action_owner.get_or_null(p_action);
|
|
ERR_FAIL_NULL(action);
|
|
|
|
if (action->handle != XR_NULL_HANDLE) {
|
|
xrDestroyAction(action->handle);
|
|
}
|
|
|
|
action_owner.free(p_action);
|
|
}
|
|
|
|
RID OpenXRAPI::get_interaction_profile_rid(XrPath p_path) {
|
|
List<RID> current;
|
|
interaction_profile_owner.get_owned_list(¤t);
|
|
for (const RID &E : current) {
|
|
InteractionProfile *ip = interaction_profile_owner.get_or_null(E);
|
|
if (ip && ip->path == p_path) {
|
|
return E;
|
|
}
|
|
}
|
|
|
|
return RID();
|
|
}
|
|
|
|
XrPath OpenXRAPI::get_interaction_profile_path(RID p_interaction_profile) {
|
|
if (p_interaction_profile.is_null()) {
|
|
return XR_NULL_PATH;
|
|
}
|
|
|
|
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
|
|
ERR_FAIL_NULL_V(ip, XR_NULL_PATH);
|
|
|
|
return ip->path;
|
|
}
|
|
|
|
RID OpenXRAPI::interaction_profile_create(const String p_name) {
|
|
if (!is_interaction_profile_supported(p_name)) {
|
|
// The extension enabling this path must not be active, we will silently skip this interaction profile
|
|
return RID();
|
|
}
|
|
|
|
InteractionProfile new_interaction_profile;
|
|
|
|
XrResult result = xrStringToPath(instance, p_name.utf8().get_data(), &new_interaction_profile.path);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to get path for ", p_name, "! [", get_error_string(result), "]");
|
|
return RID();
|
|
}
|
|
|
|
RID existing_ip = get_interaction_profile_rid(new_interaction_profile.path);
|
|
if (existing_ip.is_valid()) {
|
|
return existing_ip;
|
|
}
|
|
|
|
new_interaction_profile.name = p_name;
|
|
return interaction_profile_owner.make_rid(new_interaction_profile);
|
|
}
|
|
|
|
String OpenXRAPI::interaction_profile_get_name(RID p_interaction_profile) {
|
|
if (p_interaction_profile.is_null()) {
|
|
return String("None");
|
|
}
|
|
|
|
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
|
|
ERR_FAIL_NULL_V(ip, String());
|
|
|
|
return ip->name;
|
|
}
|
|
|
|
void OpenXRAPI::interaction_profile_clear_bindings(RID p_interaction_profile) {
|
|
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
|
|
ERR_FAIL_NULL(ip);
|
|
|
|
ip->bindings.clear();
|
|
}
|
|
|
|
bool OpenXRAPI::interaction_profile_add_binding(RID p_interaction_profile, RID p_action, const String p_path) {
|
|
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
|
|
ERR_FAIL_NULL_V(ip, false);
|
|
|
|
if (!interaction_profile_supports_io_path(ip->name, p_path)) {
|
|
return false;
|
|
}
|
|
|
|
XrActionSuggestedBinding binding;
|
|
|
|
Action *action = action_owner.get_or_null(p_action);
|
|
ERR_FAIL_COND_V(action == nullptr || action->handle == XR_NULL_HANDLE, false);
|
|
|
|
binding.action = action->handle;
|
|
|
|
XrResult result = xrStringToPath(instance, p_path.utf8().get_data(), &binding.binding);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to get path for ", p_path, "! [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
ip->bindings.push_back(binding);
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::interaction_profile_suggest_bindings(RID p_interaction_profile) {
|
|
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
|
|
|
|
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
|
|
ERR_FAIL_NULL_V(ip, false);
|
|
|
|
const XrInteractionProfileSuggestedBinding suggested_bindings = {
|
|
XR_TYPE_INTERACTION_PROFILE_SUGGESTED_BINDING, // type
|
|
nullptr, // next
|
|
ip->path, // interactionProfile
|
|
uint32_t(ip->bindings.size()), // countSuggestedBindings
|
|
ip->bindings.ptr() // suggestedBindings
|
|
};
|
|
|
|
XrResult result = xrSuggestInteractionProfileBindings(instance, &suggested_bindings);
|
|
if (result == XR_ERROR_PATH_UNSUPPORTED) {
|
|
// this is fine, not all runtimes support all devices.
|
|
print_verbose("OpenXR Interaction profile " + ip->name + " is not supported on this runtime");
|
|
} else if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to suggest bindings for ", ip->name, "! [", get_error_string(result), "]");
|
|
// reporting is enough...
|
|
}
|
|
|
|
/* For debugging:
|
|
print_verbose("Suggested bindings for " + ip->name);
|
|
for (int i = 0; i < ip->bindings.size(); i++) {
|
|
uint32_t strlen;
|
|
char path[XR_MAX_PATH_LENGTH];
|
|
|
|
String action_name = action_get_name(get_action_rid(ip->bindings[i].action));
|
|
|
|
XrResult result = xrPathToString(instance, ip->bindings[i].binding, XR_MAX_PATH_LENGTH, &strlen, path);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to retrieve bindings for ", action_name, "! [", get_error_string(result), "]");
|
|
}
|
|
print_verbose(" - " + action_name + " => " + String(path));
|
|
}
|
|
*/
|
|
|
|
return true;
|
|
}
|
|
|
|
void OpenXRAPI::interaction_profile_free(RID p_interaction_profile) {
|
|
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
|
|
ERR_FAIL_NULL(ip);
|
|
|
|
ip->bindings.clear();
|
|
|
|
interaction_profile_owner.free(p_interaction_profile);
|
|
}
|
|
|
|
bool OpenXRAPI::sync_action_sets(const Vector<RID> p_active_sets) {
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
|
|
|
|
if (!running) {
|
|
return false;
|
|
}
|
|
|
|
Vector<XrActiveActionSet> active_sets;
|
|
for (int i = 0; i < p_active_sets.size(); i++) {
|
|
ActionSet *action_set = action_set_owner.get_or_null(p_active_sets[i]);
|
|
if (action_set && action_set->handle != XR_NULL_HANDLE) {
|
|
XrActiveActionSet aset;
|
|
aset.actionSet = action_set->handle;
|
|
aset.subactionPath = XR_NULL_PATH;
|
|
active_sets.push_back(aset);
|
|
}
|
|
}
|
|
|
|
ERR_FAIL_COND_V(active_sets.is_empty(), false);
|
|
|
|
XrActionsSyncInfo sync_info = {
|
|
XR_TYPE_ACTIONS_SYNC_INFO, // type
|
|
nullptr, // next
|
|
uint32_t(active_sets.size()), // countActiveActionSets
|
|
active_sets.ptr() // activeActionSets
|
|
};
|
|
|
|
XrResult result = xrSyncActions(session, &sync_info);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to sync active action sets! [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool OpenXRAPI::get_action_bool(RID p_action, RID p_tracker) {
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
|
|
Action *action = action_owner.get_or_null(p_action);
|
|
ERR_FAIL_NULL_V(action, false);
|
|
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
|
|
ERR_FAIL_NULL_V(tracker, false);
|
|
|
|
if (!running) {
|
|
return false;
|
|
}
|
|
|
|
ERR_FAIL_COND_V(action->action_type != XR_ACTION_TYPE_BOOLEAN_INPUT, false);
|
|
|
|
XrActionStateGetInfo get_info = {
|
|
XR_TYPE_ACTION_STATE_GET_INFO, // type
|
|
nullptr, // next
|
|
action->handle, // action
|
|
tracker->toplevel_path // subactionPath
|
|
};
|
|
|
|
XrActionStateBoolean result_state;
|
|
result_state.type = XR_TYPE_ACTION_STATE_BOOLEAN,
|
|
result_state.next = nullptr;
|
|
XrResult result = xrGetActionStateBoolean(session, &get_info, &result_state);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: couldn't get action boolean! [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
return result_state.isActive && result_state.currentState;
|
|
}
|
|
|
|
float OpenXRAPI::get_action_float(RID p_action, RID p_tracker) {
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, 0.0);
|
|
Action *action = action_owner.get_or_null(p_action);
|
|
ERR_FAIL_NULL_V(action, 0.0);
|
|
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
|
|
ERR_FAIL_NULL_V(tracker, 0.0);
|
|
|
|
if (!running) {
|
|
return 0.0;
|
|
}
|
|
|
|
ERR_FAIL_COND_V(action->action_type != XR_ACTION_TYPE_FLOAT_INPUT, 0.0);
|
|
|
|
XrActionStateGetInfo get_info = {
|
|
XR_TYPE_ACTION_STATE_GET_INFO, // type
|
|
nullptr, // next
|
|
action->handle, // action
|
|
tracker->toplevel_path // subactionPath
|
|
};
|
|
|
|
XrActionStateFloat result_state;
|
|
result_state.type = XR_TYPE_ACTION_STATE_FLOAT,
|
|
result_state.next = nullptr;
|
|
XrResult result = xrGetActionStateFloat(session, &get_info, &result_state);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: couldn't get action float! [", get_error_string(result), "]");
|
|
return 0.0;
|
|
}
|
|
|
|
return result_state.isActive ? result_state.currentState : 0.0;
|
|
}
|
|
|
|
Vector2 OpenXRAPI::get_action_vector2(RID p_action, RID p_tracker) {
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, Vector2());
|
|
Action *action = action_owner.get_or_null(p_action);
|
|
ERR_FAIL_NULL_V(action, Vector2());
|
|
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
|
|
ERR_FAIL_NULL_V(tracker, Vector2());
|
|
|
|
if (!running) {
|
|
return Vector2();
|
|
}
|
|
|
|
ERR_FAIL_COND_V(action->action_type != XR_ACTION_TYPE_VECTOR2F_INPUT, Vector2());
|
|
|
|
XrActionStateGetInfo get_info = {
|
|
XR_TYPE_ACTION_STATE_GET_INFO, // type
|
|
nullptr, // next
|
|
action->handle, // action
|
|
tracker->toplevel_path // subactionPath
|
|
};
|
|
|
|
XrActionStateVector2f result_state;
|
|
result_state.type = XR_TYPE_ACTION_STATE_VECTOR2F,
|
|
result_state.next = nullptr;
|
|
XrResult result = xrGetActionStateVector2f(session, &get_info, &result_state);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: couldn't get action vector2! [", get_error_string(result), "]");
|
|
return Vector2();
|
|
}
|
|
|
|
return result_state.isActive ? Vector2(result_state.currentState.x, result_state.currentState.y) : Vector2();
|
|
}
|
|
|
|
XRPose::TrackingConfidence OpenXRAPI::get_action_pose(RID p_action, RID p_tracker, Transform3D &r_transform, Vector3 &r_linear_velocity, Vector3 &r_angular_velocity) {
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, XRPose::XR_TRACKING_CONFIDENCE_NONE);
|
|
Action *action = action_owner.get_or_null(p_action);
|
|
ERR_FAIL_NULL_V(action, XRPose::XR_TRACKING_CONFIDENCE_NONE);
|
|
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
|
|
ERR_FAIL_NULL_V(tracker, XRPose::XR_TRACKING_CONFIDENCE_NONE);
|
|
|
|
if (!running) {
|
|
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
|
|
}
|
|
|
|
ERR_FAIL_COND_V(action->action_type != XR_ACTION_TYPE_POSE_INPUT, XRPose::XR_TRACKING_CONFIDENCE_NONE);
|
|
|
|
// print_verbose("Checking " + action->name + " => " + tracker->name + " (" + itos(tracker->toplevel_path) + ")");
|
|
|
|
uint64_t index = 0xFFFFFFFF;
|
|
uint64_t size = uint64_t(action->trackers.size());
|
|
for (uint64_t i = 0; i < size && index == 0xFFFFFFFF; i++) {
|
|
if (action->trackers[i].tracker_rid == p_tracker) {
|
|
index = i;
|
|
}
|
|
}
|
|
|
|
if (index == 0xFFFFFFFF) {
|
|
// couldn't find it?
|
|
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
|
|
}
|
|
|
|
XrTime display_time = get_predicted_display_time();
|
|
if (display_time == 0) {
|
|
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
|
|
}
|
|
|
|
if (action->trackers[index].space == XR_NULL_HANDLE) {
|
|
// if this is a pose we need to define spaces
|
|
|
|
XrActionSpaceCreateInfo action_space_info = {
|
|
XR_TYPE_ACTION_SPACE_CREATE_INFO, // type
|
|
nullptr, // next
|
|
action->handle, // action
|
|
tracker->toplevel_path, // subactionPath
|
|
{
|
|
{ 0.0, 0.0, 0.0, 1.0 }, // orientation
|
|
{ 0.0, 0.0, 0.0 } // position
|
|
} // poseInActionSpace
|
|
};
|
|
|
|
XrSpace space;
|
|
XrResult result = xrCreateActionSpace(session, &action_space_info, &space);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: couldn't create action space! [", get_error_string(result), "]");
|
|
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
|
|
}
|
|
|
|
action->trackers.ptrw()[index].space = space;
|
|
}
|
|
|
|
XrSpaceVelocity velocity = {
|
|
XR_TYPE_SPACE_VELOCITY, // type
|
|
nullptr, // next
|
|
0, // velocityFlags
|
|
{ 0.0, 0.0, 0.0 }, // linearVelocity
|
|
{ 0.0, 0.0, 0.0 } // angularVelocity
|
|
};
|
|
|
|
XrSpaceLocation location = {
|
|
XR_TYPE_SPACE_LOCATION, // type
|
|
&velocity, // next
|
|
0, // locationFlags
|
|
{
|
|
{ 0.0, 0.0, 0.0, 0.0 }, // orientation
|
|
{ 0.0, 0.0, 0.0 } // position
|
|
} // pose
|
|
};
|
|
|
|
XrResult result = xrLocateSpace(action->trackers[index].space, play_space, display_time, &location);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to locate space! [", get_error_string(result), "]");
|
|
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
|
|
}
|
|
|
|
XRPose::TrackingConfidence confidence = transform_from_location(location, r_transform);
|
|
parse_velocities(velocity, r_linear_velocity, r_angular_velocity);
|
|
|
|
return confidence;
|
|
}
|
|
|
|
bool OpenXRAPI::trigger_haptic_pulse(RID p_action, RID p_tracker, float p_frequency, float p_amplitude, XrDuration p_duration_ns) {
|
|
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
|
|
Action *action = action_owner.get_or_null(p_action);
|
|
ERR_FAIL_NULL_V(action, false);
|
|
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
|
|
ERR_FAIL_NULL_V(tracker, false);
|
|
|
|
if (!running) {
|
|
return false;
|
|
}
|
|
|
|
ERR_FAIL_COND_V(action->action_type != XR_ACTION_TYPE_VIBRATION_OUTPUT, false);
|
|
|
|
XrHapticActionInfo action_info = {
|
|
XR_TYPE_HAPTIC_ACTION_INFO, // type
|
|
nullptr, // next
|
|
action->handle, // action
|
|
tracker->toplevel_path // subactionPath
|
|
};
|
|
|
|
XrHapticVibration vibration = {
|
|
XR_TYPE_HAPTIC_VIBRATION, // type
|
|
nullptr, // next
|
|
p_duration_ns, // duration
|
|
p_frequency, // frequency
|
|
p_amplitude, // amplitude
|
|
};
|
|
|
|
XrResult result = xrApplyHapticFeedback(session, &action_info, (const XrHapticBaseHeader *)&vibration);
|
|
if (XR_FAILED(result)) {
|
|
print_line("OpenXR: failed to apply haptic feedback! [", get_error_string(result), "]");
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void OpenXRAPI::register_composition_layer_provider(OpenXRCompositionLayerProvider *provider) {
|
|
composition_layer_providers.append(provider);
|
|
}
|
|
|
|
void OpenXRAPI::unregister_composition_layer_provider(OpenXRCompositionLayerProvider *provider) {
|
|
composition_layer_providers.erase(provider);
|
|
}
|
|
|
|
const XrEnvironmentBlendMode *OpenXRAPI::get_supported_environment_blend_modes(uint32_t &count) {
|
|
count = num_supported_environment_blend_modes;
|
|
return supported_environment_blend_modes;
|
|
}
|
|
|
|
bool OpenXRAPI::is_environment_blend_mode_supported(XrEnvironmentBlendMode p_blend_mode) const {
|
|
ERR_FAIL_NULL_V(supported_environment_blend_modes, false);
|
|
|
|
for (uint32_t i = 0; i < num_supported_environment_blend_modes; i++) {
|
|
if (supported_environment_blend_modes[i] == p_blend_mode) {
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool OpenXRAPI::set_environment_blend_mode(XrEnvironmentBlendMode p_blend_mode) {
|
|
if (emulate_environment_blend_mode_alpha_blend && p_blend_mode == XR_ENVIRONMENT_BLEND_MODE_ALPHA_BLEND) {
|
|
requested_environment_blend_mode = XR_ENVIRONMENT_BLEND_MODE_ALPHA_BLEND;
|
|
environment_blend_mode = XR_ENVIRONMENT_BLEND_MODE_OPAQUE;
|
|
return true;
|
|
}
|
|
// We allow setting this when not initialized and will check if it is supported when initializing.
|
|
// After OpenXR is initialized we verify we're setting a supported blend mode.
|
|
else if (!is_initialized() || is_environment_blend_mode_supported(p_blend_mode)) {
|
|
requested_environment_blend_mode = p_blend_mode;
|
|
environment_blend_mode = p_blend_mode;
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
void OpenXRAPI::set_emulate_environment_blend_mode_alpha_blend(bool p_enabled) {
|
|
emulate_environment_blend_mode_alpha_blend = p_enabled;
|
|
}
|
|
|
|
OpenXRAPI::OpenXRAlphaBlendModeSupport OpenXRAPI::is_environment_blend_mode_alpha_blend_supported() {
|
|
if (is_environment_blend_mode_supported(XR_ENVIRONMENT_BLEND_MODE_ALPHA_BLEND)) {
|
|
return OPENXR_ALPHA_BLEND_MODE_SUPPORT_REAL;
|
|
} else if (emulate_environment_blend_mode_alpha_blend) {
|
|
return OPENXR_ALPHA_BLEND_MODE_SUPPORT_EMULATING;
|
|
}
|
|
return OPENXR_ALPHA_BLEND_MODE_SUPPORT_NONE;
|
|
}
|