Gestures

From Resonite Wiki
Revision as of 17:59, 23 January 2024 by 989onan (talk | contribs) (created page with some information)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

This article or section is a Stub. You can help the Resonite Wiki by expanding it.


Preface

Facial animations in Resonite are a way of showing expression on your avatar. Whether that be an avatar with stars in its eyes when you press a controller button, when you tilt your head you automatically give a confused look, or when you frown with a facial tracker your ears flop down, there are many cool and unique ways you can make facial expressions on an avatar.

Controller Buttons

The most affordable but more difficult way for facial animations is controller input. There are many different nodes that can provide controller inputs using ProtoFlux. Some of these include Standard Controller, Index Controller, and Touch Controller to name a few from the Devices/Controllers ProtoFlux Category. These nodes can provide different data like booleans for button inputs and float2s for joystick and for Index the position you are touching your TouchPad with your thumb at. These values can be used with the boolean logic nodes and basic math nodes to drive shapekeys and facial expressions.

Example circuits:

This article or section is a Stub. You can help the Resonite Wiki by expanding it.


You can also use the rotation or pointing direction of your body limbs to drive shapekeys and actions, using nodes from vector math and even transform math nodes from Transform to make body gesture based facial expressions, which can be useful for users who have only one controller, missing support for their controller, or have no hands to use said controllerw.

Example circuits:

This article or section is a Stub. You can help the Resonite Wiki by expanding it.



Facial Tracker Based

If you are using a facial tracker, Resonite itself and various community modifications supports almost every facial tracker on the market. With Resonite, supported facial tracker connection is easy and only requires the software that came with your facial tracker. Resonite will automatically hook into your facial trackers native driver software, and will start working right out the box, even if you're using a modification to the game to support your tracker.

List of trackers that are supported and if mods are required

Known Facial Trackers
Tracker Is Supported Required Mod Github Link Limitations
Vive standalone tracker Yes N/A None
Quest 3 Pro Yes

This article or section is a Stub. You can help the Resonite Wiki by expanding it.


None
Vive Pro Eye Yes N/A None

Once a tracker is connected to Resonite, avatars that support facial tracking will automatically have facial movements. Which driven by the raw data your facial tracker gives. These facial movements are driven by the Avatar Expression Driver Component. You can also use this component to drive a Float Value Field Component rather than shapekeys on a mesh, and use that as an input to ProtoFlux or other components.