|
1 | | ---- |
2 | | -title: "On the transferability and accessibility of human movement data" |
3 | | ---- |
| 1 | +# On the Transferability and Accessibility of Human Movement Data |
4 | 2 |
|
5 | | -## Introduction |
6 | | -Human motion capture is a suite of technologies and techniques to record a motion modality (i.e., position, speed, or acceleration) of the human moving parts and their environment, with extensive uses in movement science, rehabilitation, sports, and entertainment. However, the heterogeneity of the recorded modality and the spatial and temporal resolutions of the recordings make the utility and interpretability of the motion capture recordings challenging. [some examples of the importance and use of such helpful interpretability]. |
| 3 | +# Introduction |
7 | 4 |
|
8 | | -Motion capture is a general term for acquiring a motion-related physical quantity (i.e., motion modality), such as position, speed, acceleration, etc. [talk about the fundamental differences between having each quantity.] |
| 5 | +Human motion capture encompasses various technologies and techniques to record different modalities of human motion, such as position, speed, or acceleration. These techniques are extensively used in movement science, rehabilitation, sports, and entertainment. However, the heterogeneity of recorded modalities and the varying spatial and temporal resolutions of these recordings pose challenges for the utility and interpretability of motion capture data. This necessitates a clear, unified approach for sensor placement to ensure data transferability and accessibility across different systems. |
9 | 6 |
|
10 | | -Technologies related to motion capture mainly record one of the physical quantities, while the rest could, in theory, be driven by differentiation and integration coupled with some constraints. [talk about different techs, passive active marker-based, IMUs, ToF and dot projection, video cameras, and deep learning] |
| 7 | +Motion capture generally involves acquiring motion-related physical quantities (motion modalities) such as position, speed, and acceleration. The fundamental differences between these quantities highlight the need for different technologies and methods, including passive and active marker-based systems, IMUs, ToF, dot projection, video cameras, and deep learning techniques. |
11 | 8 |
|
12 | | -Limitations of each method for motion capture manifest in the interpretability and transferability of the data from one modality to another. [brief discussions about the limitations]. A solution to address these limitations is an adequate and precise annotation of the features affecting the motion capture interpretation quality. |
| 9 | +Each motion capture method has limitations, impacting the interpretability and transferability of data from one modality to another. To address these limitations, we propose a precise annotation of features affecting the quality of motion capture interpretation. |
13 | 10 |
|
14 | | -Here we 1. briefly describe the main features of each modality. 2. We then discuss the annotations required for each modality that would adequately help with the interpretability of the motion capture data. 3. We finally introduce a set of best practices in defining coordinate systems and sensor placement schemes that would provide a framework for the interpretability of the motion capture data from one modality to another. |
| 11 | +In this manuscript, we will briefly describe the main features of each modality and introduce a set of definitions that we will use throughout. We finally propose a scheme for unified sensor placement annotation with quantifiable levels of precision. We try to align this scheme with the currently available standards for data sharing and annotation, namely the **[Brain Imaging Data Structure (BIDS)](https://bids.neuroimaging.io/)** and the **[Hierarchical Event Descriptors (HED)](https://www.hedtags.org/)**. See [Motion-BIDS](https://bids-specification.readthedocs.io/en/stable/modality-specific-files/motion.html) and definition of body parts in [HED schema browser](https://www.hedtags.org/display_hed.html) for relevant details. |
15 | 12 |
|
| 13 | +# Types of Motion Capture |
| 14 | +## Optical Motion Capture (OMC) |
16 | 15 |
|
17 | | -## Types of motion capture |
18 | | -In the following chapter we will introduce three main types of motion capture systems, what type of data they record and the limits of transferability between them. We will also include how current specifications of sensor placements are done. |
| 16 | +**Optical Motion Capture (OMC)** systems utilize multiple cameras to track the movement of reflective markers placed on a subject's body or objects. These markers reflect light emitted by the cameras, allowing the system to triangulate their positions in three-dimensional space. The cameras record the positions of these markers at high frame rates, capturing data on **marker positions (POS)** and, optionally, **orientation (ORNT)**. Additionally, **velocity (VEL)**, **acceleration (ACCEL)**, and **angular acceleration (ANGACCEL)** data can be derived from the marker positions over time. **Marker placement** should define the type, size, and shape of markers, along with their specific placement on anatomical landmarks, such as joints and limb segments. |
19 | 17 |
|
20 | | -**Optical Motion Capture (OMC)**: |
21 | | -Marker based optical motion capture systems utilize multiple cameras to track the movement of reflective markers placed on a subject's body or objects. These markers reflect light emitted by the cameras, allowing the system to triangulate their positions in three-dimensional space. The cameras record the positions of these markers at high frame rates, typically capturing data on marker positions (POS) and, optionally, orientation (ORNT). Additionally, velocity (VEL), acceleration (ACCEL), and angular acceleration (ANGACCEL) data can be derived from the marker positions over time. The placement of markers should usually define marker type, size, and shape, along with their specific placement on anatomical landmarks or reference points, often including joints, limb segments, and other relevant locations. |
| 18 | +## Inertial Measurement Units (IMUs) |
22 | 19 |
|
23 | | -**Inertial Measurement Units (IMUs)**: |
24 | | -Inertial measurement units consist of small sensors, including accelerometers, gyroscopes, and sometimes magnetometers, attached to a subject's body. IMUs measure changes in acceleration, angular velocity, and magnetic field strength, allowing them to record acceleration (ACCEL), angular acceleration (ANGACCEL), velocity (VEL), and orientation (ORNT). IMUs are compact and versatile, making them suitable for wearable applications, such as sports performance analysis and motion tracking in remote or outdoor environments. Positioning of the IMU on the human body should contain information about the location and orientation. The location often follows free form text and the orientation is described using photos or some kind of orientation in relation to the body part it is attached to. |
| 20 | +**Inertial Measurement Units (IMUs)** consist of small sensors, including accelerometers, gyroscopes, and sometimes magnetometers, attached to a subject's body. IMUs measure changes in **acceleration (ACCEL)**, **angular acceleration (ANGACCEL)**, **velocity (VEL)**, and **orientation (ORNT)**. They are compact and versatile, making them suitable for wearable applications like sports performance analysis and motion tracking in remote or outdoor environments. **Positioning of the IMU** on the body should include details about the location and orientation, typically described using text and photographs or diagrams showing the sensor's orientation relative to the body part it is attached to. |
25 | 21 |
|
26 | | -**Markerless Motion Capture**: |
27 | | -Markerless motion capture systems rely on algorithms to track a subject's movements without the need for physical markers. These systems use cameras to capture video footage of the subject, and then some form of software processes the images to extract data on position (POS), orientation (ORNT), and sometimes velocity (VEL) and acceleration (ACCEL). Markerless motion capture is advantageous for its non-invasive nature and ability to capture natural movement, making it popular in the fields of entertainment, biomechanics, and human-computer interaction. The definition of the tracked points is often closely bound to the software and which points it allows to be tracked. |
| 22 | +## Markerless Motion Capture |
28 | 23 |
|
29 | | -## Definitions |
30 | | -**Space** |
31 | | -[in BIDS terms] BIDS defines “space” as an artificial frame of reference, created to describe different anatomies in a unifying manner (see Appendix VIII). However, data collected in studies of physical or virtual motion usually have a reference frame that is anchored to the physical lab space or the virtual environment. |
| 24 | +**Markerless Motion Capture** systems use algorithms to track a subject's movements without physical markers. Cameras capture video footage, which software processes to extract data on **position (POS)**, **orientation (ORNT)**, and sometimes **velocity (VEL)** and **acceleration (ACCEL)**. Markerless motion capture is non-invasive and captures natural movement, making it popular in entertainment, biomechanics, and human-computer interaction. The tracked points' definition is often software-specific, depending on which points the software allows to be tracked. |
32 | 25 |
|
33 | | -**Reference frame** |
34 | | -A reference frame, or a frame of reference is an abstract coordinate system whose origin, orientation, and scale are specified by a set of reference points (Kovalevsky & Mueller, 1989). Here we interpret the concept of reference frame as a broad description of the type of the space, or the context that the data is associated with. It relates to whether the space itself is fixed or moving (global or local reference frame), or the identity of the object it moves with. For example, an anatomical reference frame is fixed to the body and moves when the body itself moves through space. |
| 26 | +# Definitions |
35 | 27 |
|
36 | | -**Coordinate system** |
37 | | -Examples : Cartesian coordinates, polar coordinates, and spherical coordinates. |
38 | | -A coordinate system is fully described by (1) the origin, (2) the interpretation of the axes, and (3) the units. |
| 28 | +## Space |
| 29 | +In BIDS terms, **space** is defined as an artificial frame of reference used to describe different anatomies in a unifying manner (see Appendix VIII). Data collected in studies of physical or virtual motion typically have a reference frame anchored to the physical lab space or the virtual environment. |
39 | 30 |
|
40 | | -[in BIDS terms] A coordinate system in BIDS comprises of the information about (1) the origin relative to which the coordinate is expressed, (2) the interpretation of the three axes, and (3) the units in which the numbers are expressed |
| 31 | +## Reference Frame |
| 32 | +A **reference frame** is an abstract coordinate system with a specified origin, orientation, and scale, defined by a set of reference points (Kovalevsky & Mueller, 1989). It broadly describes the type of space or context associated with the data, whether the space is fixed or moving (global or local reference frame), or the identity of the object it moves with. For instance, an anatomical reference frame is fixed to the body and moves as the body moves through space. |
41 | 33 |
|
42 | | -Coordinate systems |
43 | | -Motion data can be expressed within different coordinate systems associated with different frames of reference. The type of reference frame that requires the most detailed description is the anatomical, body-based reference frames. |
| 34 | +## Coordinate System |
| 35 | +A **coordinate system** is fully described by (1) the origin, (2) the interpretation of the axes, and (3) the units. In BIDS terms, a coordinate system comprises information about (1) the origin relative to which the coordinate is expressed, (2) the interpretation of the three axes, and (3) the units in which the numbers are expressed. |
44 | 36 |
|
45 | | -Reference frames can have a hierarchical structure where one reference frame is nested within another. For example, the position of the torso can be expressed in a coordinate system within the room reference frame, the arm position relative to the torso, and the hand position relative to the arm. In biomechanics … |
| 37 | +## Hierarchical Structure of Reference Frames |
| 38 | +Reference frames can have a hierarchical structure, where one reference frame is nested within another. For example, the position of the torso can be expressed within a room reference frame, the arm position relative to the torso, and the hand position relative to the arm. The reference frame at the top of this hierarchy is the **global reference frame**, associated with the space through which the entire body moves. This representation is useful in scenarios where the location of the person in space is relevant rather than their posture or limb motion. |
46 | 39 |
|
47 | | -We refer to the reference frame that is at the top level of this hierarchy as the “global” reference frame. The global reference frame can be associated with the space through which the entire body moves, allowing the location of the person to be expressed as a position within that space. This representation is useful in scenarios where the location of the person in space is relevant rather than their posture or limb motion. For example, in spatial cognition research … |
| 40 | +# Unified Placement Scheme |
48 | 41 |
|
49 | | -## Anatomical coordinate system for each rigid body part (Julius) |
50 | | -Axis per body part are defined in the table, with position of fiducials and span from start (-50) to end (50). A dataset with both IMU and optical markers is used as an example. |
| 42 | +## Anatomical Coordinate System for Rigid Body Parts |
| 43 | +Axis definitions per body part are provided in the anatomical **[landmark table](./anatomical_table.qmd)**. The table includes the name of the body part, the axis, and the direction of the axis, defined using anatomical landmarks, with axis limits ranging from 0 to 100% of the distance between the landmarks. |
51 | 44 |
|
52 | | -## Unified placement scheme |
53 | | -The most common types of motion data are acceleration, angular acceleration, joint angles, magnetic field strength, orientation and position. Each type has different requirements for sensor placement. We want to present a way of encoding the marker location and orientation so that it can be unified across different data sets. … |
| 45 | +## Principles of Sensor Placement Annotation |
| 46 | +We propose a **unified placement scheme** for sensors based on anatomical landmarks and the axes defined in the anatomical landmark table. The scheme follows these principles: |
| 47 | + |
| 48 | + 1. Sensor placement should be reproducible by a human with defined precision. |
| 49 | + 2. Placement in each dimension should be related to the anatomical landmarks of the relevant body part. |
| 50 | + 3. Landmarks define the origin, direction, and limits of the axes. |
| 51 | + 4. Sensor locations should be reported as a ratio of the limits of each axis for each body part. |
| 52 | + 5. Placement precision depends on the precision of landmarks, axes, and the measurement method. |
| 53 | + |
| 54 | +## Placement Precision |
| 55 | +The precision of sensor placement is related to the precision of landmark definitions, axis orthogonality, and measurement methods. We propose the following precision levels: |
| 56 | + |
| 57 | +1. **Visual Inspection**: Placement defined by visual inspection of body parts and landmarks, limited by human eye resolution and alignment ability. Estimated precision: ~10% of the distance between landmarks. |
| 58 | +2. **Tape Measure**: Placement defined by measuring distances between landmarks and placing the sensor at a specific ratio. Limited by tape measure resolution and alignment ability. Estimated precision: ~5% of the distance between landmarks. |
| 59 | +3. **3D Scanning**: Placement defined by 3D scanning body parts and placing the sensor at a specific ratio. Limited by 3D scanner resolution and alignment ability. Estimated precision: ~1% of the distance between landmarks. |
| 60 | + |
| 61 | +**Sensor placement precision** should always be reported in the dataset metadata. |
| 62 | + |
| 63 | +## Sensor Placement Annotation |
| 64 | +Sensor placement should be annotated using a standardized format, including: |
| 65 | + |
| 66 | + 1. Body part name |
| 67 | + 2. Axis name |
| 68 | + 3. Axis direction |
| 69 | + 4. Axis limits |
| 70 | + 5. Sensor location as a ratio of the axis limits |
| 71 | + 6. Sensor placement precision |
| 72 | + |
| 73 | +### Hierchical Event Descriptors (HED) for Sensor Placement |
| 74 | +We *propose* using Hierarchical Event Descriptors (HED) to annotate sensor placement in a standardized format. A possible HED tag for sensor placement could be: |
| 75 | + |
| 76 | + (Body-part, (X-position/#, Y-position/#, Z-position/#), Precision) |
| 77 | + |
| 78 | +Where: |
| 79 | + |
| 80 | +- Body-part: The name of the body part where the sensor is placed. |
| 81 | +- X-position, Y-position, Z-position: The sensor's position along the X, Y, and Z axes, respectively, expressed as a ratio of the axis limits. |
| 82 | +- Precision: The precision level of sensor placement (e.g., Visual Inspection, Tape Measure, 3D Scanning). |
| 83 | + |
| 84 | +NOTE: This proopsal is still in discussion and we are open to feedback and suggestions. Importantly, we aim to establish a **HED partnered schemaw** to define (1) the anatomical landmarks, (2) axis limits, and (3) axis direction for each body part. |
0 commit comments