Loading [MathJax]/jax/output/SVG/jax.js
Research article Special Issues

Fate factors and emission flux estimates for emerging contaminants in surface waters

  • Pharmaceuticals, personal care products, hormones, and wastewater products are emerging environmental concerns for manifold reasons, including the potential of some compounds found in these products for endocrine disruption at a very low chronic exposure level. The environmental occurrences and sources of these contaminants in the water, soil, sediment and biota in European nations and the United States are well documented. This work reports a screening-level emission and fate assessment of thirty compounds, listed in the National Reconnaissance of the United States Geological Survey (USGS, 1999–2000) as the most frequently detected organic wastewater contaminants in U.S. streams and rivers. Estimations of the surface water fate factors were based on Level II and Level III multimedia fugacity models for a 1000 km2 model environment, the size of a typical county in the eastern United States. The compounds are categorized into three groups based upon the sensitivity of their predicted surface water fate factors to uncertainties in their physicochemical property values and the landscape parameters. The environmental fate factors, mass distributions, and loss pathways of all of the compounds are strongly affected by their assumed modes of entry into the environment. It is observed that for thirteen of the thirty organic wastewater contaminants most commonly detected in surface waters, conventional treatment strategies may be ineffective for their removal from wastewater effluents. The surface water fate factors predicted by the fugacity models were used in conjunction with the surface water concentrations measured in the USGS reconnaissance to obtain emission flux estimates for the compounds into U.S. streams and rivers. These include estimated fluxes of 6.8 × 10−5 to 0.30 kg/h km2 for the biomarker coprostanol; 1.7 × 10−5 to 6.5 × 10−5 kg/h km2 for the insect repellent N,N-diethyltoluamide; and 4.3 × 10−6 to 3.1 × 10−5 kg/h km2 for the steroid estriol.

    Citation: Hoa T. Trinh, Peter Adriaens, Christian M. Lastoskie. Fate factors and emission flux estimates for emerging contaminants in surface waters[J]. AIMS Environmental Science, 2016, 3(1): 21-44. doi: 10.3934/environsci.2016.1.21

    Related Papers:

    [1] Hong Yuan, Jing Huang, Jin Li . Protein-ligand binding affinity prediction model based on graph attention network. Mathematical Biosciences and Engineering, 2021, 18(6): 9148-9162. doi: 10.3934/mbe.2021451
    [2] Saranya Muniyappan, Arockia Xavier Annie Rayan, Geetha Thekkumpurath Varrieth . DTiGNN: Learning drug-target embedding from a heterogeneous biological network based on a two-level attention-based graph neural network. Mathematical Biosciences and Engineering, 2023, 20(5): 9530-9571. doi: 10.3934/mbe.2023419
    [3] Dong Ma, Shuang Li, Zhihua Chen . Drug-target binding affinity prediction method based on a deep graph neural network. Mathematical Biosciences and Engineering, 2023, 20(1): 269-282. doi: 10.3934/mbe.2023012
    [4] Lei Chen, Kaiyu Chen, Bo Zhou . Inferring drug-disease associations by a deep analysis on drug and disease networks. Mathematical Biosciences and Engineering, 2023, 20(8): 14136-14157. doi: 10.3934/mbe.2023632
    [5] Ze-Yu Zhang, Zhu-Jun Mao, Ye-ping Ruan, Xin Zhang . Computational identification of Shenshao Ningxin Yin as an effective treatment for novel coronavirus infection (COVID-19) with myocarditis. Mathematical Biosciences and Engineering, 2022, 19(6): 5772-5792. doi: 10.3934/mbe.2022270
    [6] Deepalakshmi Sarkarai, Kalyani Desikan . QSPR/QSAR analysis of some eccentricity based topological descriptors of antiviral drugs used in COVID-19 treatment via $ \mathscr{D}\varepsilon $- polynomials. Mathematical Biosciences and Engineering, 2023, 20(9): 17272-17295. doi: 10.3934/mbe.2023769
    [7] Mingju Chen, Sihang Yi, Mei Yang, Zhiwen Yang, Xingyue Zhang . UNet segmentation network of COVID-19 CT images with multi-scale attention. Mathematical Biosciences and Engineering, 2023, 20(9): 16762-16785. doi: 10.3934/mbe.2023747
    [8] Huiqing Wang, Sen Zhao, Jing Zhao, Zhipeng Feng . A model for predicting drug-disease associations based on dense convolutional attention network. Mathematical Biosciences and Engineering, 2021, 18(6): 7419-7439. doi: 10.3934/mbe.2021367
    [9] Ali Al Khabyah, Muhammad Kamran Jamil, Ali N. A. Koam, Aisha Javed, Muhammad Azeem . Partition dimension of COVID antiviral drug structures. Mathematical Biosciences and Engineering, 2022, 19(10): 10078-10095. doi: 10.3934/mbe.2022471
    [10] Shizhen Huang, ShaoDong Zheng, Ruiqi Chen . Multi-source transfer learning with Graph Neural Network for excellent modelling the bioactivities of ligands targeting orphan G protein-coupled receptors. Mathematical Biosciences and Engineering, 2023, 20(2): 2588-2608. doi: 10.3934/mbe.2023121
  • Pharmaceuticals, personal care products, hormones, and wastewater products are emerging environmental concerns for manifold reasons, including the potential of some compounds found in these products for endocrine disruption at a very low chronic exposure level. The environmental occurrences and sources of these contaminants in the water, soil, sediment and biota in European nations and the United States are well documented. This work reports a screening-level emission and fate assessment of thirty compounds, listed in the National Reconnaissance of the United States Geological Survey (USGS, 1999–2000) as the most frequently detected organic wastewater contaminants in U.S. streams and rivers. Estimations of the surface water fate factors were based on Level II and Level III multimedia fugacity models for a 1000 km2 model environment, the size of a typical county in the eastern United States. The compounds are categorized into three groups based upon the sensitivity of their predicted surface water fate factors to uncertainties in their physicochemical property values and the landscape parameters. The environmental fate factors, mass distributions, and loss pathways of all of the compounds are strongly affected by their assumed modes of entry into the environment. It is observed that for thirteen of the thirty organic wastewater contaminants most commonly detected in surface waters, conventional treatment strategies may be ineffective for their removal from wastewater effluents. The surface water fate factors predicted by the fugacity models were used in conjunction with the surface water concentrations measured in the USGS reconnaissance to obtain emission flux estimates for the compounds into U.S. streams and rivers. These include estimated fluxes of 6.8 × 10−5 to 0.30 kg/h km2 for the biomarker coprostanol; 1.7 × 10−5 to 6.5 × 10−5 kg/h km2 for the insect repellent N,N-diethyltoluamide; and 4.3 × 10−6 to 3.1 × 10−5 kg/h km2 for the steroid estriol.


    Physical training, an integral aspect of modern life, encompasses various systematic and targeted exercise activities aimed at improving one's physical fitness, overall health and well-being [1]. The significance of physical training extends beyond individual benefits, as it not only helps to prevent various health issues such as obesity, cardiovascular diseases, and diabetes, but it also enhances mental health, boosts self-esteem and fosters social connections [2][5]. From a societal perspective, promoting physical training contributes to reduced healthcare costs, increased productivity and improved quality of life, making it a vital factor in the well-being of communities and nations alike [6]. As the demand for efficient and personalized physical training methods grows, there is an emerging need to leverage advanced technologies to optimize training outcomes and maximize the benefits derived from exercise.

    Physical training has evolved significantly. Figure 1 illustrates the technological advancements in fitness equipment over time. In the 18th and early 19th centuries, physical training primarily served military purposes, emphasizing the development of physical prowess for combat readiness [7]. The inception of the modern Olympic Games in the early 20th century marked a turning point, as various forms of exercise and fitness training emerged, including aerobic exercise, strength training, yoga and Pilates [8]. The advent of modern technology has further revolutionized physical training, employing tools such as a smart bracelet and intelligent body fat scales [9]. However, current physical training methodologies face notable challenges. Most devices on the market are limited to single-modality monitoring, providing an incomplete understanding of an individual's overall physical condition [10][12]. Addressing this gap, multi-modal sensor fusion systems that combine data from various sensors may serve as solutions to provide a comprehensive picture of an individual's fitness [13]. These systems have the potential to significantly enhance the effectiveness of physical training by offering more accurate, personalized feedback and recommendations based on a holistic understanding of the trainee's physical state.

    Figure 1.  The digitization process for fitness equipment.

    In light of these challenges, the primary objective of this research was to validate the effectiveness of the proposed multi-modal sensor fusion system in enhancing physical training performance. The developed system integrates a diverse range of sensors, including inertial measurement units (IMUs), heart rate sensors, electromyography (EMG) sensors, pressure sensors, GPS sensors, cameras, computer vision systems and environmental sensors. The research methodology here combines principles from geometry, biology and kinematics to develop a comprehensive understanding of an individual's physical state and performance during training. Figure 2 presents the design idea of the system and the core technologies involved.

    Figure 2.  The system design and core technologies.

    This study had the following goals. First, we introduce a multi-modal sensor fusion system that revolutionizes the approach to physical training. Our system has been designed, implemented and thoroughly evaluated to provide a distinctive and innovative approach, surpassing the capabilities of conventional single-modality systems. Integrating multiple sensors enables the acquisition of richer and more comprehensive data and insights, facilitating a deeper understanding of an individual's physical condition during training.

    Second, the system incorporates advanced sensor technology and employs sophisticated data processing algorithms. This integration ensures the system's ability to monitor various aspects of an individual's physical condition with high accuracy and reliability. Through the utilization of cutting-edge sensors and advanced algorithms, we enhance the precision of data collection and analysis, enabling more effective monitoring of the trainee's physical state.

    Last, the proposed multi-modal sensor fusion system showcases its potential to significantly enhance physical training outcomes. By leveraging the comprehensive data and insights obtained from multiple sensors, the system provides tailored feedback and recommendations. This personalized approach, rooted in a comprehensive understanding of the trainee's fitness level and performance, has the potential to optimize training routines and unlock improved outcomes in physical training.

    In essence, this paper delves into the utilization of multi-modal sensor fusion systems for physical training. The subsequent sections are organized as follows. Section 2 provides an extensive literature review, examining prior work in the field. Building upon this foundation, section 3 outlines the proposed methodology, elucidating the system design. In Section 4, the implementation of our multi-modal sensor fusion system is presented. Section 5 verifies the effectiveness of our approach through experimental demonstrations. Finally, Section 6 concludes the paper by summarizing the principal contributions and suggesting future research directions.

    Within the realm of physical training, various single-modality monitoring approaches have been utilized, such as force sensors, depth cameras, IMUs and heart rate monitors, to track different facets of human performance. However, these individual methods are not without their limitations, including decreased precision and unreliable data. To address these shortcomings, a wearable sensing system was developed to enhance rehabilitation for neurological patients by employing an electronic glove equipped with compact force sensors and a modified interface to measure fingertip force [14]. Additionally, a novel teaching methodology was introduced, employing a gamified learning process that promotes physical activity through the use of interactive balls in educational activities [15]. Notably, this system utilized a Kinect depth camera for image recognition, enabling the detection of ball hits on activity projections. In [16], to prevent cognitive and physical decline in older individuals with mild cognitive impairment, an exergame was designed. This exergame employed IMUs worn on the wrists and feet of the participants to track their movements as they navigated through the game environment. Within the field of rehabilitation training, Batalik et al. [17] explored the potential of telerehabilitation with wrist heart rate monitors as an alternative to traditional outpatient cardiac rehabilitation. The study demonstrated significant improvements in physical fitness for both regular outpatient training and interventional home-based telerehabilitation groups. Most recently, Bo and Sun [18] developed a heart-rate monitoring platform for college sports training, utilizing wireless networks. This platform featured multi-level link equalization, a transmission numerical model for heart-rate monitoring channels, data redundancy verification and a topology stage structure.

    The utilization of multi-modal sensor fusion systems has garnered considerable attention in diverse domains, including affective computing, robotics and automated driving [19], [20]. These systems combine data from multiple sensors to provide a more comprehensive understanding of various phenomena. A crucial aspect of our research is the application of such systems in biomedical engineering, especially in enhancing physical training. In the domain of wearables, a wearable multi-modal biosensing system has been developed to collect, synchronize, record and transmit data from diverse biosensors, such as systems of PPG, EEG, eye-gaze headset, body motion capture and GSR [21]. The performance of these sensors was evaluated by comparing them to standard research-grade biosensors. Comparatively, a study [22] yielded a scalable design model for artificial fingers in anthropomorphic robotic and prosthetic hands. This design integrated mechanical components, embedded electronics and a multi-modal sensor system. The fully parametric design enabled automated scaling, and interchangeable electronic modules facilitated mechanical adjustments. In the context of urban autonomous driving, Marco et al. [23] proposed a multi-modal sensor fusion scheme for accurately estimating 3D vehicle velocity, pitch and roll angles. The method simultaneously estimated gyroscope and accelerometer biases to enhance accuracy, proving itself to be effective during regular urban drives and collision avoidance maneuvers.

    A multitude of multi-modal sensor fusion systems have emerged, aiming to augment physical training outcomes and yield an extensive comprehension of human performance. These systems amalgamate diverse sensor types, such as GPS, heart-rate monitor, IMUs and motion-capture devices, to capture a broad range of physical activity aspects. Noteworthy contributions include Ma and Hu's development [24] of a home-oriented cyber-physical system that enhances motion coordination capabilities via physical training. Their system employs a thermal camera to record leg and foot motions, as well as insole pressure sensors for plantar pressure measurement. Innovative algorithms for leg skeleton extraction and motion signal auto-segmentation, recognition and analysis were devised.

    As early as 2016, Torreño et al. [25] conducted an investigation by utilizing GPS and heart-rate technology to explore the match running profiles, heart rates and performance efficiency indices of professional soccer players during official club-level games. Similarly, the utilization of a Polar GPS device alongside heart-rate sensors allowed for the determination of motion intensity in physical education settings [26]. To address the rehabilitation requirements, a novel virtual personal trainer system, tailored for rehabilitation purposes, was proposed. This system employed nine IMU sensors and a 3D camera for comprehensive full-body motion tracking [27]. The incorporation of a 3D camera mitigated IMU sensor output instability, counteracted gyroscopic drift and reduced electromagnetic interference. The resulting wireless full-body sensor array system provided a cost-effective alternative to on-site trainers. Stroke rehabilitation-focused physical training involved the development of a portable rehabilitation platform, encompassing both mental and physical training modalities [28]. This platform encompassed an EEG-based BCI system for mental training, a force sensor-embedded orthosis for elbow extension/flexion and an FES unit for hand extension. Later, a group [29] proposed a data fusion algorithm that combined skeletal data from three Kinect devices to overcome the limitations of single-device usage for human skeletons and motion tracking within sports medicine and rehabilitation. This approach offered comprehensive 3D spatial coverage of subjects and demonstrated an enhanced tracking accuracy improvement of 15.7%. As multi-modal sensor fusion technology advances, substantial strides are anticipated in the areas of data fusion algorithms and machine learning techniques, elevating the capabilities of these systems even further [30].

    Multi-modal sensor fusion systems play a pivotal role in advancing research and technology [31]. Researchers have focused on various aspects of these systems, such as PRF-PIR, unified camera tracking approaches and novel methods for human activity recognition. In the realm of PRF-PIR, Yuan et al. [32] proposed a passive, multi-modal sensor fusion system named PRF-PIR. This system consists of a software-defined radio device and a novel passive infrared (PIR) sensor system. Using a recurrent neural network (RNN) as the human identification and activity recognition model, the PRF-PIR system accurately and non-intrusively monitors human activity in indoor environments. The system's effectiveness was validated through data collection from 11 activities performed by 12 human subjects, employing explainable artificial intelligence methodologies. In [33], Oskiper et al. came up with a unified approach for a camera tracking system, employing an error-state Kalman filter algorithm. This approach utilizes relative and global measurements obtained from image-based motion estimation, landmark matching and radio frequency-ranging radios. The proposed approach demonstrated long-term stability and overall accuracy under vision-aided and vision-impaired conditions, as evidenced by the rendering of views from a 3D graphical model and actual video images. Regarding human activity recognition, an efficient method that fuses data from inertial sensors (IMUs), surface EMG, and visual depth sensors have been developed. This approach exhibits superior performance compared to single- or dual-sensor approaches, with high robustness against occlusions, data losses and other non-structured environmental events [34].

    The integration of artificial intelligence into sensor fusion technology, encompassing data fusion algorithms and machine learning methods, holds immense potential for advancements [35][37]. Xu and his colleagues [38] posited that a self-adaptive wavelet transform-based data fusion algorithm, suitable for both static and dynamic systems, can achieve an optimal estimation of a measurand with minimum mean square error in multisensor systems. Similarly, the GAPSOBP algorithm, which intelligently combines a BP neural network, genetic algorithm and particle swarm optimization algorithm, exhibits efficiency in reducing the volume of data transmitted to a base station or sink node within wireless sensor networks, thus conserving network energy [39]. In the field of physical training, the utilization of data fusion algorithms and machine learning methods has substantial value. They enable the development of multi-modal sensor fusion systems that yield richer data and insights than those of conventional single-modality systems [40]. Furthermore, the synergy between multi-modal sensor fusion systems and virtual reality (VR) presents an opportunity to enhance various applications. Through the integration of data from diverse sensors, multi-modal sensor fusion systems offer a comprehensive and precise understanding of the physical environment and human behavior. In contrast, VR creates immersive and interactive virtual environments that simulate real-world scenarios, allowing users to engage with virtual objects and surroundings. The combination of these technologies results in a robust and authentic training or simulation platform, providing users with an immersive, interactive experience that delivers accurate and comprehensive feedback [41][43].

    Table 1 provides a comprehensive overview of the diverse sensing modalities employed in mobile and wearable sensor-based multi-modal applications.

    Table 1.  Sensor modalities, applications and methods.
    Applications Refs. Year Multi-sensor Methods
    Neurologic Rehabilitation Trainings [14] 2018 FSR, Wearable+force+embedded sensors Rigid discs
    Electronic circuit
    Physical Activity [15] 2021 Kinect sensor
    ECG
    MOCAP
    TEL
    Heart-Rate Monitor [17] 2020 Wearable sensors Polar Flow web
    [18] 2023 Temperature+heart rate+acceleration+smoke sensors Wireless Network
    Bio signals [21] 2018 PPG
    EEG Body motion capture
    GSR
    Inertial sensors
    SSVEP
    Vehicle Motion State Estimation [23] 2020 ESC
    GNSS
    IMU+Series-grade chassis sensors
    Intrinsic Euler angles
    Immersion transformation
    Human Identification and Activity [32] 2022 PIR
    PRF-PIR
    IMU
    MI-PIR
    RGBD
    CNN
    NBC
    FMCW
    RNN
    Infrastructure-free Localization in Vision-impaired Environments [33] 2010 EO
    IMU
    RF
    HMD
    Kalman filter

     | Show Table
    DownLoad: CSV

    This section presents the proposed methodology, expounding on the intricate configuration and execution of our multi-modal sensor fusion system. In Figure 3, an overview of the system's use in physical training is depicted. In the following subsections, we will describe each module in detail.

    1. Skeleton tracker (Section 3.1), which integrates depth detection, bone tracking, face recognition and voice recognition using diverse cameras.

    2. Body angle measurement (Section 3.2), which utilizes specialized equipment to collect EMG signals from various arm segments, providing real-time display and raw data export.

    3. Finger motion monitor (Section 3.3), which captures real-time movement data from individual hand joints, utilizing non-sensing wear technology and facilitating data transfer.

    4. VR headset (Section 3.4), allowing users to immerse themselves in a VR environment with the freedom to navigate and observe in any direction.

    Figure 3.  Multi-modal sensor fusion system in physical training.

    The skeleton tracker's primary function in this research is to construct an accurate and dynamic model of the user's body motion during physical training. It uses multiple cameras and chips to detect depth, track bones and recognize faces and voices. Meanwhile, it allows for high-resolution output (1920 × 1080, 30 fps in color; 512 × 424, 30 fps in depth) with remarkable accuracy. Depth detection relies on the projection of infrared rays through the infrared camera to generate reflected light, which facilitates the determination of an object's position and the creation of a depth image based on spare flight time. Skeleton tracking was primarily designed to construct a bone map for up to six players and accurately track the corresponding bone nodes throughout the physical training session.

    The body angle measurement apparatus was designed to capture the subtleties of body movement, focusing primarily on the arm segment, by using EMG signals. It gives real-time data on muscle activity during training. The EMG signal collector primarily comprises a wireless collector, a wireless EMG sensor (one channel) and a wireless biaxial joint goniometer (two tracks). In our system, the Biometrics GZ12 EMG signal collector records electrical muscle activity in response to different physical actions. This information is then exported for in-depth analysis, allowing us to examine the efficiency of the training, identify potential strain or injury risks and suggest improvements in a training method.

    Accurate measurement and analysis of hand movements necessitates sophisticated algorithms and advanced sensors. The finger motion monitor, represented here by our data glove, is a critical component that captures nuanced data about hand joint movements during physical training. Equipped with 9-axis MEMS inertial sensors and a vibration feedback module, the glove captures and communicates detailed hand motion data. By employing inverse dynamics, the glove accurately reconstructs bone motion, enabling the faithful reproduction of natural movement in virtual environments. Furthermore, the palm of the glove incorporates a built-in vibration feedback module, which triggers vibration effects corresponding to different scenarios, thereby enhancing the immersive experience in computer designs. To ensure seamless communication, the gloves utilize 24-GHz wireless transmission, achieving a high frame rate of 120 Hz for one hand and 240 Hz for both hands. This high frame rate, coupled with a low delay transmission effect within 10 ms, ensures smooth and responsive interactions. In this research context, we used it to understand the involvement of hand and finger movements during the exercise process. These data are particularly valuable for the evaluation of the precision and accuracy of hand movements during training, allowing us to devise more effective training regimens or corrective measures as required.

    The Pico 4 VR headset is our system's main user interface. The headset provides a fully immersive VR environment, allowing the user to perform physical training actions in a controlled and responsive setting. The Pico 4 is a standalone headset with six-degree-of-freedom inside-out tracking, the weighing 450 g. It prioritizes comfort with an adjustable head strap and face cushion, offering a smooth and immersive experience with a 101-degree FOV and 75 Hz refresh rate. Powered by a Qualcomm Snapdragon 845 processor, it has 4 GB of RAM and 128 GB of storage. The Pico 4 includes built-in stereo speakers, a 3.5 mm audio jack, two six-degree-of-freedom controllers for intuitive interaction, WiFi 6, Bluetooth 5.0 and hand-tracking capabilities, and it runs on the proprietary PICOS. In summary, the Pico 4 VR headset offers immersive experiences without external sensors or cameras. It is well suited for enhancing physical training and facilitating engaging virtual environments.

    Our hand-tracking system represents a significant advancement in the field of the precise tracking of complex hand movements (Figure 4). Unlike traditional systems that are confined to a fixed workspace, ours allows users to move freely while wearing the equipment, thus introducing a more natural interactive user experience.

    Figure 4.  The system overview. On the left, a detailed schematic showcases the specific sensors and equipment related to hand tracking, meticulously designed for accurate data acquisition of hand movements. On the right, an illustrative depiction of a user engaged in physical training, demonstrating the freedom and flexibility afforded by the wearable equipment in real-world applications.

    The system comprises two individual subsystems for each hand, each utilizing a glove equipped with a constellation of sensors. Each glove houses six nine-axis IMUs and five flex sensors, providing a total of 12 IMUs and 10 flex sensors for both hands. The IMUs incorporate a gyroscope, accelerometer and magnetometer with measurement ranges of ±2000 dps, ±16 g and ±8.1 Ga, respectively. These sensors work in unison to provide accurate tracking of the hand's spatial orientation and acceleration, as well as the earth's magnetic field, to determine directional heading.

    The flex sensors, strategically positioned, monitor the bending of the fingers, thereby capturing even the subtlest of hand movements. These sensors are calibrated to deliver a static accuracy of 0.2° RMS for roll/pitch and a dynamic accuracy of 1.0° RMS. The angular measurement resolution stands at an impressive 0.02°.

    The gloves employ dual-band (2.4 GHz/5.8 GHz) wireless communication technology, achieving a synchronization accuracy of 10 µs, an industry-leading figure. Furthermore, the system's data communication supports manual and automatic channel switching, thereby mitigating the influence of potential wireless interference in the surroundings.

    The system's hardware components are engineered for compatibility with 100BASE-T interfaces and the IEEE 802.3 af/at standards for Power over Ethernet, making it a versatile solution for various user requirements.

    Further enhancing the tracking capabilities of our system is a Microsoft Kinect V2, incorporated for the capture of body skeleton data and recognition of overall body movements. The amalgamation of hand movement data from our custom gloves and body movement data from Kinect V2 results in a comprehensive and seamless body tracking experience for the user.

    Augmenting the sensor suite of each glove is a wireless EMG data collector from Biometrics. This collector gathers critical data pertaining to forearm muscle activity, providing another data source for enhanced movement recognition. Our approach of extensive sensor fusion, utilizing data from multiple sensor types, allows the system to achieve an unprecedented level of precision in hand movement recognition.

    Powering our tracking system is the Robot Operating System (ROS), which was run on an Ubuntu 20 machine. This computer communicates with the gloves via a high-speed wireless router, ensuring minimal latency and maximum throughput in data transmission. Furthermore, MATLAB was employed to process the collected data, which were subject to interaction in real-time within a VR scene in the PICO 4 system. This VR scene, developed in Unity, maintains a real-time communication link with MATLAB, thus enabling fluid interaction between the user and the virtual environment.

    Given its precision, our system is optimal for intricate physical training regimens, including those that emphasize precise hand movements such as rehabilitation exercises. It is invaluable for trainers seeking detailed feedback on a trainee's form, grip and finger positioning.

    Table 2.  Device and software parameters.
    Device/Software Version/Parameters
    Inertial Sensor Gyroscope: ±2000 dps, Accelerometer: ±16 g, Magnetometer: ±8.1 Ga
    Flex Sensor 10 (5 per hand)
    Kinect V2 Body skeleton tracking
    EMG Data Collector Biometrics
    Communication Technology Dual-band (2.4 GHz/5.8 GHz), Sync accuracy: 10 µs
    ROS Melodic Morenia (1.14.11)
    Ubuntu 20.04 LTS
    MATLAB R2023a
    PICO 4 VR System Latest firmware as of June 2023
    Unity 2023.1.2

     | Show Table
    DownLoad: CSV

    In conclusion, our system stands out due to its detailed hand tracking, full-body motion capture and virtual environment integration. It provides a flexible solution that fits various use cases.

    First, we capture the basic posture of the human body through the skeleton tracking feature of the Kinect V2. For each joint, Kinect V2 provides its position in 3D space. To compute the angle between two joints, we can use the dot product of vectors formed by the joints, which is calculated as follows:

    θ=cos1(ab|a||b|)

    where a and b are the vectors formed by the joints, and θ is the angle between these vectors.

    Kinect provides the 3D positions of various joints in the human body. Each joint Ji is defined in a 3D Cartesian space, and thus has a position vector in this space denoted by Pi=(Xi,Yi,Zi). The data acquisition rate of Kinect V2 is 30 frames per second.

    Now, consider three joints J1, J2 and J3 in sequence, where J2 is the joint of interest and J1 and J3 are adjacent joints. The vectors formed from J2 to J1 and J2 to J3 are respectively denoted by A and B, which can be calculated as follows:

    A=P1P2=(X1X2,Y1Y2,Z1Z2)

    B=P3P2=(X3X2,Y3Y2,Z3Z2)

    The angle θ between these two vectors, which corresponds to the angle of the joint J2, can be calculated by using the dot product of A and B:

    θ=cos1(ABAB)=cos1(AxBx+AyBy+AzBz(A2x+A2y+A2z)(B2x+B2y+B2z))

    where ||A|| and ||B|| are the magnitudes of vectors A and B respectively.

    The glove is furnished with inertial sensors and the flex sensors, capturing the movements of fingers. The fusion of data from these sensors allows us to achieve accurate finger movement estimation. The IMU sensors operate at a sampling rate of 200 Hz, and flex sensors operate at a rate of 100 Hz. All sensors are synchronized via a central microcontroller that timestamps the data before passing it on for processing.

    We leverage an adaptive neural network algorithm to process the sensor data and predict hand movements. It helps to overcome the limitations of traditional methods, like sensitivity to the initial conditions and the constraints of linear systems, thus providing robust and precise hand tracking.

    The flex sensors in the glove measure the amount of bend in each finger joint. For each joint Fi we have a corresponding bending measurement i.

    The IMU provides data for the orientation of the glove (and, consequently, the hand and the fingers) in the form of quaternion values Q=(qw,qx,qy,qz). The Euler angles can be calculated from the quaternion as follows:

    roll=atan2(2(qyqz+qwqx),q2wq2xq2y+q2z)pitch=atan2(2(qxqz+qwqy),q2w+q2xq2yq2z)yaw=asin(2(qxqyqwqz))

    For precise measurement of the finger joint angle θFi, sensor fusion is implemented by using the Kalman filter. The state estimate ˆxk|k and estimate uncertainty Pk|k are calculated as follows:

    ˆxk|k=ˆxk|k1+Kk(zkHˆxk|k1)Pk|k=(IKkH)Pk|k1

    Where Kk is the Kalman gain, zk is the actual sensor reading, H is the transformation matrix, I is the identity matrix, ˆxk|k1 is the a priori estimate, and Pk|k1 is the a priori estimate uncertainty.

    The bending measurement i and IMU orientation information are used as inputs to the Kalman filter to obtain the accurate estimate of the joint angle θFi.

    In the pursuit of optimal sensor fusion, algorithm selection is crucial. For this research, the gated recurrent unit-RNN has been employed to predict the finger joint angles based on the sequential sensor readings. The primary reason for adopting the GRU-RNN model over other deep learning or traditional algorithms lies in its inherent capabilities that are tailored for sequential data. The GRU makes use of two gating mechanisms, namely the reset gate and the update gate, which control the flow of information inside of the unit. The sensors' raw data feed into the GRU model at their respective sampling rates, with missing data points interpolated to ensure continuous input.

    For a given sequence of sensor readings {s1,s2,,st}, at each time step t, the update gate zt and reset gate rt are computed as follows:

    zt=σ(Wz[st1,st]+bz)rt=σ(Wr[st1,st]+br)

    where Wz, Wr are weight matrices, bz, br are bias terms and σ is the sigmoid function.

    The candidate activation ˜ht is then calculated as

    ˜ht=tanh(W[st1rt,st]+b)

    Finally, the hidden state ht is updated as follows:

    ht=(1zt)ht1+zt˜ht

    This hidden state ht represents the predicted joint angle at time step t.

    A key property of GRUs (and RNNs in general) is that they have the ability to handle sequences of data and model temporal dependencies, which make them particularly suited for our application. However, the learning of GRUs involves the challenging task of learning long-term dependencies due to the vanishing gradients problem and requires careful initialization and potentially regularization methods to ensure stable convergence.

    To interact with the PICO 4 VR system, we used Unity3D to develop a VR scene and establish communication with MATLAB. The human movement data processed by MATLAB are sent to Unity in real-time, which then updates the VR scene accordingly. This provides the user with an immersive experience as their physical movements are accurately reflected in the virtual environment, which is especially beneficial for physical exercise applications.

    For future work, we intend to enhance the system by incorporating more advanced sensor fusion algorithms and neural networks, further improving the robustness and accuracy of human movement tracking in VR-based human-computer interaction. We also plan to expand its application areas to other domains, such as medical rehabilitation and professional sports training.

    In order to evaluate the performance of the proposed GRU method and compare it with other algorithms, a comprehensive set of experiments were conducted. A dataset comprising time-series data from the motion-tracking system was prepared. This dataset consisted of sensor readings from the hand-tracking gloves, Kinect V2 and EMG data collectors.

    Each time step in the data corresponds to a specific joint angle configuration for the human subject. The goal of the algorithms is to predict the joint angles at the next time step given the current and previous sensor readings. The ground truth for these predictions comes from the actual joint angles measured by the motion-tracking system.

    The experiment involved two main steps: feature extraction and model training. The data from the Kinect and glove sensors were pre-processed and segmented into windows of 2 second with 50% overlap. For each window, a set of features were extracted, which included statistical features (mean, variance, skewness and kurtosis), frequency-domain features (FFT coefficients) and time-domain features (zero-crossings, peak values). These features were then used as input to the machine learning models.

    The models were trained on 80% of the data, with the remaining 20% used for testing. The performance of each model was evaluated by using five metrics: accuracy, precision, recall, F1 score and average prediction time. The accuracy was calculated as the proportion of correctly predicted samples. The precision was calculated as the ratio of correctly predicted positive observations to the total predicted positive observations. The recall was calculated as the ratio of correctly predicted positive observations to all observations in the actual class. The F1 score was calculated as 2(Recall * Precision) / (Recall + Precision). The average prediction time was measured by averaging the time taken by the model to predict the joint angles for all samples in the test set.

    The accuracy, precision, recall, and F1 score are popular metrics used in machine learning, and they were computed as follows:

    Accuracy=Number of Correct PredictionsTotal Number of Predictions

    Precision=True PositivesTrue Positives+False Positives

    Recall=True PositivesTrue Positives+False Negatives

    F1 Score=2×Precision×RecallPrecision+Recall

    The comparison of the GRU-RNN algorithm with other machine learning methods is summed up in Table 3. It is clear from the results that the GRU-RNN algorithm outperforms the other algorithms in terms of accuracy, precision, recall and F1 score. This implies that the GRU-RNN algorithm is more successful in terms of correctly predicting the joint angles from the sensor data. Furthermore, the GRU-RNN algorithm also demonstrates superiority in terms of the average prediction time, further enhancing its suitability for real-time applications. The superior performance of the GRU-RNN algorithm can be attributed to its ability to effectively capture the temporal dependencies in the sensor data, which is crucial for accurate joint angle prediction. This makes the GRU-RNN algorithm a robust and efficient choice for the motion-tracking system.

    Table 3.  Comparison of different algorithms on joint angle prediction task.
    Algorithm Accuracy Precision Recall F1 Score Average Prediction Time (ms)
    GRU-RNN 0.97 0.98 0.96 0.97 4.5
    Feedforward Neural Network 0.85 0.87 0.83 0.85 5.2
    Support Vector Machine 0.81 0.83 0.79 0.81 6.3
    Random Forest 0.76 0.77 0.75 0.76 7.8

     | Show Table
    DownLoad: CSV

    As observed in Table 3, our proposed method outperforms the other techniques in terms of accuracy, precision, recall and F1 score. Furthermore, it provides the best real-time processing capabilities, with a shorter computational time than other methods. Thus, our system's application of networks demonstrates superior performance, further bolstering its potential for precise, robust and real-time hand motion tracking in dynamic environments.

    The bar chart provided in Figure 5 presents a comparison of the GRU-RNN algorithm with other common machine learning algorithms in terms of accuracy.

    Figure 5.  Comparison of the accuracy of different algorithms.

    Evidently, the GRU-RNN outperforms the other algorithms in terms of accuracy. It is important to note that the superiority of the GRU-RNN is attributable to its inherent ability to capture temporal dependencies in sequential data. In the case of joint angle prediction, the current joint angles are closely related to the previous states due to the physical constraints of human body motion. This context is something that the GRU-RNN captures well.

    The feedforward neural network, support vector machine (SVM) and random forest methods, though being competent machine learning models, are unable to handle sequential data as effectively. This is due to their inability to maintain a “memory” of previous inputs in the sequence.

    The feedforward neural network, in this case, fails to consider the temporal sequence of data. On the other hand,the SVM and random forest work on the assumption that the data samples are independent, which is not the case with sequential sensor data. Hence, their performance falls short as compared to the GRU-RNN.

    In conclusion, for tasks that involve sequential data, such as the joint angle prediction in our motion-capture system, the use of the GRU-RNN provides significant advantages over other methods in terms of accuracy. This finding is vital for the enhancement of the effectiveness and reliability of our system in order to deliver a more immersive VR experience for users.

    We have successfully developed a multi-modal sensor fusion system that can enhance physical training by delivering detailed, real-time feedback on the user's performance. Through the use of diverse sensor sources including IMUs, flex sensors, EMG sensors and Microsoft's Kinect V2, our system can capture comprehensive body and hand motion data. Our results demonstrate that the GRU-RNN algorithm, as employed in our system, outperforms traditional machine learning algorithms in terms of accuracy, precision, recall and F1 score. Furthermore, the integration of our system with the PICO 4 VR environment offers an immersive and interactive experience, fostering user engagement in physical training. Unlike many traditional systems, ours overcomes the constraint of a fixed workspace, allowing users the flexibility to train in a free-form environment. Future work will focus on further improving the accuracy of motion estimation and enhancing the immersive experience of the VR environment.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    [1] Daughton CG, Ternes TA (1999) Pharmaceuticals and personal care products in the environment: agents of subtle change? Environ Health Perspect Supp 107: 907-938. doi: 10.1289/ehp.99107s6907
    [2] Sedlak DL, Gray JL, Pinkston KE (2000) Understanding microcontaminants in recycled water. Environ Sci Technol 34: 508A-515A. doi: 10.1021/es003513e
    [3] Asano T, Levine AD (2004) Recovering sustainable waste from wastewater. Environ Sci Technol 38: 201A-208A. doi: 10.1021/es040504n
    [4] Halling-Sorensen B, Nielsen N, Lansky PF, et al. (1998) Occurrence, fate and effects of pharmaceutical substances in the environment- A review. Chemosphere 36: 357-393. doi: 10.1016/S0045-6535(97)00354-8
    [5] Heberer T (2002) Occurrence, fate, and removal of pharmaceutical residues in the aquatic environment: a review of recent research data. Toxicol Lett 131: 5-17. doi: 10.1016/S0378-4274(02)00041-3
    [6] Kolpin DW, Furlong ET, Meyer MT, et al. (2002) Pharmaceuticals, hormones, and other organic wastewater contaminants in U.S. streams, 1999-2000: A national reconnaissance. Environ Sci Technol 36: 1202-1211. doi: 10.1021/es011055j
    [7] Panter GH, Thompson RS, Sumpter JP (2000) Intermittent exposure of fish to estradiol. Environ Sci Technol 34: 2756-2760. doi: 10.1021/es991117u
    [8] Purdom CE, Hardiman PA, Bye VJ, et al. (1994) Estrogenic effects of effluents from sewage treatment works. Chem Ecol 8: 275-285. doi: 10.1080/02757549408038554
    [9] Davis DL, Bradlow HL (1995) Can environmental estrogens cause breast cancer? Sci Am 273: 144-149. doi: 10.1038/scientificamerican0995-144
    [10] D'Costa VM, McGrann KM, Hughes DW, et al. (2006) Sampling the antibiotic resistome. Science 311: 374-377. doi: 10.1126/science.1120800
    [11] Ternes TA, Joss A, Siegrist H (2004) Scrutinizing pharmaceuticals and personal care products in wastewater treatment. Environ Sci Technol 38: 392A-399A. doi: 10.1021/es040639t
    [12] Ternes TA, Herrmann N, Bonerz M, et al. (2004) A rapid method to measure the solid–water distribution coefficient Kd. for pharmaceuticals and musk fragrances in sewage sludge. Water Res 38: 4075-4084. doi: 10.1016/j.watres.2004.07.015
    [13] Rogers HR (1996) Sources, behaviour and fate of organic contaminants during sewage treatment and in sewage sludges. Sci Total Environ 185: 3-26. doi: 10.1016/0048-9697(96)05039-5
    [14] Golet EM, Xifra I, Siegrist H, et al. (2003) Environmental exposure assessment of fluoroquinolone antibacterial agents from sewage to soil. Environ Sci Technol 37: 3243-3249. doi: 10.1021/es0264448
    [15] Buser HR, Poiger T, Mueller MD (1998) Occurrence and fate of the pharmaceutical drug diclofenac in surface waters: rapid photodegradation in a lake. Environ Sci Technol 32: 3449-3456. doi: 10.1021/es980301x
    [16] Ternes TA (1998) Occurrence of drugs in German sewage treatment plants and rivers. Water Res 32: 3245-3260. doi: 10.1016/S0043-1354(98)00099-2
    [17] Huber MM, Korhonen S, Ternes TA, et al. (2005) Oxidation of pharmaceuticals during water treatment with chlorine dioxide. Water Res 39: 3607-3617. doi: 10.1016/j.watres.2005.05.040
    [18]   Joss A, Keller E, Alder AC, et al. Removal of pharmaceuticals and fragrances in biological wastewater treatment. Water Res 39: 3139-3152. doi: 10.1016/j.watres.2005.05.031
    [19] Mackay D (2001) Multimedia environmental models — the fugacity approach. 2nd ed.: Lewis Publishers, Chelsea, MI.
    [20] Cowan CE, Mackay D, Feijtel TCJ, et al. (1995) The multi-media fate model: a vital tool for predicting the fate of chemicals.: SETAC Press, Penascola, FL.
    [21] De Haes HAU, Jolliet O, Norris G, et al. (2002) Background, aims and scope. Int J LCA 4: 192-195.
    [22] De Haes HAU, Heijungs R, Suh S, et al. (2004) Three strategies to overcome the limitations of Life-Cycle Assessment. J Ind Ecol 8: 19-32.
    [23] Bare JC, Norris GA, Pennington DW, et al. (2002) TRACI: The tool for the reduction and assessment of chemical and other environmental impacts. J Ind Ecol 6: 49-78. doi: 10.1162/108819802766269539
    [24] Mackay D, Paterson S (1991) Evaluating the multimedia fate of organic chemicals: a level III fugacity model. Environ Sci Technol 25: 427-432. doi: 10.1021/es00015a008
    [25] Mackay D, Di Guardo A, Paterson S, et al. (1996) Evaluating the environmental fate of a variety of types of chemicals using the EQC model. Environ Toxicol Chem 15: 1627-1637. doi: 10.1002/etc.5620150929
    [26] Di Guardo A, Calamari D, Benfenati E, et al. (2001) In Pharmaceuticals in the environment: sources, fate, effects and risks. Kummerer, K., Ed.: Springer-Verlag, Berlin, 91-102.
    [27] Khan SJ, Ongerth JE (2004) Modelling of pharmaceutical residues in Australian sewage by quantities of use and fugacity calculations. Chemosphere 54: 355-367. doi: 10.1016/j.chemosphere.2003.07.001
    [28] Zukowska B, Breivik K, Wania F (2006) Evaluating the environmental fate of pharmaceuticals using a level III model based on poly-parameter linear free energy relationships. Sci Tot Environ 359: 177-187. doi: 10.1016/j.scitotenv.2005.05.033
    [29] MacLeod M, Fraser AJ, Mackay D (2002) Evaluating and expressing the propagation of uncertainty in chemical fate and bioaccumulation models. Environ Toxicol Chem 21: 700-709. doi: 10.1002/etc.5620210403
    [30] Hertwich EG, McKone TE, Pease WS (1999) Parameter uncertainty and variability in evaluative fate and exposure models. Risk Anal 19: 1193-1204.
    [31] Hollander A, Pistocchi A, Huijbregts MA, et al. (2009) Substance or space? The relative importance of substance properties and environmental characteristics in modeling the fate of chemicals in Europe. Environ Toxicol Chem 28: 44-51.
    [32] EPA Office of Pollution Prevention Toxics and Syracuse Research Corporation, Predictive Models and Tools for Assessing Chemicals under the Toxic Substances Control Act (TSCA). 2015. Available from: http://www.epa.gov/opptintr/exposure/pubs/episuitedl.htm.
    [33] Lyman WJ, Rosenblatt DH (1982) Handbook of chemical property estimation methods: Environmental behavior of organic compounds: McGraw-Hill.
    [34] Schwarzenbach RP, Gschwend PM, Imboden DM (2003) Environmental organic chemistry, 2nd ed.: Wiley-Interscience.
    [35] Hansch C, Leo A, Hoekman D (1995) Exploring QSAR - Hydrophobic, electronic, and steric constants: American Chemical Society, Washington, DC.
    [36] Lintelman J, Katayama A, Kurihara N, et al. (2003) Endocrine disruptors in the environment IUPAC Technical Report. Pure Appl Chem 75: 631-691.
    [37] Platts JA, Abraham MH, Butina D, et al. (2000) Estimation of molecular linear free energy relationship descriptors by a group contribution approach. 2. Prediction of partition coefficients. J Chem Inf Comput Sci 40: 71-80. doi: 10.1021/ci990427t
    [38] Zhao YH, Abraham MH, Le J, et al. (2002) Rate-limited steps of human oral absorption and QSAR studies. Pharm Res 19: 1446-1457. doi: 10.1023/A:1020444330011
    [39] Goss KU, Schwarzenbach RP (2001) Linear free energy relationships used to evaluate equilibrium partitioning of organic compounds. Environ Sci Technol 35: 1-9. doi: 10.1021/es000996d
    [40] Westerhoff P, Yoon Y, Snyder S, et al. (2005) Fate of endocrine-disruptor, pharmaceutical, and personal care product chemicals during simulated drinking water treatment processes. Environ Sci Technol 39: 6649-6663. doi: 10.1021/es0484799
    [41] Loffler D, Rombke J, Meller M, et al. Environmental fate of pharmaceuticals in water/sediment systems. Environ Sci Technol 39: 5209-5218. doi: 10.1021/es0484146
    [42] United States Census Bureau, State Area Measurements and Internal Point Coordinates. 2015. Available from: http://www.census.gov/geo/reference/state-area.html.
    [43] MacLeod M, Mackay D (1999) An assessment of the environmental fate and exposure of benzene and the chlorobenzenes in Canada. Chemosphere 38: 1777-1796. doi: 10.1016/S0045-6535(98)00394-4
    [44] USGS Toxic Substances Hydrology Program, Emerging Contaminants In the Environment. Available from: http://toxics.usgs.gov/regional/emc.html.
    [45] Arnot JA, Mackay D, Webster E, et al. (2006) Screening level risk assessment model for chemical fate and effects in the environment. Environ Sci Technol 40: 2316-2323. doi: 10.1021/es0514085
    [46] Gouin T, Mackay D, Webster E, et al. (2000) Screening chemicals for persistence in the environment. Environ Sci Technol 34: 881-884. doi: 10.1021/es991011z
    [47] Mikes O, Trapp S (2010) Acute toxicity of the dissociating veterinary antibiotics trimethoprim to willow trees at varying pH. Bull Environ Contam Toxicol 85: 556-561. doi: 10.1007/s00128-010-0150-6
    [48] LeBlanc LA, Latimer JS, Ellis JT, et al. (1992) The geochemistry of coprostanol in waters and surface sediments from Narragansett Bay. Est Coast Shelf Sci 34: 439-458. doi: 10.1016/S0272-7714(05)80116-7
    [49] Zhang QQ, Ying GG, Chen ZF, et al. (2015) Basin-scale emission and multimedia fate of triclosan in whole China. Environ Sci Pollut Res 22: 10130-10143. doi: 10.1007/s11356-015-4218-z
    [50] Zhang QQ, Ying GG, Pan CG, et al. (2015) Comprehensive evaluation of antibiotics emission and fate in the river basins of China: source analysis, multimedia modeling, and linkage to bacterial resistance. Environ Sci Technol 49: 6772-6782. doi: 10.1021/acs.est.5b00729
    [51] Zhang QQ, Zhao JL, Ying GG, et al. (2014) Emission estimation and multimedia fate modeling of seven steroids at the river basin scale in China. Environ Sci Technol 48: 7982-7992. doi: 10.1021/es501226h
    [52] Trapp S, Franco A, Mackay D (2010) Activity-based concept for transport and partitioning of ionizing organics. Environ Sci Technol 44: 6123-6129. doi: 10.1021/es100509x
  • This article has been cited by:

    1. Peng Li, Mosharaf Md Parvej, Chenghao Zhang, Shufang Guo, Jing Zhang, Advances in the Development of Representation Learning and Its Innovations against COVID-19, 2023, 3, 2673-8112, 1389, 10.3390/covid3090096
    2. Seema Yadav, Abhishek Singh, Rishika Singhal, Jagat Pal Yadav, Revolutionizing drug discovery: The impact of artificial intelligence on advancements in pharmacology and the pharmaceutical industry, 2024, 2, 2949866X, 367, 10.1016/j.ipha.2024.02.009
  • Reader Comments
  • © 2016 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(7898) PDF downloads(1475) Cited by(7)

Figures and Tables

Figures(7)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog