Summary of P2lhap:wearable Sensor-based Human Activity Recognition, Segmentation and Forecast Through Patch-to-label Seq2seq Transformer, by Shuangjian Li et al.
P2LHAP:Wearable sensor-based human activity recognition, segmentation and forecast through Patch-to-Label Seq2Seq Transformer
by Shuangjian Li, Tao Zhu, Mingxing Nie, Huansheng Ning, Zhenyu Liu, Liming Chen
First submitted to arxiv on: 13 Mar 2024
Categories
- Main: Computer Vision and Pattern Recognition (cs.CV)
- Secondary: Artificial Intelligence (cs.AI)
GrooveSquid.com Paper Summaries
GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!
Summary difficulty | Written by | Summary |
---|---|---|
High | Paper authors | High Difficulty Summary Read the original abstract here |
Medium | GrooveSquid.com (original content) | Medium Difficulty Summary The novel Patch-to-Label Seq2Seq framework, P2LHAP, is introduced to simultaneously segment, recognize, and forecast human activities from sensor data. This framework divides sensor data streams into patches, serving as input tokens, and outputs a sequence of patch-level activity labels including predicted future activities. A unique smoothing technique based on surrounding patch labels identifies activity boundaries accurately. The framework learns patch-level representation using sensor signal channel-independent Transformer encoders and decoders. All channels share embedding and Transformer weights across all sequences. P2LHAP significantly outperforms the state-of-the-art in all three tasks, demonstrating its effectiveness and potential for real-world applications. |
Low | GrooveSquid.com (original content) | Low Difficulty Summary P2LHAP is a new way to understand human activities using sensors. It can tell what people are doing and will do next by breaking down sensor data into small pieces called “patches”. This helps with healthcare and assisted living because it’s important to know what’s happening in real-time. P2LHAP does this more accurately than other methods, making it useful for many applications. |
Keywords
» Artificial intelligence » Embedding » Seq2seq » Transformer