Difference between revisions of "Resource:Seminar"

From MobiNetS
Jump to: navigation, search
(8 intermediate revisions by 2 users not shown)
Line 1: Line 1:
{{SemNote
{{SemNote
|time='''Thursday 9:00-10:30'''
|time='''Friday 10:30-12:00'''
|addr=4th Research Building A518
|addr=4th Research Building A518
|note=Useful links: [[Resource:Reading_List|Readling list]]; [[Resource:Seminar_schedules|Schedules]]; [[Resource:Previous_Seminars|Previous seminars]].
|note=Useful links: [[Resource:Reading_List|Readling list]]; [[Resource:Seminar_schedules|Schedules]]; [[Resource:Previous_Seminars|Previous seminars]].
Line 7: Line 7:
===Latest===
===Latest===
{{Latest_seminar
{{Latest_seminar
|abstract=The massive connection of LoRa brings serious collision interference. Existing collision decoding methods cannot effectively deal with the adjacent collisions that occur when the collided symbols are adjacent in the frequency spectrum. The decoding features relied on by the existing methods will be corrupted by adjacent collisions. To address these issues, we propose Paralign, which is the first LoRa collision decoder supporting decoding LoRa collisions with confusing symbols via parallel alignment. The key enabling technology behind Paralign is tha there is no spectrum leakage with periodic truncation of a chirp. Paralign leverages the precise spectrum obtained by aligning the de-chirped periodic signals from each packet in parallel for spectrum filtering and power filtering. To aggregate correlation peaks in different windows of the same symbol, Paralign matches the peaks of multiple interfering windows to the interested window based on the time offset between collided packets. Moreover, a periodic truncation method is proposed to address the multiple candidate peak problem caused by side lobes of confusing symbols. We evaluate Paralign using USRP N210 in a 20-node network. Experimental results demonstrate that Paralign can significantly improve network throughput, which is over 1.46× higher than state-of-the-art methods.
|abstract=We present NeuriCam, a novel deep learning-based system to achieve video capture from low-power dual-mode IoT camera systems. Our idea is to design a dual-mode camera system where the first mode is low power (1.1 mW) but only outputs grey-scale, low resolution and noisy video and the second mode consumes much higher power (100 mW) but outputs color and higher resolution images. To reduce total energy consumption, we heavily duty cycle the high power mode to output an image only once every second. The data for this camera system is then wirelessly sent to a nearby plugged-in gateway, where we run our real-time neural network decoder to reconstruct a higher-resolution color video. To achieve this, we introduce an attention feature filter mechanism that assigns different weights to different features, based on the correlation between the feature map and the contents of the input frame at each spatial location. We design a wireless hardware prototype using off-the-shelf cameras and address practical issues including packet loss and perspective mismatch. Our evaluations show that our dual-camera approach reduces energy consumption by 7x compared to existing systems. Further, our model achieves an average greyscale PSNR gain of 3.7 dB over prior single and dual-camera video super-resolution methods and 5.6 dB RGB gain over prior color propagation methods.
|confname=ToSN '23
|confname=MobiCom 2023
|link=https://dl.acm.org/doi/10.1145/3571586
|link=https://dl.acm.org/doi/10.1145/3570361.3592523
|title=Decoding LoRa Collisions via Parallel Alignment
|title=NeuriCam: Key-Frame Video Super-Resolution and Colorization for IoT Cameras
|speaker=Kai Chen
|speaker=Jiyi
|date=2024-01-04}}
|date=2024-04-12}}
{{Latest_seminar
{{Latest_seminar
|abstract=Human activity recognition is important for a wide range of applications such as surveillance systems and human-computer interaction. Computer vision based human activity recognition suffers from performance degradation in many real-world scenarios where the illumination is poor. On the other hand, recently proposed WiFi sensing that leverage ubiquitous WiFi signal for activity recognition is not affected by illumination but has low accuracy in dynamic environments. In this paper, we propose WiMix, a lightweight and robust multimodal system that leverages both WiFi and vision for human activity recognition. To deal with complex real-world environments, we design a lightweight mix cross attention module for automatic WiFi and video weight distribution. To reduce the system response time while ensuring the sensing accuracy, we design an end-to-end framework together with an efficient classifier to extract spatial and temporal features of two modalities. Extensive experiments are conducted in the real-world scenarios and the results demonstrate that WiMix achieves 98.5% activity recognition accuracy in 3 scenarios, which outperforms the state-of-the-art 89.6% sensing accuracy using WiFi and video modalities. WiMix can also reduce the inference latency from 1268.25ms to 217.36ms, significantly improving the response time.
|abstract=The dominant sequence transduction models are based on complex recurrent or convolutional neural networks in an encoder-decoder configuration. The best performing models also connect the encoder and decoder through an attention mechanism. We propose a new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two machine translation tasks show these models to be superior in quality while being more parallelizable and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task, improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training costs of the best models from the literature. We show that the Transformer generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data.
|confname=MASS '23
|confname=Neurips 2017
|link=https://ieeexplore.ieee.org/abstract/document/10298524
|link=https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf
|title=WiMix: A Lightweight Multimodal Human Activity Recognition System based on WiFi and Vision
|title=Attention Is All You Need
|speaker=Haotian
|speaker=Qinyong
|date=2024-01-04}}
|date=2024-04-12}}
{{Latest_seminar
|abstract=In wireless networks, the Named Data Networking (NDN) architecture maximizes contents’ availability throughout multiple network paths based on multicast-based communication combined with stateful forwarding and caching in intermediate nodes. Despite its benefits, the downside of the architecture resides in the packet flooding not efficiently prevented by current forwarding strategies and mainly associated with the native flooding of the architecture or malicious packets from Interest Flooding Attack (IFA). This work introduces iFLAT, a multicriteria-based forwarding strategy for i nterest FL ooding mitig AT ion on named data wireless networking. It follows a cross-layer approach that considers: (i) the Received Signal Strength (RSS) to adjust itself to the current status of the wireless network links, (ii) the network traffic ( meInfo ) to detect flooding issues and traffic anomalies, and (iii) a fake Interest Blacklist to identify IFA-related flooding. In doing so, the strategy achieves better efficiency on Interest flooding mitigation, addressing both native and IFA types of flooding. We evaluated the proposed strategy by reproducing a Flying Ad hoc Network (FANET) composed of Unmanned Aerial Vehicles (UAVs) featured by the NDN stack deployed in all nodes. Simulation results show that iFLAT can effectively detect and mitigate flooding, regardless of its origin or nature, achieving greater packet forwarding efficiency than competing strategies. In broadcast storm and IFA scenarios, it achieved 25.75% and 37.37% of traffic reduction, whereas Interest satisfaction rates of 16.36% and 51.78% higher, respectively.
|confname=TMC '23
|link=https://ieeexplore.ieee.org/document/9888056
|title=A Multicriteria-Based Forwarding Strategy for Interest Flooding Mitigation on Named Data Wireless Networking
|speaker=Zhenghua
|date=2024-01-04}}
{{Latest_seminar
|abstract=While recent work explored streaming volumetric content on-demand, there is little effort on live volumetric video streaming that bears the potential of bringing more exciting applications than its on-demand counterpart. To fill this critical gap, in this paper, we propose MetaStream, which is, to the best of our knowledge, the first practical live volumetric content capture, creation, delivery, and rendering system for immersive applications such as virtual, augmented, and mixed reality. To address the key challenge of the stringent latency requirement for processing and streaming a huge amount of 3D data, MetaStream integrates several innovations into a holistic system, including dynamic camera calibration, edge-assisted object segmentation, cross-camera redundant point removal, and foveated volumetric content rendering. We implement a prototype of MetaStream using commodity devices and extensively evaluate its performance. Our results demonstrate that MetaStream achieves low-latency live volumetric video streaming at close to 30 frames per second on WiFi networks. Compared to state-of-the-art systems, MetaStream reduces end-to-end latency by up to 31.7% while improving visual quality by up to 12.5%.
|confname=MobiCom '23
|link=https://dl.acm.org/doi/abs/10.1145/3570361.3592530
|title=MetaStream: Live Volumetric Content Capture, Creation, Delivery, and Rendering in Real Time
|speaker=Jiale
|date=2024-01-04}}
{{Resource:Previous_Seminars}}
{{Resource:Previous_Seminars}}

Revision as of 15:10, 9 April 2024

Time: Friday 10:30-12:00
Address: 4th Research Building A518
Useful links: Readling list; Schedules; Previous seminars.

Latest

  1. [MobiCom 2023] NeuriCam: Key-Frame Video Super-Resolution and Colorization for IoT Cameras, Jiyi
    Abstract: We present NeuriCam, a novel deep learning-based system to achieve video capture from low-power dual-mode IoT camera systems. Our idea is to design a dual-mode camera system where the first mode is low power (1.1 mW) but only outputs grey-scale, low resolution and noisy video and the second mode consumes much higher power (100 mW) but outputs color and higher resolution images. To reduce total energy consumption, we heavily duty cycle the high power mode to output an image only once every second. The data for this camera system is then wirelessly sent to a nearby plugged-in gateway, where we run our real-time neural network decoder to reconstruct a higher-resolution color video. To achieve this, we introduce an attention feature filter mechanism that assigns different weights to different features, based on the correlation between the feature map and the contents of the input frame at each spatial location. We design a wireless hardware prototype using off-the-shelf cameras and address practical issues including packet loss and perspective mismatch. Our evaluations show that our dual-camera approach reduces energy consumption by 7x compared to existing systems. Further, our model achieves an average greyscale PSNR gain of 3.7 dB over prior single and dual-camera video super-resolution methods and 5.6 dB RGB gain over prior color propagation methods.
  2. [Neurips 2017] Attention Is All You Need, Qinyong
    Abstract: The dominant sequence transduction models are based on complex recurrent or convolutional neural networks in an encoder-decoder configuration. The best performing models also connect the encoder and decoder through an attention mechanism. We propose a new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two machine translation tasks show these models to be superior in quality while being more parallelizable and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task, improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training costs of the best models from the literature. We show that the Transformer generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data.

History

2024

2023

2022

2021

2020

  • [Topic] [ The path planning algorithm for multiple mobile edge servers in EdgeGO], Rong Cong, 2020-11-18

2019

2018

2017

Template loop detected: Resource:Previous Seminars

Instructions

请使用Latest_seminar和Hist_seminar模板更新本页信息.

    • 修改时间和地点信息
    • 将当前latest seminar部分的code复制到这个页面
    • 将{{Latest_seminar... 修改为 {{Hist_seminar...,并增加对应的日期信息|date=
    • 填入latest seminar各字段信息
    • link请务必不要留空,如果没有link则填本页地址 https://mobinets.org/index.php?title=Resource:Seminar
  • 格式说明
    • Latest_seminar:

{{Latest_seminar
|confname=
|link=
|title=
|speaker=
}}

    • Hist_seminar

{{Hist_seminar
|confname=
|link=
|title=
|speaker=
|date=
}}