HANDS-ON 3D EM DESIGN WORKSHOP USING EMPIRE XPU

Virtual: https://events.vtools.ieee.org/m/420515

A virtual technical workshop will be organized by the experts of Empire XPU team, where they will give hands on demonstration of their cutting edge EM simulation software and cover important topics related to EM engineering. Specifically, the following topics would be covered: - Advanced EM simulation using fast FDTD algorithm - Antenna and Radar applications - Selected Empire Features for antenna and Radar simulations - Antenna Array modelling and design - Radar Rx phase difference evaluation Virtual: https://events.vtools.ieee.org/m/420515

IEEE SPS SBC Webinar: Face video compression with generative networks (By Dr. Yan Ye)

Virtual: https://events.vtools.ieee.org/m/422747

Video coding is a fundamental and ubiquitous technology in the modern society. Generations of international video coding standards, such as the widely-deployed H.264/AVC and H.265/HEVC and the latest H.266/VVC, provide essential means for enabling video conferencing, video streaming, video sharing, e-commerce, entertainment, and many more video applications. These existing standards all rely on the fundamental theory of signal processing and information theory to encode generic video efficiently with a favorable rate distortion behavior. In recent years, rapid advancement in deep learning and artificial intelligence technology has allowed people to manipulate images and videos using deep generative models. Among these, of particular interest to the field of video coding is the application of deep generative models towards compressing talking-face video at ultra-low bit rates. By focusing on talking faces, generative models can effectively learn the inherent structure about composition, movement and posture of human faces and deliver promising results using very little bandwidth resource. At ultra-low bit rates, when even the latest video coding standard H.266/VVC is apt to suffer from significant blocking artifacts and blurriness beyond the point of recognition, generative methods can maintain clear facial features and vivid expression in the reconstructed video. Further, generative face video coding techniques are inherently capable of manipulating the reconstructed face and promise to deliver a more interactive experience. In this talk, we start with a quick overview of traditional and deep learning-based video coding techniques. We then focus on face video coding with generative networks, and present two schemes that send different deep information in the bitstream, one sending compact temporal motion features and the other sending 3D facial semantics. We compare their compression efficiency and visual quality with that of the latest H.266/VVC standard, and showcase the power of deep generative models in preserving vivid facial images with little bandwidth resource. We also present visualization results to exhibit the capability of the 3D facial semantics-based scheme in terms of interacting with the reconstructed face video and animating virtual faces. Speaker(s): Dr. Yan Ye, Virtual: https://events.vtools.ieee.org/m/422747