no code implementations • ICLR 2019 • Lili Meng, Bo Zhao, Bo Chang, Gao Huang, Frederick Tung, Leonid Sigal
Our model is efficient, as it proposes a separable spatio-temporal mechanism for video attention, while being able to identify important parts of the video both spatially and temporally.
Action Recognition In Videos Temporal Action Localization +1
no code implementations • 3 Oct 2023 • Qi Yan, Raihan Seraj, JiaWei He, Lili Meng, Tristan Sylvain
Following this, the chosen articles are subjected to zero-shot summarization to attain succinct context.
1 code implementation • 13 Sep 2023 • Mahsa Keramati, Lili Meng, R. David Evans
ConR addresses the continuous nature of label space with two main strategies in a contrastive manner: incorrect proximities are penalized proportionate to the label similarities and the correct ones are encouraged to model local similarities.
no code implementations • 21 Jun 2023 • Chiyu Zhang, Qi Yan, Lili Meng, Tristan Sylvain
Despite these advances, there remains a significant gap in understanding the impact of different SSCL strategies on time series forecasting performance, as well as the specific benefits that SSCL can bring.
no code implementations • 18 Jun 2023 • Fanxin Xia, Jian Jin, Lili Meng, Feng Ding, Huaxiang Zhang
GAN-based image compression schemes have shown remarkable progress lately due to their high perceptual quality at low bit rates.
1 code implementation • 30 May 2023 • Melissa Mozifian, Tristan Sylvain, Dave Evans, Lili Meng
Attention-based sequential recommendation methods have shown promise in accurately capturing users' evolving interests from their past interactions.
no code implementations • 25 Feb 2023 • Feng Ding, Jian Jin, Lili Meng, Weisi Lin
After combining them together, we can better assign the distortion in the compressed image with the guidance of JND to preserve the high perceptual quality.
1 code implementation • 16 Aug 2022 • Jian Jin, Yuan Xue, Xingxing Zhang, Lili Meng, Yao Zhao, Weisi Lin
However, they have a major drawback that the generated JND is assessed in the real-world signal domain instead of in the perceptual domain in the human brain.
1 code implementation • 8 Jun 2022 • Amin Shabani, Amir Abdi, Lili Meng, Tristan Sylvain
The performance of time series forecasting has recently been greatly improved by the introduction of transformers.
no code implementations • 1 Mar 2022 • Jian Jin, Dong Yu, Weisi Lin, Lili Meng, Hao Wang, Huaxiang Zhang
Besides, the JND of the red and blue channels are larger than that of the green one according to the experimental results of the proposed model, which demonstrates that more changes can be tolerated in the red and blue channels, in line with the well-known fact that the human visual system is more sensitive to the green channel in comparison with the red and blue ones.
no code implementations • 7 Jan 2022 • Jian Jin, Xingxing Zhang, Lili Meng, Weisi Lin, Jie Liang, Huaxiang Zhang, Yao Zhao
Experimental results show that the VSD can be accurately estimated with the weights learnt by the nonlinear mapping function once its associated S-VSDs are available.
no code implementations • 19 Dec 2021 • Sien Chen, Jian Jin, Lili Meng, Weisi Lin, Zhuo Chen, Tsui-Shan Chang, Zhengguang Li, Huaxiang Zhang
Meanwhile, an image predictor is designed and trained to achieve the general-quality image reconstruction with the 16-bit gray-scale profile and signal features.
no code implementations • 7 Aug 2019 • Zhiwei Deng, Megha Nawhal, Lili Meng, Greg Mori
In this paper, we propose Continuous Graph Flow, a generative continuous flow based method that aims to model complex distributions of graph-structured data.
no code implementations • CVPR 2019 • Bo Zhao, Lili Meng, Weidong Yin, Leonid Sigal
The representation of each object is disentangled into a specified/certain part (category) and an unspecified/uncertain part (appearance).
Ranked #2 on Layout-to-Image Generation on Visual Genome 64x64
no code implementations • 1 Oct 2018 • Lili Meng, Bo Zhao, Bo Chang, Gao Huang, Wei Sun, Frederich Tung, Leonid Sigal
Inspired by the observation that humans are able to process videos efficiently by only paying attention where and when it is needed, we propose an interpretable and easy plug-in spatial-temporal attention mechanism for video action recognition.
3 code implementations • 25 Jan 2018 • Bo Chang, Qiong Zhang, Shenyi Pan, Lili Meng
Our method is applied not only to commonly used Chinese characters but also to calligraphy work with aesthetic values.
no code implementations • 28 Oct 2017 • Lili Meng, Frederick Tung, James J. Little, Julien Valentin, Clarence de Silva
Camera relocalization plays a vital role in many robotics and computer vision tasks, such as global localization, recovery from tracking failure and loop closure detection.
no code implementations • ICLR 2018 • Bo Chang, Lili Meng, Eldad Haber, Frederick Tung, David Begert
Deep residual networks (ResNets) and their variants are widely used in many computer vision applications and natural language processing tasks.
1 code implementation • 22 Oct 2017 • Lili Meng, Jianhui Chen, Frederick Tung, James J. Little, Julien Valentin, Clarence W. de Silva
Camera relocalization plays a vital role in many robotics and computer vision tasks, such as global localization, recovery from tracking failure, and loop closure detection.
2 code implementations • 12 Sep 2017 • Bo Chang, Lili Meng, Eldad Haber, Lars Ruthotto, David Begert, Elliot Holtham
In this work, we interpret deep residual networks as ordinary differential equations (ODEs), which have long been studied in mathematics and physics with rich theoretical and empirical success.
Ranked #49 on Image Classification on STL-10
no code implementations • 9 Jul 2017 • Lijun Zhao, Jie Liang, Huihui Bai, Lili Meng, Anhong Wang, Yao Zhao
Both frameworks employ the division of gradient and the local activity measurement to achieve noise removal.