no code implementations • 2 May 2024 • Daxin Li, Yuanchao Bai, Kai Wang, Junjun Jiang, Xianming Liu, Wen Gao
To further expedite the network inference, we introduce context cache optimization to GroupedMixer, which caches attention activation values in cross-group token-mixers and avoids complex and duplicated computation.
no code implementations • 25 Jan 2024 • Daxin Li, Yuanchao Bai, Kai Wang, Junjun Jiang, Xianming Liu
Recent advancements in neural compression have surpassed traditional codecs in PSNR and MS-SSIM measurements.