Search Results for author: Shengnan Wang

Found 4 papers, 0 papers with code

XL3M: A Training-free Framework for LLM Length Extension Based on Segment-wise Inference

no code implementations28 May 2024 Shengnan Wang, Youhui Bai, Lin Zhang, Pingyi Zhou, Shixiong Zhao, Gong Zhang, Sen Wang, Renhai Chen, Hua Xu, Hongwei Sun

Under the XL3M framework, the input context will be firstly decomposed into multiple short sub-contexts, where each sub-context contains an independent segment and a common ``question'' which is a few tokens from the end of the original context.

Language Modelling Large Language Model

Digital twin-assisted three-dimensional electrical capacitance tomography for multiphase flow imaging

no code implementations22 Dec 2023 Shengnan Wang, Yi Li, Zhou Chen, Yunjie Yang

Three-dimensional electrical capacitance tomography (3D-ECT) has shown promise for visualizing industrial multiphase flows.

Computational Efficiency

Progressively Stacking 2.0: A Multi-stage Layerwise Training Method for BERT Training Speedup

no code implementations27 Nov 2020 Cheng Yang, Shengnan Wang, Chao Yang, Yuechuan Li, Ru He, Jingqiao Zhang

In BERT training, the backward computation is much more time-consuming than the forward computation, especially in the distributed training setting in which the backward computation time further includes the communication time for gradient synchronization.

CoRe: An Efficient Coarse-refined Training Framework for BERT

no code implementations27 Nov 2020 Cheng Yang, Shengnan Wang, Yuechuan Li, Chao Yang, Ming Yan, Jingqiao Zhang, Fangquan Lin

In the second phase, we transform the trained relaxed BERT model into the original BERT and further retrain the model.

Cannot find the paper you are looking for? You can Submit a new open access paper.