Seguir
Shaohan Huang
Shaohan Huang
Microsoft Research Asia
Email confirmado em microsoft.com
Título
Citado por
Citado por
Ano
Layoutlm: Pre-training of text and layout for document image understanding
Y Xu, M Li, L Cui, S Huang, F Wei, M Zhou
Proceedings of the 26th ACM SIGKDD international conference on knowledge …, 2020
7462020
Superagent: A customer service chatbot for e-commerce websites
L Cui, S Huang, F Wei, C Tan, C Duan, M Zhou
Proceedings of ACL 2017, system demonstrations, 97-102, 2017
4792017
Neural document summarization by jointly learning to score and select sentences
Q Zhou, N Yang, F Wei, S Huang, M Zhou, T Zhao
arXiv preprint arXiv:1807.02305, 2018
4432018
Kosmos-2: Grounding multimodal large language models to the world
Z Peng, W Wang, L Dong, Y Hao, S Huang, S Ma, F Wei
arXiv preprint arXiv:2306.14824, 2023
4142023
Language is not all you need: Aligning perception with language models
S Huang, L Dong, W Wang, Y Hao, S Singhal, S Ma, T Lv, L Cui, ...
Advances in Neural Information Processing Systems 36, 72096-72109, 2023
3792023
Learning to generate product reviews from attributes
L Dong, S Huang, F Wei, M Lapata, M Zhou, K Xu
Proceedings of the 15th Conference of the European Chapter of the …, 2017
2172017
Tablebank: Table benchmark for image-based table detection and recognition
M Li, L Cui, S Huang, F Wei, M Zhou, Z Li
Proceedings of the Twelfth Language Resources and Evaluation Conference …, 2020
1992020
Minilmv2: Multi-head self-attention relation distillation for compressing pretrained transformers
W Wang, H Bao, S Huang, L Dong, F Wei
arXiv preprint arXiv:2012.15828, 2020
1962020
Retentive network: A successor to transformer for large language models
Y Sun, L Dong, S Huang, S Ma, Y Xia, J Xue, J Wang, F Wei
arXiv preprint arXiv:2307.08621, 2023
1932023
DocBank: A benchmark dataset for document layout analysis
M Li, Y Xu, L Cui, S Huang, F Wei, Z Li, M Zhou
arXiv preprint arXiv:2006.01038, 2020
1882020
Hitanomaly: Hierarchical transformers for anomaly detection in system log
S Huang, Y Liu, C Fung, R He, Y Zhao, H Yang, Z Luan
IEEE transactions on network and service management 17 (4), 2064-2076, 2020
1702020
Promptbert: Improving bert sentence embeddings with prompts
T Jiang, J Jiao, S Huang, Z Zhang, D Wang, F Zhuang, F Wei, H Huang, ...
arXiv preprint arXiv:2201.04337, 2022
1642022
Response generation by context-aware prototype editing
Y Wu, F Wei, S Huang, Y Wang, Z Li, M Zhou
Proceedings of the AAAI Conference on Artificial Intelligence 33 (01), 7281-7288, 2019
1382019
Deepnet: Scaling transformers to 1,000 layers
H Wang, S Ma, L Dong, S Huang, D Zhang, F Wei
IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024
1362024
Language generation with multi-hop reasoning on commonsense knowledge graph
H Ji, P Ke, S Huang, F Wei, X Zhu, M Huang
arXiv preprint arXiv:2009.11692, 2020
1322020
XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
Z Chi
arXiv preprint arXiv:2106.16138, 2021
1242021
A length-extrapolatable transformer
Y Sun, L Dong, B Patra, S Ma, S Huang, A Benhaim, V Chaudhary, ...
arXiv preprint arXiv:2212.10554, 2022
1172022
Longnet: Scaling transformers to 1,000,000,000 tokens
J Ding, S Ma, L Dong, X Zhang, S Huang, W Wang, N Zheng, F Wei
arXiv preprint arXiv:2307.02486, 2023
1082023
Language models are general-purpose interfaces
Y Hao, H Song, L Dong, S Huang, Z Chi, W Wang, S Ma, F Wei
arXiv preprint arXiv:2206.06336, 2022
942022
The era of 1-bit llms: All large language models are in 1.58 bits
S Ma, H Wang, L Ma, L Wang, W Wang, S Huang, L Dong, R Wang, J Xue, ...
arXiv preprint arXiv:2402.17764, 2024
772024
O sistema não pode efectuar a operação agora. Tente mais tarde.
Artigos 1–20