Advancing AGI for humanity
⇛
Research
Blog
About
Star
...
LLM Accelerator: Lossless Acceleration of Large Language Models
Apr 11, 2023
Kosmos-1: A Multimodal Large Language Model (MLLM)
Feb 28, 2023
VALL-E (X): Neural Codec Language Models are Zero-Shot Text to Speech Synthesizers
Jan 6, 2023
A Length-Extrapolatable Transformer
Dec 20, 2022
Why Can GPT Learn In-Context? Language Models Secretly Perform Gradient Descent as Meta Optimizers
Dec 20, 2022
Promptist - Optimizing Prompts for Text-to-Image Generation
Dec 19, 2022
Structured Prompting: Scaling In-Context Learning to 1,000 Examples
Dec 12, 2022
Prompt Intelligence - Extensible Prompts
Dec 2, 2022
TorchScale - Transformers at (Any) Scale
Nov 24, 2022
Foundation Transformers
October 13, 2022
BEiT-3: A General-Purpose Multimodal Foundation Model
Aug 30, 2022
Language Models are General-Purpose Interfaces
June 13, 2022
DeepNet: Scaling Transformers to 1,000 Layers
Mar 1, 2022
BEiT: BERT Pre-Training of Image Transformers
June 15, 2021
XLM-E: Efficient Multilingual Language Model Pre-training
June 30, 2021
UniLM: Unified Language Model Pre-training
May 8, 2019