All public logs
Jump to navigation
Jump to search
Combined display of all available logs of Robowaifu Institute of Technology. You can narrow down the view by selecting a log type, the username (case-sensitive), or the affected page (also case-sensitive).
- 15:25, 28 April 2023 RobowaifuDev talk contribs created page Scaling Transformer to 1M tokens and beyond with RMT (Created page with "Category:2023Category:April 2023Category:Research paperCategory:Natural language processing '''Scaling Transformer to 1M tokens and beyond with RMT''' is a research paper on the application of a recurrent memory mechanism to extend the context length of BERT, one of the most effective Transformer-based models in natural language processing.<ref>Bulatov et al. "Scaling Transformer to 1M tokens and beyond with RMT." 2023. [h...") Tag: Visual edit: Switched