-
Notifications
You must be signed in to change notification settings - Fork 13
szhangtju/The-compression-of-Transformer
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
## Introduction This directory contains our pytorch implementation of Tensorized Transformer. The code is based on the code( https://github.com/kimiyoung/transformer-xl). Note that our state-of-the-art results reported in the paper were obtained by training the model on a GPU. ## Prerequisite - Pytorch 1.0.0 ## Data Prepration(Language Modeling) PTB, Wiki-Text-103, One-Billion. ## Training and Evaluation if run Language Modeling Test in the case of one core, then bash run_(ptb/wt1o3).sh if run Language Modeling Test in the case of multiple core, then first set the hyperparameter core_nums in the transformer_upload.py file to 2, then run the bash file.
About
No description, website, or topics provided.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published