exdysa commited on
Commit
c5c1f62
·
verified ·
1 Parent(s): eb94961

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -46
README.md CHANGED
@@ -1,53 +1,24 @@
1
  ---
2
- datasets:
3
- - Skylion007/openwebtext
4
- language:
5
- - en
6
- library_name: transformers
7
  license: apache-2.0
8
- metrics:
9
- - perplexity
10
  pipeline_tag: text-generation
 
 
 
11
  ---
12
-
13
- ## Using DUO
14
- To use the pre-trained model for masked language modeling, use the following snippet:
15
- ```python
16
- from transformers import AutoModelForMaskedLM, AutoTokenizer
17
-
18
- # See the `MDLM` collection page on the hub for list of available models.
19
- tokenizer = transformers.AutoTokenizer.from_pretrained('gpt2')
20
- model = AutoModelForMaskedLM.from_pretrained('s-sahoo/duo')
21
- ```
22
- For a hands-on example, check out this [Colab notebook](https://colab.research.google.com/drive/1Sf7R-dqdR6gq-H8nyZ9E3ZkyvqMTqcwq?usp=sharing).
23
- For more information and implementation details, visit our github repository: [DUO](https://github.com/s-sahoo/duo) and project page: [Project Page](https://s-sahoo.com/duo)
24
-
25
- ## Model Details
26
- The model, which has a context length of `1024` and is similar in size to GPT2-medium with approximately `130 million` non-embedding parameters,
27
- was trained for 1M steps on the OpenWebText corpus.
28
-
29
- For more details, please see our paper: [The Diffusion Duality](https://openreview.net/forum?id=CB0Ub2yXjC).
30
-
31
-
32
-
33
- ## Citation
34
-
35
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
36
- Please cite our work using the bibtex below:
37
-
38
- **BibTeX:**
39
 
40
  ```
41
- @inproceedings{
42
- sahoo2025the,
43
- title={The Diffusion Duality},
44
- author={Subham Sekhar Sahoo and Justin Deschenaux and Aaron Gokaslan and Guanghan Wang and Justin T Chiu and Volodymyr Kuleshov},
45
- booktitle={Forty-second International Conference on Machine Learning},
46
- year={2025},
47
- url={https://openreview.net/forum?id=9P9Y8FOSOk}
48
- }
49
  ```
50
-
51
-
52
- ## Model Card Contact
53
- Subham Sekhar Sahoo (ssahoo@cs.cornell.edu)
 
1
  ---
2
+ name: duo
3
+ base_model: s-sahoo/duo
 
 
 
4
  license: apache-2.0
 
 
5
  pipeline_tag: text-generation
6
+ tasks: text-generation
7
+ language: en
8
+ datasets: Skylion007/openwebtext
9
  ---
10
+ > [!IMPORTANT]
11
+ > Original Model Link : [https://huggingface.co/s-sahoo/duo](https://huggingface.co/s-sahoo/duo)
12
+ >
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
  ```
15
+ name: duo
16
+ base_model: s-sahoo/duo
17
+ license: apache-2.0
18
+ pipeline_tag: text-generation
19
+ tasks: text-generation
20
+ language: en
21
+ datasets: Skylion007/openwebtext
 
22
  ```
23
+ # Duo
24
+ Duo is a diffusion masked language model with 130m parameters. This repo makes the `flash_attn` and CUDA libraries optional so that all frameworks can be supported. [Code for this model](https://github.com/s-sahoo/duo)