justheuristic commited on
Commit
fce987b
Β·
1 Parent(s): ea1b028

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +39 -2
README.md CHANGED
@@ -14,13 +14,50 @@ pinned: false
14
 
15
  ---
16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  <div class="lg:col-span-3">
18
 
19
- <img src="https://raw.githubusercontent.com/NCAI-Research/CALM/main/assets/logo.png" width="380" alt="CALM Logo" />
20
 
21
  <!-- this is actually not markdown, please use regular HTML -->
22
 
23
- # CALM: Collaborative Arabic Language Model
24
 
25
  The CALM project is joint effort lead by <a href="https://sdaia.gov.sa/ncai/?Lang=en">NCAI</a> in collaboration with <a href="https://yandex.com/">Yandex</a> and <a href="https://huggingface.co/">HuggingFace</a> to train an Arabic language model with volunteers from around the globe. The project is an adaptation of the framework proposed at the NeurIPS 2021 demonstration: <a href="https://huggingface.co/training-transformers-together">Training Transformers Together</a>.
26
 
 
14
 
15
  ---
16
 
17
+
18
+ <!-- The classes below are necessary for correct rendering -->
19
+ <div class="lg:col-span-3">
20
+ <img src="https://raw.githubusercontent.com/NCAI-Research/CALM/main/assets/logo.png" width="380" alt="CALM Logo" />
21
+ <p class="mb-2">
22
+ This organization is a part of the NeurIPS 2021 demonstration <u><a href="https://training-transformers-together.github.io/">"Training Transformers Together"</a></u>.
23
+ </p>
24
+ <p class="mb-2">
25
+ In this demo, we train a model similar to <u><a target="_blank" href="https://openai.com/blog/dall-e/">OpenAI DALL-E</a></u> β€”
26
+ a Transformer "language model" that generates images from text descriptions.
27
+ Training happens collaboratively β€” volunteers from all over the Internet contribute to the training using hardware available to them.
28
+ We use <u><a target="_blank" href="https://laion.ai/laion-400-open-dataset/">LAION-400M</a></u>,
29
+ the world's largest openly available image-text-pair dataset with 400 million samples. Our model is based on
30
+ the <u><a target="_blank" href="https://github.com/lucidrains/DALLE-pytorch">dalle‑pytorch</a></u> implementation
31
+ by <u><a target="_blank" href="https://github.com/lucidrains">Phil Wang</a></u> with a few tweaks to make it communication-efficient.
32
+ </p>
33
+ <p class="mb-2">
34
+ See details about how to join and how it works on <u><a target="_blank" href="https://training-transformers-together.github.io/">our website</a></u>.
35
+ </p>
36
+ <p class="mb-2">
37
+ This organization gathers people participating in the collaborative training and provides links to the necessary resources:
38
+ </p>
39
+ <ul class="mb-2">
40
+ <li>πŸ‘‰ Starter kits for <u><a target="_blank" href="https://colab.research.google.com/drive/1BqTWcfsvNQwQqqCRKMKp1_jvQ5L1BhCY?usp=sharing">Google Colab</a></u> and <u><a target="_blank" href="https://www.kaggle.com/yhn112/training-transformers-together/">Kaggle</a></u> (easy way to join the training)</li>
41
+ <li>πŸ‘‰ <u><a target="_blank" href="https://huggingface.co/spaces/training-transformers-together/Dashboard">Dashboard</a></u> (the current training state: loss, number of peers, etc.)</li>
42
+ <li>πŸ‘‰ <u><a target="_blank" href="https://colab.research.google.com/drive/1Vkb-4nhEEH1a5vrKtpL4MTNiUTPdpPUl?usp=sharing">Colab notebook for running inference</a></u>
43
+ <li>πŸ‘‰ <u><a target="_blank" href="https://huggingface.co/training-transformers-together/dalle-demo-v1">Model weights</a></u> (the latest checkpoint)</li></li>
44
+ <li>πŸ‘‰ Weights & Biases plots for <u><a target="_blank" href="https://wandb.ai/learning-at-home/dalle-hivemind/runs/3l7q56ht">aux peers</a></u> (aggregating the metrics) and actual <u><a target="_blank" href="https://wandb.ai/learning-at-home/dalle-hivemind-trainers">trainers</a></u> (contributing with their GPUs)</li>
45
+ <li>πŸ‘‰ <u><a target="_blank" href="https://github.com/learning-at-home/dalle-hivemind">Code</a></u></li>
46
+ <li>πŸ‘‰ <u><a target="_blank" href="https://huggingface.co/datasets/laion/laion_100m_vqgan_f8">Dataset</a></u></li>
47
+ </ul>
48
+ <p class="mb-2">
49
+ Feel free to reach us on <u><a target="_blank" href="https://discord.gg/uGugx9zYvN">Discord</a></u> if you have any questions πŸ™‚
50
+ </p>
51
+ </div>
52
+
53
+
54
  <div class="lg:col-span-3">
55
 
56
+
57
 
58
  <!-- this is actually not markdown, please use regular HTML -->
59
 
60
+ #
61
 
62
  The CALM project is joint effort lead by <a href="https://sdaia.gov.sa/ncai/?Lang=en">NCAI</a> in collaboration with <a href="https://yandex.com/">Yandex</a> and <a href="https://huggingface.co/">HuggingFace</a> to train an Arabic language model with volunteers from around the globe. The project is an adaptation of the framework proposed at the NeurIPS 2021 demonstration: <a href="https://huggingface.co/training-transformers-together">Training Transformers Together</a>.
63