xrsrke commited on
Commit
ecc4472
·
1 Parent(s): 5cc7587

add oslo and tunibai to appendex

Browse files
Files changed (2) hide show
  1. dist/index.html +10 -0
  2. src/index.html +10 -0
dist/index.html CHANGED
@@ -2386,6 +2386,11 @@
2386
  <a href="https://github.com/kakaobrain/torchgpipe"><strong>torchgpipe</strong></a>
2387
  <p>A GPipe implementation in PyTorch.</p>
2388
  </div>
 
 
 
 
 
2389
 
2390
  <h3>Debugging</h3>
2391
 
@@ -2504,6 +2509,11 @@
2504
  <a href="https://www.harmdevries.com/post/context-length/"><strong>Harm's blog for long context</strong></a>
2505
  <p>Investigation into long context training in terms of data and training cost.</p>
2506
  </div>
 
 
 
 
 
2507
 
2508
  <h2>Appendix</h2>
2509
 
 
2386
  <a href="https://github.com/kakaobrain/torchgpipe"><strong>torchgpipe</strong></a>
2387
  <p>A GPipe implementation in PyTorch.</p>
2388
  </div>
2389
+
2390
+ <div>
2391
+ <a href="https://github.com/EleutherAI/oslo"><strong>OSLO</strong></a>
2392
+ <p>OSLO: Open Source for Large-scale Optimization.</p>
2393
+ </div>
2394
 
2395
  <h3>Debugging</h3>
2396
 
 
2509
  <a href="https://www.harmdevries.com/post/context-length/"><strong>Harm's blog for long context</strong></a>
2510
  <p>Investigation into long context training in terms of data and training cost.</p>
2511
  </div>
2512
+
2513
+ <div>
2514
+ <a href="https://github.com/tunib-ai/large-scale-lm-tutorials"><strong>TunibAI's 3D parallelism tutorial</strong></a>
2515
+ <p>Large-scale language modeling tutorials with PyTorch.</p>
2516
+ </div>
2517
 
2518
  <h2>Appendix</h2>
2519
 
src/index.html CHANGED
@@ -2386,6 +2386,11 @@
2386
  <a href="https://github.com/kakaobrain/torchgpipe"><strong>torchgpipe</strong></a>
2387
  <p>torchgpipe: On-the-fly Pipeline Parallelism for Training Giant Models.</p>
2388
  </div>
 
 
 
 
 
2389
 
2390
  <h3>Debugging</h3>
2391
 
@@ -2504,6 +2509,11 @@
2504
  <a href="https://www.harmdevries.com/post/context-length/"><strong>Harm's blog for long context</strong></a>
2505
  <p>Investigation into long context training in terms of data and training cost.</p>
2506
  </div>
 
 
 
 
 
2507
 
2508
  <h2>Appendix</h2>
2509
 
 
2386
  <a href="https://github.com/kakaobrain/torchgpipe"><strong>torchgpipe</strong></a>
2387
  <p>torchgpipe: On-the-fly Pipeline Parallelism for Training Giant Models.</p>
2388
  </div>
2389
+
2390
+ <div>
2391
+ <a href="https://github.com/EleutherAI/oslo"><strong>OSLO</strong></a>
2392
+ <p>OSLO: Open Source for Large-scale Optimization.</p>
2393
+ </div>
2394
 
2395
  <h3>Debugging</h3>
2396
 
 
2509
  <a href="https://www.harmdevries.com/post/context-length/"><strong>Harm's blog for long context</strong></a>
2510
  <p>Investigation into long context training in terms of data and training cost.</p>
2511
  </div>
2512
+
2513
+ <div>
2514
+ <a href="https://github.com/tunib-ai/large-scale-lm-tutorials"><strong>TunibAI's 3D parallelism tutorial</strong></a>
2515
+ <p>Large-scale language modeling tutorials with PyTorch.</p>
2516
+ </div>
2517
 
2518
  <h2>Appendix</h2>
2519