Skip to content

Commit

Permalink
Updated MLO
Browse files Browse the repository at this point in the history
  • Loading branch information
ineiti committed Apr 12, 2024
1 parent eae03eb commit 90b9a2b
Show file tree
Hide file tree
Showing 4 changed files with 133 additions and 51 deletions.
1 change: 1 addition & 0 deletions data.py
Original file line number Diff line number Diff line change
Expand Up @@ -139,6 +139,7 @@
"commercial",
"non-commercial",
"other",
"various"
])),
sy.Optional("information"): sy.Seq(
sy.Map({
Expand Down
140 changes: 95 additions & 45 deletions data/MLO/projects.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -10,9 +10,6 @@ projects:
Inspired by the PowerSGD algorithm for centralized deep learning, this algorithm uses power iteration steps to maximize the
information transferred per bit. We prove that our method requires no additional hyperparameters, converges faster than prior methods,
and is asymptotically independent of both the network and the compression.
contacts:
- name: Thijs Vogels
email: [email protected]
tags:
- Deep Neural Networks
type: Library
Expand All @@ -31,7 +28,7 @@ projects:
text: NeurIPS 2020
url: https://proceedings.neurips.cc/paper/2020/hash/a376802c0811f1b9088828288eb0d3f0-Abstract.html
date_added: 2021-03-05
date_updated: 2022-07-06
date_updated: 2024-04-09

chocosgd:
name: chocoSGD
Expand All @@ -43,9 +40,6 @@ projects:
tech_desc: >
Communication-efficient decentralized ML training (both deep learning, compatible with PyTorch, and traditional convex machine
learning models).
contacts:
- name: Tao Lin
email: [email protected]
code:
type: Lab GitHub
url: https://github.com/epfml/ChocoSGD
Expand All @@ -66,7 +60,7 @@ projects:
language: Python
license: Apache-2.0
date_added: 2019-07-30
date_updated: 2022-07-06
date_updated: 2024-04-09
maturity: 1

cola:
Expand Down Expand Up @@ -109,7 +103,7 @@ projects:
license: Apache-2.0
type: Application
date_added: 2019-07-30
date_updated: 2022-07-06
date_updated: 2024-04-09

mlbench:
name: MLBench
Expand Down Expand Up @@ -138,17 +132,14 @@ projects:
url: https://mlbench.github.io/blog/
tags:
- Benchmark
c4dt_contact:
name: Linus Gasser
email: [email protected]
language: Python
type: Framework
license: Apache-2.0
incubator:
type: retired
work: 2020/Q4 evaluated and tested the project
date_added: 2019-07-30
date_updated: 2023-03-22
date_updated: 2024-04-09
maturity: 2

sent2vec:
Expand All @@ -174,7 +165,7 @@ projects:
- Natural Language
type: Library
date_added: 2019-03-18
date_updated: 2023-03-22
date_updated: 2024-04-09
maturity: 1

powersgd:
Expand All @@ -189,13 +180,10 @@ projects:
compressed gradients using all-reduce, and iii) achieve test performance on par with SGD. The proposed algorithm is the only method
evaluated that achieves consistent wall-clock speedups when benchmarked against regular SGD with an optimized communication backend.
We demonstrate reduced training times for convolutional networks as well as LSTMs on common datasets.
contacts:
- name: Thijs Vogels
email: [email protected]
code:
type: Lab GitHub
url: https://github.com/epfml/powersgd
date_last_commit: 2022-11-22
date_last_commit: 2023-07-04
information:
- type: Paper
title: "PowerSGD: Practical Low-Rank Gradient Compression for Distributed Optimization"
Expand All @@ -217,7 +205,7 @@ projects:
type: Application
license: MIT
date_added: 2020-05-01
date_updated: 2023-03-22
date_updated: 2024-04-09
maturity: 2

relaysgd:
Expand All @@ -235,13 +223,10 @@ projects:
the RelaySum mechanism for information propagation in decentralized learning. RelaySum uses spanning trees to distribute information
exactly uniformly across all workers with finite delays depending on the distance between nodes. In contrast, the typical gossip
averaging mechanism only distributes data uniformly asymptotically while using the same communication volume per step as RelaySum.
contacts:
- name: Thijs Vogels
email: [email protected]
code:
type: La GitHub
type: Lab GitHub
url: https://github.com/epfml/relaysgd
date_last_commit: 2021-10-27
date_last_commit: 2023-04-21
language: Python
type: Library, Experiments
license: MIT
Expand All @@ -254,7 +239,7 @@ projects:
title: RelaySum for Decentralized Deep Learning on Heterogeneous Data
url: https://arxiv.org/pdf/2110.04175v1.pdf
date_added: 2021-11-04
date_updated: 2022-07-06
date_updated: 2024-04-09

disco:
name: Disco
Expand All @@ -271,16 +256,10 @@ projects:
creating machine learning models and people having relevant data. By running without installation nor complex configuration, anyone
can help. It's also a breeding ground for new machine learning technologies.
url: https://epfml.github.io/disco
c4dt_contact:
name: Linus Gasser
email: [email protected]
contacts:
- name: Martin Jaggi
email: [email protected]
code:
type: Lab GitHub
url: https://github.com/epfml/disco
date_last_commit: 2023-03-21
date_last_commit: 2024-04-03
language: TypeScript
license: Apache-2.0
incubator:
Expand Down Expand Up @@ -335,7 +314,7 @@ projects:
url: http://proceedings.mlr.press/v139/karimireddy21a.html
maturity: 2
date_added: 2022-02-18
date_updated: 2023-03-22
date_updated: 2024-04-09

byzantine-robust-optimizer:
name: Byzantine Robust Optimizer
Expand All @@ -351,8 +330,6 @@ projects:
worker momentum to overcome time-coupled attacks. This is the first provably robust method for the standard stochastic optimization
setting.
contacts:
- name: Sai Praneeth Reddy Karimireddy
email: [email protected]
- name: Lie He
email: [email protected]
code:
Expand All @@ -375,7 +352,7 @@ projects:
text: ICML 2021
url: https://icml.cc/Conferences/2021
date_added: 2021-11-04
date_updated: 2022-07-06
date_updated: 2024-04-09

quasi-global-momentum:
name: Quasi-Global Momentum
Expand All @@ -389,9 +366,6 @@ projects:
realistic learning scenarios, the presence of heterogeneity across different clients' local datasets poses an optimization challenge
and may severely deteriorate the generalization performance. We propose a novel momentum-based method to mitigate this decentralized
training difficulty.
contacts:
- name: Tao Lin
email: [email protected]
code:
type: Lab GitHub
url: https://github.com/epfml/quasi-global-momentum
Expand All @@ -412,7 +386,7 @@ projects:
text: ICML 2021
url: https://icml.cc/Conferences/2021
date_added: 2021-11-04
date_updated: 2023-03-22
date_updated: 2024-04-09

hyperaggregate:
name: HyperAggregate
Expand All @@ -438,7 +412,7 @@ projects:
title: "HyperAggregate: A sublinear secure aggregation protocol"
url: https://infoscience.epfl.ch/record/286909
date_added: 2021-11-04
date_updated: 2022-07-06
date_updated: 2024-04-09

phantomedicus:
name: PhantoMedicus
Expand All @@ -458,7 +432,7 @@ projects:
tags:
- Predictor
date_added: 2022-07-06
date_updated: 2022-07-06
date_updated: 2024-04-09

paxlib:
name: PAXlib
Expand All @@ -477,7 +451,7 @@ projects:
tags:
- PyTorch
date_added: 2022-07-07
date_updated: 2022-07-07
date_updated: 2024-04-09

byzantine-robust-noniid-optimizer:
name: Byzantine-Robust optimizer
Expand All @@ -496,7 +470,7 @@ projects:
tags:
- PyTorch
date_added: 2022-07-07
date_updated: 2022-07-07
date_updated: 2024-04-09

anomaly-detection:
name: Distributed Homomorphic Anomaly Detection
Expand All @@ -523,4 +497,80 @@ projects:
- Distributed Learning
- Homomorphic Encryption
date_added: 2024-01-03
date_updated: 2024-01-03
date_updated: 2024-04-09

meditron:
name: Meditron
description: >
Open-source medical language model for clinical decision support
type: "Application"
categories:
- "Learning"
applications:
- "Health"
tags:
- Machine Learning
- Natural Language
layman_desc: >
MEDITRON is an AI model designed to help doctors and healthcare
professionals make better decisions by providing access to medical
knowledge. It was trained on high-quality medical information from
research papers and guidelines. Unlike other medical AI models that
are closed-source, MEDITRON's code and training data are open, allowing
transparency and further improvement by researchers. The goal is to
safely bring the benefits of AI to healthcare in an ethical way.
tech_desc: >
MEDITRON is a pair of open-source large language models (LLMs) with 7
and 70 billion parameters, tailored to the medical domain. It was
trained on carefully curated medical data sources, including
peer-reviewed literature and clinical practice guidelines. MEDITRON
outperforms other open-source models and closed models like GPT-3.5 on
medical benchmarks, coming within 5-10% of GPT-4 and Med-PaLM-2.
code:
type: Lab Github
url: https://github.com/epfLLM/meditron
date_last_commit: 2024-04-10
language: Python
license: Apache-2.0
url: https://actu.epfl.ch/news/epfl-s-new-large-language-model-for-medical-knowle/
information:
- type: Paper
title: "MEDITRON-70B: Scaling Medical Pretraining for Large Language Models"
url: https://arxiv.org/abs/2311.16079
date_added: 2024-04-12

# Most common fields:
megatron:
name: Megatron-LLM
description: >
Large language model training library
type: "Application"
categories:
- "Learning"
applications:
- "Infra"
tags:
- Machine Learning
- Natural Language
layman_desc: >
Megatron-LLM is a software library that allows researchers and
developers to train and fine-tune large language models, which are powerful
AI systems that can understand and generate human-like text. It supports
various model architectures and enables training on regular hardware by
distributing the workload across multiple machines. The library offers advanced
features to improve model performance and integrates with popular tools for
tracking training progress and sharing models.
tech_desc: >
Megatron-LLM enables pre-training and fine-tuning of large language
models (LLMs) at scale. It supports architectures like Llama, Llama 2, Code
Llama, Falcon, and Mistral. The library allows training of large models (up to
70B parameters) on commodity hardware using tensor, pipeline, and data
parallelism. It provides features like grouped-query attention, rotary position
embeddings, BF16/FP16 training, and integration with Hugging Face and WandB.
code:
type: Lab Github
url: https://github.com/epfLLM/Megatron-LLM/
date_last_commit: 2023-12-03
language: Python
license: various
date_added: 2024-04-12
8 changes: 3 additions & 5 deletions data/labs.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -359,12 +359,10 @@ labs:
description: >
The Machine Learning and Optimization Laboratory is interested in machine learning, optimization algorithms and text understanding, as
well as several application domains.
contacts:
- name: Aymeric Dieuleveut
email: [email protected]
- name: Mikhail Langovoy
email: [email protected]
url: https://mlo.epfl.ch
information:
- title: Github Repo
url: https://github.com/epfml

MMSPG:
name: Multimedia Signal Processing Group
Expand Down
35 changes: 34 additions & 1 deletion updates/2024-big-update.md
Original file line number Diff line number Diff line change
Expand Up @@ -246,7 +246,40 @@ Linus
- Sent email on 2024-03-22
- MLO - Martin Jaggi

[//]: # " 20 "
- updated last commits
- update contacts
- existing projects
- disco: https://epfml.github.io/disco/#/list - says disco server is not running
- lab page:
- Empty or broken links:
- `AISTATS Paper on Adaptive First-Order Optimization`
- `AISTATS results on Frank-Wolfe, and Optimal Transport`
- https://thijsvogels.nl
- Links to AMLD workshops:
- `NLP track`: `https://appliedmldays.org/tracks`
- `Theory meets Practice Workshop`: `https://appliedmldays.org/workshops/advances-in-ml-theory-meets-practice`
- `PyTorch Workshop`: `https://appliedmldays.org/workshops/pytorch-tutorial`
- `Industry Track`: `https://www.appliedmldays.org/2019/tracks/6`
- `the first instance of the Applied ML Days`: `https://www.appliedmldays.org/2017`
- other links:
- `Electric Vehicle Charging`: `https://www.faculty.ece.vt.edu/kekatos/papers/CDC2016.pdf`
- `Summer school`: `https://www.ds3-datascience-polytechnique.fr/day-4-5-sessions/`
- `Machine Learning Course Started`: `https://mlo.epfl.ch/page-146520.html`
- `Google Focused Research Award`: `https://ai.google/research/outreach/focused-research-awards/`
- `Greedy and Coordinate Algorithms at AISTATS`: `http://www.aistats.org/accepted.html`
- `Welcome Mikhail and Sebastian`: `/mlo/page-135382-en-html/`
- `Machine Learning Course Started`: `mlo.epfl.ch/page-136795.html`
- `news article`: `https://www.inf.ethz.ch/news-and-events/spotlights/semeval2016.html`
- `here`: `http://www.jmlr.org/proceedings/papers/v37/mab15.html`
- `Algorithms in the wild`: `https://blogs.nvidia.com/blog/2018/03/20/big-blue-touts-partnership-with-nvidia-at-ibm-think-confab/`
- `SDSC fellowship`: `https://datascience.ch/academic-projects/`
- `Paper “Primal-Dual Rates and Certificates” at ICML`: `http://jmlr.org/proceedings/papers/v48/dunner16.html`
- new projects
- Meditron
- Megatron
- Also various new github repos - which ones are worth to be shown?
- Sent email on 2024-04-12
[//]: # " 20 "

- RS3LAB - Sanidhya Kashyap
- SACS - Anne-Marie Kermarrec
Expand Down

0 comments on commit 90b9a2b

Please sign in to comment.