Skip to content

FlexFlow Serve: Low-Latency, High-Performance LLM Serving

License

Notifications You must be signed in to change notification settings

Marsella8/FlexFlow

This branch is 23 commits ahead of, 4 commits behind flexflow/flexflow-train:master.

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Feb 9, 2025
97227c6 · Feb 9, 2025
Jan 15, 2025
Jan 31, 2025
Feb 1, 2025
Jan 20, 2025
Feb 1, 2025
Mar 16, 2024
Jan 20, 2025
Feb 9, 2025
Oct 6, 2023
May 13, 2022
Aug 4, 2023
Dec 12, 2022
Jan 20, 2025
Jun 4, 2024
Jan 15, 2025
Mar 16, 2024
Jul 6, 2023
Jan 17, 2025
Oct 6, 2022
Sep 14, 2024
Jan 25, 2021
Jan 20, 2025
Jun 30, 2023
Jan 8, 2020
Jan 20, 2025
Jan 20, 2025
Jan 20, 2025
Feb 1, 2025

Repository files navigation

FlexFlow Train

clang-format Check per-lib-checks shell-check Documentation Status

Warning

The FlexFlow repository has been split into separate flexflow-train and flexflow-serve repositories. You are currently viewing flexflow-train. For anything inference/serving-related, go to flexflow-serve.

FlexFlow Train is a deep learning framework that accelerates distributed DNN training by automatically searching for efficient parallelization strategies.

Contributing

Please let us know if you encounter any bugs or have any suggestions by submitting an issue.

For instructions on how to contribute code to FlexFlow Train, see CONTRIBUTING.md.

We welcome all contributions to FlexFlow Train from bug fixes to new features and extensions.

Citations

The Team

FlexFlow Train is developed and maintained by teams at CMU, Facebook, Los Alamos National Lab, MIT, Stanford, and UCSD (alphabetically).

License

FlexFlow Train uses Apache License 2.0.

About

FlexFlow Serve: Low-Latency, High-Performance LLM Serving

Resources

License

Code of conduct

Stars

Watchers

Forks

Packages

No packages published

Languages

  • C++ 69.8%
  • Cuda 13.2%
  • Python 13.1%
  • Shell 1.6%
  • CMake 1.1%
  • C 0.8%
  • Other 0.4%