Add "smart" layer partition #30
Labels
enhancement
New feature or request
extras
Not directly related to the thesis, low priority
idea
A new idea - may or may not bring improvements
The idea is to implement a benchmark function, e.g., running a predefined ML task with Torch to evaluate the computation capabilities of a specific computer. These capabilities may be quantified, e.g., as the flops.
Then, at model partition (on the starter node or in the
src/llama/prepare_model.py
script), it would be possible to assign a number of layers proportional to this value to each node to split the model in a balanced way.Ideally, this benchmark would be performed once on each node to get the flops, and then this value would be inserted in the config.json and used to initialize the node.
Then, before partitioning the model, the starter node would request this value through a GET.
Example benchmark:
The text was updated successfully, but these errors were encountered: