Fleets
Fleets is Zerve's built-in feature for parallel processing. With a single line of code, users can parallelize code to run across multiple blocks.
Fleets in Zerve provide an easy-to-use parallel processing feature. By leveraging the spread()
function, you can initialize parallel tasks with a simple Python command. The spread()
function takes a list as input, fanning out multiple compute blocks corresponding to each element of the list. After execution, the gather()
function collects all results into a list for subsequent processing.
ML Pipeline in Fleets:
Parallel processing in Zerve's fleets can significantly enhance the efficiency of building machine learning (ML) pipelines. Once data processing is complete, multiple algorithms can be applied simultaneously to different subsets of data. This approach allows data scientists to explore various ML models in parallel, quickly identifying which algorithms yield the best results for a given dataset.
Hyperparameter tuning, an essential step in optimizing ML models, can also benefit greatly from the parallel processing capabilities of fleets. By running hyperparameter tuning in fleets, multiple combinations of parameters can be tested simultaneously. This not only speeds up the process but also helps in discovering the most effective configuration to enhance model performance across varied datasets.





Data Processing with Fleets:
Fleets in Zerve also enable faster data processing by allowing category-level processing within a dataset column. For instance, when dealing with large datasets that contain categorical variables, fleets can process each category in parallel. This strategy reduces processing time substantially, enabling quicker insights and data-driven decisions.


LLM Evaluation with Fleets:
Furthermore, testing large language models (LLMs) with multiple prompts across different models can become significantly more efficient using fleets. By distributing the workload across different compute blocks, teams can evaluate LLMs against a variety of inputs in a fraction of the time it would take to perform sequential testing. This parallel approach not only saves time but also provides a more comprehensive understanding of how different models perform under diverse conditions.


Last updated