As I just read from Appwrite Twitter account, you're cooking up a new runtime which is designed to perform machine learning tasks. Because Appwrite is scalable with Docker Swarm, does that mean that if a ML task is given to the swarm, then the task is automatically distributed across the swarm nodes to speed up the learning process? Technically, would this mean that we are able to easily scale ML tasks across multiple server machines? Also, would the ML runtime support distributed training of large language models like LLaMA?
Hi - The runtime will be basically a Python runtime with all system libraries needed to do machine learning things so it becomes easier for ML devs to work with it. Ofcourse it would be open to gather community feedback to see how we want to improve it and make it better. That being said, No, currently in first iteration of release, ML runtime will not support distributed training of large language models like LLaMA, but in next iterations (if community feedbacks suggest) it will be added.
It will scale exactly the same as any other Appwrite Function BUT it will proper access to host machine GPU, which is almost necessary for machine learning.
Thank you Jyoti! 🙂 I would strongly support the ability to distribute LLM training over multiple Appwrite swarm nodes.
Thanks for the feedback! We will definitely consider it ❤️
Recommended threads
- Python-ml-3.12
Hi, when this runtime will be available for the cloud. I could see the runtimes repo has this added to Runtimes.php but I'm not able to use it to run locally or...
- Function Building show 500 Error
Region: sfo Temple: Starter Function Runtime: Nodejs 22 What happened? Before 2 minutes it show ‘server error.","code":500,"version":"0.7.24"}’ after it show c...
- Function for long running task (other 60...
My function need long running other 60 seconds. But async mode return empty body. How to deal with that please ?