As I just read from Appwrite Twitter account, you're cooking up a new runtime which is designed to perform machine learning tasks. Because Appwrite is scalable with Docker Swarm, does that mean that if a ML task is given to the swarm, then the task is automatically distributed across the swarm nodes to speed up the learning process? Technically, would this mean that we are able to easily scale ML tasks across multiple server machines? Also, would the ML runtime support distributed training of large language models like LLaMA?
Hi - The runtime will be basically a Python runtime with all system libraries needed to do machine learning things so it becomes easier for ML devs to work with it. Ofcourse it would be open to gather community feedback to see how we want to improve it and make it better. That being said, No, currently in first iteration of release, ML runtime will not support distributed training of large language models like LLaMA, but in next iterations (if community feedbacks suggest) it will be added.
It will scale exactly the same as any other Appwrite Function BUT it will proper access to host machine GPU, which is almost necessary for machine learning.
Thank you Jyoti! 🙂 I would strongly support the ability to distribute LLM training over multiple Appwrite swarm nodes.
Thanks for the feedback! We will definitely consider it ❤️
Recommended threads
- dynamic key missing scopes for database ...
Here are the scopes listed, I get permission errors for reading row and document. Appears to be missing since last time i checked. Database 6 Scopes policies....
- Worker functions stuck on "Fetched 0 fun...
Appwrite Version: 1.9.0 Bug Description: The appwrite-worker-functions container gets stuck in an infinite loop logging "Fetched 0 functions..." while scheduled...
- I am using s3 for app storage but is it ...
_APP_STORAGE_DEVICE=s3 puts everything to the s3 storage but i need to be able to keep the function builds and site in the local and not waste the cloud storage...