Is TrueFoundry ML Platform Right for you?

June 5, 2023
Share this post
https://www.truefoundry.com/blog/is-truefoundry-ml-platform-right-for-you
URL

The ML infrastructure landscape is filled with some of the most impressive solutions out there to simplify the ML pipeline. TrueFoundry can be a solution if you relate to some of the problems mentioned below:

Its taking us quite long to get our ML models into production, and there is a lot of dependency among multiple stakeholders

The biggest reason we have found about delays in timelines is dependency between teams and lack of skillsets with different personas. TrueFoundry makes it easy for Data Scientists to train and deploy on Kubernetes using Python. It also allows infra teams to setup security constraints and cost budgets. In most companies we have talked to, the flow of implementation is something like below:

ML workflow with TrueFoundry
ML Workflow is broken
TrueFoundry helps you to reduce the development time by at least 3-4x by empowering Data Scientists to deploy and evaluate the model on their own without reliance on the infra/DevOps team.

With TrueFoundry, the flow is similar to the one below:

ML workflow with TrueFoundry
ML workflow with TrueFoundry

Want to use our standard Kubernetes infrastructure for ML training and deployments

TrueFoundry is Kubernetes native and it works over EKS, AKS, GKE (standard and autopilot clusters) or on-prem clusters. ML requires a few custom things compared to standard software infrastructure - like dynamic node provisioning, GPU support, volumes for faster access, cost budgeting and developer autonomy. We take care of all the nitty-gritty details across the clusters so that you can focus on building the best applications over a state of the art infrastructure.

Data Scientists shouldn't have to deal with infra or YAML

We provide Python APIs - so you never need to interact with YAML. We do provide YAML support also if you want to use it in your CI/CD pipelines. For e.g, using TrueFoundry, you can deploy an inference api using the code below:

service = Service(
   name="fastapi",
   image=Build(
       build_spec=PythonBuild(
           command="uvicorn app:app --port 8000 --host 0.0.0.0",
           requirements_path="requirements.txt",
       )
   ),
   ports=[
       Port(
           port=8000,
           host="<Provide a host value based on your configured domain>"
       )
   ],
   resources=Resources(
       cpu_request=0.5,
       cpu_limit=1,
       memory_request=1000,
       memory_limit=1500
   ),
   env={
       "UVICORN_WEB_CONCURRENCY": "1",
       "ENVIRONMENT": "dev"
   }
)
service.deploy(workspace_fqn="tfy-cluster/my-workspace")

Need ML infrastructure that doesn't require us to move data outside of our Cloud

TrueFoundry gets deployed entirely on your own Kubernetes cluster. The data stays in your own VPC, docker images get saved in your own docker registry and all the models stay in your own blob storage system. You can read more about the TrueFoundry architecture here.

Autoscaling of models is quite slow due to the download time of ML models

Kubernetes usually supports autoscaling using HPA based on CPU and memory. However, for ML workloads, autoscaling based on request counts is a lot better in many cases. Another challenge in autoscaling can be the high startup time of models because of large image sizes and model download times. Truefoundry solves these problems by providing container startup time in seconds, caching of models for faster loading and providing faster inference times.

Want to use the power of LLMs for our business, but we cannot let the data out of our environment

Can we use some open-source LLM models?

TrueFoundry allows you to deploy and finetune the open-source LLMs on your own infrastructure. We have already figured out the best settings for the most common open-source models so that you can train and deploy them at the optimal settings and lowest cost.

Want to allow all my developers to try out different LLMs quickly

We host an internal LLM playground where you can decide which LLMs you want to whitelist for the company developers, including internally hosted ones and different developers can experiment with the internal data. Here is a quick video on same:

Fine tune & deploy LLMs on your cloud

Want to provide Jupyter Notebooks to Data Scientists on a self-serve basis in a multi-tenant cost-optimized way

Jupyter Notebooks are essential to the Data Scientist's daily development cycle. Running Jupyter Notebooks locally on one's own machine is not always an option because of the following reasons:

  1. We might require higher resources which might not be available on a local laptop
  2. Data access might not be allowed in the local environment.

We have put in a lot of effort to seamlessly run Jupyter Notebooks on Kubernetes. Jupyter Notebooks on TrueFoundry provide the following benefits compared to JupyterLab or Kubeflow Notebooks:

  1. Fast startup time of Notebooks (under 10 seconds)
  2. Auto-stopping feature which shuts down the notebooks after certain configurable period of inactivity. This reduces cost since a data scientist might only work 8 hours a day. This will reduce cost by around 60% compared to running Jupyter on EC2 instances.
  3. Persistence of environment, data and Python dependencies across restarts.
  4. Ability to add dependencies dynamically to the base image.
  5. Ability to share notebooks with other team members.
  6. Ability to configure dataset access using service accounts instead of keys/passwords.

Want to keep track of all models inside the company in one place, and figure out which ones are deployed in what environment?

TrueFoundry provides a model registry that can track which models are in what stage and the schema and API of all the models in the registry.

Want to mirror or split traffic to my new version of the model so that we can test it on online traffic before rolling it out completely?

TrueFoundry allows splitting or mirroring traffic from one model to another. This is especially useful when you want to test a new model version on live traffic for some time before rolling it to production. Truefoundry also supports canary and blue-green rollout strategies in model deployment.

Want to use hardware and compute across clouds and on-prem. How do I connect them so that developers can seamlessly move workloads from one environment to other?

We have put in a lot of effort to make sure we take care of the nitty-gritty differences of the Kubernetes clusters across clouds. Developers can write and deploy the same code in any environment without worrying about the underlying infrastructure. We take care of checking if underlying components of Kubernetes are installed, checking incompatible migrations and informing developers accordingly.

Incurring a lot of cost on our ML infra and its becoming difficult to track and reduce it.

We expose the cost visibility of services to developers and provide insights to reduce the cost. All our current customers have seen atleast 30% cost reduction after adopting truefoundry.

TrueFoundry is a ML Deployment PaaS over Kubernetes to speed up developer workflows while allowing them full flexibility in testing and deploying models while ensuring full security and control for the Infra team. Through our platform, we enable Machine learning Teams to deploy and monitor models in 15 minutes with 100% reliability, scalability, and the ability to roll back in seconds - allowing them to save cost and release Models to production faster, enabling real business value realisation.  

Discover More

September 7, 2023

Machine Learning Tool Integrations #3 Label Studio for All your Labelling Needs

Engineering and Product
Use Cases
April 13, 2023

Understanding ChatGPT plugins

Engineering and Product
Use Cases
LLMs & GenAI
June 15, 2023

Machine Learning Tool Integrations #2 DVC for your Data Versioning

Engineering and Product
Use Cases
May 3, 2023

Machine Learning Platform Integrations #1: Weights & Biases

Use Cases
Engineering and Product

Related Blogs

No items found.

Blazingly fast way to build, track and deploy your models!

pipeline