Is it possible to download a fine-tuned OSS model on Vertex AI ?

Hello, I would like to inquire about the possibility of downloading a fine-tuned model using GCP's computing resources and then performing inference on a local machine. Is it possible to do this with OSS models from Model Garden?

Solved Solved
2 3 247
1 ACCEPTED SOLUTION

Yes, it is possible to download a fine-tuned model from Google Cloud Platform's (GCP) Model Garden and perform inference on a local machine. The Model Garden provides open-source machine learning models, including pre-trained and fine-tuned models, which you can use for various tasks such as image classification, object detection, text generation, etc.

Here's a general outline of the steps you would typically take:

Choose a Model: Select the pre-trained or fine-tuned model that best suits your task from the Model Garden. You can find these models in TensorFlow or PyTorch formats.

Download the Model: Once you've selected the model, download the model files to your local machine. These files usually include the model architecture, pre-trained weights, and any other necessary files.

Set Up Inference Environment: Set up your local machine for performing inference with the downloaded model. This may involve installing the required deep learning framework (e.g., TensorFlow or PyTorch), as well as any other dependencies specific to your task.

Perform Inference: Use the downloaded model files to perform inference on your local machine. Depending on your specific use case, you may need to write some code to load the model, preprocess input data, and interpret the output predictions.

Optimization (Optional): Depending on your performance requirements and constraints, you may need to optimize the inference process for your local machine. This could involve techniques such as model quantization, optimizing input data preprocessing, or using hardware accelerators like GPUs or TPUs if available.

Deployment: If you plan to deploy the model for production use, you may need to package the inference code and model files into a deployable format suitable for your deployment environment.

Remember that some fine-tuned models might have specific licensing or usage restrictions, so make sure to review any applicable terms and conditions before downloading and using the model.

View solution in original post

3 REPLIES 3

Yes, it is possible to download a fine-tuned model from Google Cloud Platform's (GCP) Model Garden and perform inference on a local machine. The Model Garden provides open-source machine learning models, including pre-trained and fine-tuned models, which you can use for various tasks such as image classification, object detection, text generation, etc.

Here's a general outline of the steps you would typically take:

Choose a Model: Select the pre-trained or fine-tuned model that best suits your task from the Model Garden. You can find these models in TensorFlow or PyTorch formats.

Download the Model: Once you've selected the model, download the model files to your local machine. These files usually include the model architecture, pre-trained weights, and any other necessary files.

Set Up Inference Environment: Set up your local machine for performing inference with the downloaded model. This may involve installing the required deep learning framework (e.g., TensorFlow or PyTorch), as well as any other dependencies specific to your task.

Perform Inference: Use the downloaded model files to perform inference on your local machine. Depending on your specific use case, you may need to write some code to load the model, preprocess input data, and interpret the output predictions.

Optimization (Optional): Depending on your performance requirements and constraints, you may need to optimize the inference process for your local machine. This could involve techniques such as model quantization, optimizing input data preprocessing, or using hardware accelerators like GPUs or TPUs if available.

Deployment: If you plan to deploy the model for production use, you may need to package the inference code and model files into a deployable format suitable for your deployment environment.

Remember that some fine-tuned models might have specific licensing or usage restrictions, so make sure to review any applicable terms and conditions before downloading and using the model.

Thank you for your answer, it has been very helpful to me. I have an additional question: If I fine-tune an open-source model using GPU/TPU on GCP, can I also download it afterwards?

A sprobuje dzisiaj pobrac 😄