Skip to content

Latest commit

 

History

History
67 lines (48 loc) · 1.93 KB

fine-tuned-model.md

File metadata and controls

67 lines (48 loc) · 1.93 KB

Fine-tuned model

These steps walk you through downloading the fine-tuned model from Hugging Face and uploads the data into the model GCS bucket for use within the guide for the respective use case.

These prereqs were developed to be run on the playground AI/ML platform. If you are using a different environment the scripts and manifest will need to be modified for that environment.

  • Ensure that your MLP_ENVIRONMENT_FILE is configured

    cat ${MLP_ENVIRONMENT_FILE} && \
    source ${MLP_ENVIRONMENT_FILE}

    You should see the various variables populated with the information specific to your environment.

  • Download the fine-tuned model from Hugging Face and copy it into the GCS bucket.

    NOTE: Due to the limitations of Cloud Shell’s storage and the size of our model we need to run this job to perform the transfer to GCS on the cluster.

    • Get credentials for the GKE cluster

      gcloud container fleet memberships get-credentials ${MLP_CLUSTER_NAME} --project ${MLP_PROJECT_ID}
    • Replace the respective variables required for the job

      MODEL_REPO=gcp-acp/Llama-gemma-2-9b-it-ft
      
      sed \
        -i -e "s|V_KSA|${MLP_MODEL_EVALUATION_KSA}|" \
        -i -e "s|V_BUCKET|${MLP_MODEL_BUCKET}|" \
        -i -e "s|V_MODEL_REPO|${MODEL_REPO}|" \
        manifests/transfer-to-gcs.yaml
    • Deploy the job

      kubectl apply --namespace ${MLP_KUBERNETES_NAMESPACE} \
        -f manifests/transfer-to-gcs.yaml
    • Trigger the wait for job completion (the job will take ~5 minutes to complete)

      kubectl --namespace ${MLP_KUBERNETES_NAMESPACE} wait \
        --for=condition=complete --timeout=900s job/transfer-to-gcs
    • Example output of the job completion

      job.batch/transfer-to-gcs condition met

    NOTE: Return to the respective use case instructions you were following.