New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Vertex AI pipeline - IndexError: Invalid key: 0 is out of bounds for size 0 #2813
Comments
@kk2491 can you please let me know which notebook you ran? |
Hi @gericdong I am using the below notebook
Thank you, |
@genquan9: can you please assist with this? Thank you. |
If you do training from HF datasets, you can input sth like: timdettmers/openassistant-guanaco directly. but, if you use dataset json stored in gcs, you should use the format as: |
The team is verifying the notebook with pipelines again. |
@genquan9 @gericdong Sorry to bother you. Did you get chance to look into the above issue? Thank you, |
Hi @kk2491, I was able to reproduce the issue. Please try again but set the evaluation_limit to 100. |
@jismailyan-google Thanks for the suggestion. Thank you, |
@jismailyan-google Looks like the notebook for vertex-ai pipeline has been removed. |
@genquan9 @gericdong Did you get chance to look into the above issue? Thank you, |
Hi @kk2491, I was able to get the tuning completed with your dataset. Also, please note the updated
Let me know if this works. |
@jismailyan-google I tried again this time with Vertex GUI (looks like the notebook for fine-tune with vertex-ai has been removed). Now I am 100% sure that I am doing some silly mistake here.. !!! |
I am running into the same error when trying to specify a custom dataset: # Hugging Face dataset name or gs:// URI to a custom JSONL dataset.
dataset_name = "gs://llama-fine-tuning/training_data.jsonl" # @param {type:"string"}
# Name of the dataset column containing training text input.
instruct_column_in_dataset = "text" # @param {type:"string"}
# Optional. Template name or gs:// URI to a custom template.
template = "" # @param {type:"string"} I haven't looked, but I suspect that the image running the I saw the following comment by @genquan9:
I haven't tried this yet, but it seems that the instruct lora task needs to account for gs:// URI somehow. Does it? |
@Joshwani-broadcom Here is how I was able to fix the error. (Worth giving a try, if not tried yet)
Looks like all of your samples are getting dropped due to one of the above reasons. You can also find more details here. Kindly let me know if you face any other issues. Thank you, |
Thank you @kk2491 - Is it true that you are using a huggingface dataset? Did you ever find success using a gs:// uri in the notebook like this: dataset_name = "gs://llama-fine-tuning/training_data.jsonl" ? |
Yea initially I tried with huggingface dataset and got it working. Later with the same dataset I migrated to Google bucket, it worked as expected. Thank you, |
Expected Behavior
The fine-tuning of the foundation model should complete without any issues.
Actual Behavior
The fine-tuning step gets terminated. The details provided below:
Training framework - Google collab
Model used - Llama2-7B
Fine-tuning method - PEFT
Number of samples in Training Set - 100
Number of samples in Eval Set - 20
Format of the training data - jsonl
Example sample is given below -
Vertex pipeline parameters :
When I execute the training process, I get the below error:
Can you please help in understanding the below question?
Steps to Reproduce the Problem
Specifications
The text was updated successfully, but these errors were encountered: