I am working on a project where I have to work with 3D arrays, basically human motion data. The arrays are stored in npy files. The goal is to finetune this data on a LLM and be able to use it for analyzing any human motion. But, I realized that using npy files might not be ideal for fine tuning. My advisor said converting them to csv/json might not be ideal since that would change the nature of the arrays, from binary to string. So, my question is, how would I need to prompt this data for fine tuning?
Related topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
What is the text dataset format for fintune LLM? | 2 | 2742 | June 8, 2023 | |
TypeError: Object of type ndarray is not JSON serializable | 0 | 1559 | August 19, 2022 | |
Fine Tuning Format/Structure for data for llma3.1 models | 0 | 63 | October 28, 2024 | |
Best practice for finetune LLM | 0 | 653 | June 21, 2023 | |
DialoGPT fine-tuning dataset format | 3 | 726 | April 27, 2021 |