You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Dec 21, 2023. It is now read-only.
I've gathered approximately 5 GB of time series data from my Apple Watch, categorizing activities into two classes with proper labeling (such as walking or not).
The model was trained successfully, and the attached photo shows the best model's results.
However, upon testing with an unseen dataset, I find the performance unsatisfactory.
Despite achieving high accuracy, F1, precision, and recall values during training, deploying the model for real-time activity classification on the Apple Watch yields disappointing results.
I only see log_loss couldn't see valid/train loss graph.
While I understand my dataset might not be considered real Big Data, I believe that 5 GB of data should suffice for distinguishing between two activities, such as walking or not.
I think metrics like accuracy, recall, precision, and F1 aren't adequate to take this model in practical deployment. How do I handle this challenge to build a trustful model?
How can I ensure the reliability of my model in such a scenario?
Ps: Tucreate: Training, validation, and testing pipeline was done considering human activity documents.
Subsequently, the model deployment encountered unsatisfactory results on the watch app.
The text was updated successfully, but these errors were encountered:
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Hi,
I've gathered approximately 5 GB of time series data from my Apple Watch, categorizing activities into two classes with proper labeling (such as walking or not).
The model was trained successfully, and the attached photo shows the best model's results.
However, upon testing with an unseen dataset, I find the performance unsatisfactory.
Despite achieving high accuracy, F1, precision, and recall values during training, deploying the model for real-time activity classification on the Apple Watch yields disappointing results.
I only see log_loss couldn't see valid/train loss graph.
While I understand my dataset might not be considered real Big Data, I believe that 5 GB of data should suffice for distinguishing between two activities, such as walking or not.
I think metrics like accuracy, recall, precision, and F1 aren't adequate to take this model in practical deployment. How do I handle this challenge to build a trustful model?
How can I ensure the reliability of my model in such a scenario?
Ps: Tucreate: Training, validation, and testing pipeline was done considering human activity documents.
Subsequently, the model deployment encountered unsatisfactory results on the watch app.
The text was updated successfully, but these errors were encountered: