- Here is a link to the API guide which goes over all the jobs that are possible to run on ThingWorx Analytics. For descriptions on the Prediction Models I would Check out the Prediction Mode Generation Section and Scoring sections of the Document.
- ThingWorx Analytics self-validates its predictions by putting aside a relevant set of data (holdout set) to verify against. Prediction accuracy is qualified on several performance metrics, primarily RMSE and Pearson Correlation. So the model accuracy was so high because that is how the hold out set performed when it was scored against the generated model.
Let me know if you have any further questions.
Now as far understanding the outcome is concerned in analytics is on a case by case basis so no general interpretations could be provided for specific results. However, a broad definition of the concepts you mentioned would be :
-The Goal: A “Goal” in ThingWorx Analytics is the same as an outcome variable or “dependent variable”.
-The features or Independent Variables : is the information available to help us predict or explain an outcome whereas the dependent variable (Goal) is the outcome we are trying to predict or understand.
-The SignalsA “Signal” represents the strength or weakness a particular input (independent variable) has in relation to the Topic (dependent variable)
-Model: This is a mathematical Equation representing the relationship between the Goal and the independent Variables created through ThingWorx Analytics Machine learning algorithms. Its accuracy is measured through different methods like RMSE or Pearson correlations. The Model is trained based on the historical Data available in the Dataset you provide
For the interpretation of the results, you would look at the signals first. These would provide a first view on the descriptive analytics side towards the goal you are trying to predict. ThingWorx Analytics would not create those signals but rather reveal them through trends in your Dataset.
And for your model I would suggest to remove the feature with the strongest signal from the Dataset (through the use of an exclusive filter). After that try training your model again. If you are obtaining a "too good to be true" model this feature might have biased correlation to the goal.(although it is not necessarily the case ).
I hope this helps.