Calculating error between non-linear dataset and model

My model is a rotated ellipse, for which I have already worked out parameters for the ellipse which best fits my data using the ellipse fit here: https://uk.mathworks.com/matlabcentral/fileexchange/3215-fit_ellipse .
I have a set of pseudo-elliptical data points. I would like to calculate the error between my data and the model; however, the data is not at regular intervals and I am unsure how to calculate the error if I do not have points which correspond. I suppose I should be able to calculate the y-values of the ellipse for a given x value of my data - how would I go about this?
Apologies if this is just maths but if anyone can help that would be great!

Answers (0)

Asked:

on 20 Mar 2019

Edited:

on 20 Mar 2019

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!