What do you mean by "compute"? Do you want to graph it? Factor it? Calculate it's function given a set of points that lie on it? If you're looking to compute the function given three points that fall on the parabola, then I have just the code for you. If you're given three points, (x1, y1), (x2, y2) and (x3, y3), then you can compute the coefficients of your quadratic equation like this: a = (y1 * (x2 - x3) + y2 * (x3 - x1) + y3 * (x1 - x2)) / (x1 * x1 * (x2 - x3) + x2 * x2 * (x3 - x1) + x3 * x3 * (x1 - x2)) b = (y1 - y2) / (x1 - x2) - a * (x1 + x2); c = y1 - (x1 * x1) * a - x1 * b; You now can calculate the y co-ordinate of any point given it's x co-ordinate by saying: y = a * x * x + b * x + c;
y(i) = a + b1.x1(i) + b2.x2(i) + b3.x3(i) + ... + bk.xk(i) + e(i)where i = 1, 2, ... n are n observations ofthe independent variables x1, x2, ... xk,y is the dependent variablea and the b are regression parameters.The e are independent, identically distributed random variables (representing the error).
You need two variables. Make them x and y make y your dependent variable and x your independent variable. y=x1 this is a line y=4x1 this is a line y=4x1+3 this is a line y=x2 this is a curve y=x3 this is a curve make sure your x value has no power to it except 1 x=x1
multiple correlation: Suppose you calculate the linear regression of a single dependent variable on more than one independent variable and that you include a mean in the linear model. The multiple correlation is analogous to the statistic that is obtainable from a linear model that includes just one independent variable. It measures the degree to which the linear model given by the linear regression is valuable as a predictor of the independent variable. For calculation details you might wish to see the wikipedia article for this statistic. partial correlation: Let's say you have a dependent variable Y and a collection of independent variables X1, X2, X3. You might for some reason be interested in the partial correlation of Y and X3. Then you would calculate the linear regression of Y on just X1 and X2. Knowing the coefficients of this linear model you would calculate the so-called residuals which would be the parts of Y unaccounted for by the model or, in other words, the differences between the Y's and the values given by b1X1 + b2X2 where b1 and b2 are the model coefficients from the regression. Now you would calculate the correlation between these residuals and the X3 values to obtain the partial correlation of X3 with Y given X1 and X2. Intuitively, we use the first regression and residual calculation to account for the explanatory power of X1 and X2. Having done that we calculate the correlation coefficient to learn whether any more explanatory power is left for X3 to 'mop up'.
This is actual question SUPPOSE X1 X2 X3, Xn form a random sample from a population with density function f(x,y) = 1/y where 0<x<y,y>0 where y is unknown parameter .let T=max(X1,X2,....Xn) show that Y (estimate) ... Y=(1+1/n) is unbiased estimator of Y?
Any model can be linear/nonlinear. Linearity can be in parameters or in variables. In Y=a+ b*x1 + c*x2 + d*x3 the model is linear in both parameters (b,c,d) and variables(x1,x2,x3) In Y=a+ (b+c)x1 + c*x2 + d*x3 the model is nonlinear in parameters (b,c,d) and linear in variables(x1,x2,x3) In Y=a+ bx1 + c*x2*x3 + d*x3 the model is linear in parameters (b,c,d) and nonlinear in variables(x1,x2,x3) In Y=a+ bx1 + c*x2*x3 + exp(b+d)*x3 the model is nonlinear in parameters (b,c,d) and nonlinear in variables(x1,x2,x3)
What do you mean by "compute"? Do you want to graph it? Factor it? Calculate it's function given a set of points that lie on it? If you're looking to compute the function given three points that fall on the parabola, then I have just the code for you. If you're given three points, (x1, y1), (x2, y2) and (x3, y3), then you can compute the coefficients of your quadratic equation like this: a = (y1 * (x2 - x3) + y2 * (x3 - x1) + y3 * (x1 - x2)) / (x1 * x1 * (x2 - x3) + x2 * x2 * (x3 - x1) + x3 * x3 * (x1 - x2)) b = (y1 - y2) / (x1 - x2) - a * (x1 + x2); c = y1 - (x1 * x1) * a - x1 * b; You now can calculate the y co-ordinate of any point given it's x co-ordinate by saying: y = a * x * x + b * x + c;
y(i) = a + b1.x1(i) + b2.x2(i) + b3.x3(i) + ... + bk.xk(i) + e(i)where i = 1, 2, ... n are n observations ofthe independent variables x1, x2, ... xk,y is the dependent variablea and the b are regression parameters.The e are independent, identically distributed random variables (representing the error).
You need two variables. Make them x and y make y your dependent variable and x your independent variable. y=x1 this is a line y=4x1 this is a line y=4x1+3 this is a line y=x2 this is a curve y=x3 this is a curve make sure your x value has no power to it except 1 x=x1
This is actual question SUPPOSE X1 X2 X3, Xn form a random sample from a population with density function f(x,y) = 1/y where 0<x<y,y>0 where y is unknown parameter .let T=max(X1,X2,....Xn) show that Y (estimate) ... Y=(1+1/n) is unbiased estimator of Y?
multiple correlation: Suppose you calculate the linear regression of a single dependent variable on more than one independent variable and that you include a mean in the linear model. The multiple correlation is analogous to the statistic that is obtainable from a linear model that includes just one independent variable. It measures the degree to which the linear model given by the linear regression is valuable as a predictor of the independent variable. For calculation details you might wish to see the wikipedia article for this statistic. partial correlation: Let's say you have a dependent variable Y and a collection of independent variables X1, X2, X3. You might for some reason be interested in the partial correlation of Y and X3. Then you would calculate the linear regression of Y on just X1 and X2. Knowing the coefficients of this linear model you would calculate the so-called residuals which would be the parts of Y unaccounted for by the model or, in other words, the differences between the Y's and the values given by b1X1 + b2X2 where b1 and b2 are the model coefficients from the regression. Now you would calculate the correlation between these residuals and the X3 values to obtain the partial correlation of X3 with Y given X1 and X2. Intuitively, we use the first regression and residual calculation to account for the explanatory power of X1 and X2. Having done that we calculate the correlation coefficient to learn whether any more explanatory power is left for X3 to 'mop up'.
the equation is A= y2-y2/x3-x2 after that you find the y-intercept by doing, b= y1+y2+y3-A(x1+x2+x3)/3
The step-work involved in proving this would be too long and detailed to show here, but the end result is this: We're given three verticies, defined by the points (x1, y1), (x2, y2), and (x3, y3). We want to use them to define a parabola in this format: f(x) = ax2 + bx + c We can find our a, b and c coefficients with the following equations: a = [y1(x2 - x3) + y2(x3 - x1) + y3(x1 - x2)] / [x12(x2 - x3) + x22(x3 - x1) + x32(x1 - x2)] b = (y1 - y2) / (x1 - x2) - a(x1 + x2) c = y1 - a(x12) - b(x1)
The x-coordinate of the centroid is the arithmetic mean of the x-coordinates of the three vertices. And likewise, the y-coordinate of the centroid is the arithmetic mean of the y-coordinates of the three vertices. Thus, if A = (x1, y1), B = (x2, y2) and C = (x3, y3) then the coordinates of the centroid, G = [(x1,+ x2 + x3)/3, (y1 + y2 + y3)/3].
This question can only be answered if the probability distribution functions of X1, X2 and X3 are known. They are not and so the question cannot be answered.
For three points, (x1,y1), (x2,y2) & (x3,y3), you can set up 3 distance equations with variables x, y & z: z^2 = (x-x1)^2 + (y-y1)^2 z^2 = (x-x2)^2 + (y-y2)^2 z^2 = (x-x3)^2 + (y-y3)^2 3 equations and 3 variables....Solve away! z is your distance. x & y are the coordinates of the equidistant point.
x3