How to estimate $alpha$ given a discrete sequence $a(n) = n^alpha+ n^1-alpha+b$.

The name of the pictureThe name of the pictureThe name of the pictureClash Royale CLAN TAG#URR8PPP











up vote
0
down vote

favorite












I have a data sequence of length $10^6$ which I know it can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$ ($b$ is an unknown constant, $0<alpha <1$). But how to estimate $alpha$ here?



Update:



Sorry, I made some mistakes above. I sincerely apologize for the mistakes. I revise the problem and explain my concerns more in this update.



The data sequence is generated from simulations of one algorithm. It should have the order $O(n^max(alpha, 1-alpha))$. I know it can be approximately modeled as $a(n) = b_1n^alpha+ b_2n^1-alpha+b_3$ ($b_1,b_2,b_3$ are unknown constants) when $n$ is suffciently large.



I only care about the order $max(alpha, 1-alpha)$ to justify the performance of the algorithm.



I have tried using $frac1t_1-t_0sum_t=t_0+1^t_1log_2 fraca(2t)a(t)$. The problem is the graph I got does not get the minimum at $alpha = 0.5$. How to eliminate the numerical errors?



Here's a link of the new post of the problem.



enter image description here







share|cite|improve this question

















  • 1




    You can consider $0<alpha<frac12$.
    – Yves Daoust
    Jul 23 at 8:26










  • Is the data noisy or smooth ? Do the initial terms match the model ?
    – Yves Daoust
    Jul 23 at 9:06










  • What are you after, an estimate of the asymptotic behavior or just a best fit to the million points ?
    – Yves Daoust
    Jul 23 at 9:33















up vote
0
down vote

favorite












I have a data sequence of length $10^6$ which I know it can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$ ($b$ is an unknown constant, $0<alpha <1$). But how to estimate $alpha$ here?



Update:



Sorry, I made some mistakes above. I sincerely apologize for the mistakes. I revise the problem and explain my concerns more in this update.



The data sequence is generated from simulations of one algorithm. It should have the order $O(n^max(alpha, 1-alpha))$. I know it can be approximately modeled as $a(n) = b_1n^alpha+ b_2n^1-alpha+b_3$ ($b_1,b_2,b_3$ are unknown constants) when $n$ is suffciently large.



I only care about the order $max(alpha, 1-alpha)$ to justify the performance of the algorithm.



I have tried using $frac1t_1-t_0sum_t=t_0+1^t_1log_2 fraca(2t)a(t)$. The problem is the graph I got does not get the minimum at $alpha = 0.5$. How to eliminate the numerical errors?



Here's a link of the new post of the problem.



enter image description here







share|cite|improve this question

















  • 1




    You can consider $0<alpha<frac12$.
    – Yves Daoust
    Jul 23 at 8:26










  • Is the data noisy or smooth ? Do the initial terms match the model ?
    – Yves Daoust
    Jul 23 at 9:06










  • What are you after, an estimate of the asymptotic behavior or just a best fit to the million points ?
    – Yves Daoust
    Jul 23 at 9:33













up vote
0
down vote

favorite









up vote
0
down vote

favorite











I have a data sequence of length $10^6$ which I know it can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$ ($b$ is an unknown constant, $0<alpha <1$). But how to estimate $alpha$ here?



Update:



Sorry, I made some mistakes above. I sincerely apologize for the mistakes. I revise the problem and explain my concerns more in this update.



The data sequence is generated from simulations of one algorithm. It should have the order $O(n^max(alpha, 1-alpha))$. I know it can be approximately modeled as $a(n) = b_1n^alpha+ b_2n^1-alpha+b_3$ ($b_1,b_2,b_3$ are unknown constants) when $n$ is suffciently large.



I only care about the order $max(alpha, 1-alpha)$ to justify the performance of the algorithm.



I have tried using $frac1t_1-t_0sum_t=t_0+1^t_1log_2 fraca(2t)a(t)$. The problem is the graph I got does not get the minimum at $alpha = 0.5$. How to eliminate the numerical errors?



Here's a link of the new post of the problem.



enter image description here







share|cite|improve this question













I have a data sequence of length $10^6$ which I know it can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$ ($b$ is an unknown constant, $0<alpha <1$). But how to estimate $alpha$ here?



Update:



Sorry, I made some mistakes above. I sincerely apologize for the mistakes. I revise the problem and explain my concerns more in this update.



The data sequence is generated from simulations of one algorithm. It should have the order $O(n^max(alpha, 1-alpha))$. I know it can be approximately modeled as $a(n) = b_1n^alpha+ b_2n^1-alpha+b_3$ ($b_1,b_2,b_3$ are unknown constants) when $n$ is suffciently large.



I only care about the order $max(alpha, 1-alpha)$ to justify the performance of the algorithm.



I have tried using $frac1t_1-t_0sum_t=t_0+1^t_1log_2 fraca(2t)a(t)$. The problem is the graph I got does not get the minimum at $alpha = 0.5$. How to eliminate the numerical errors?



Here's a link of the new post of the problem.



enter image description here









share|cite|improve this question












share|cite|improve this question




share|cite|improve this question








edited Jul 24 at 15:59
























asked Jul 23 at 7:21









Jimmy Kang

647




647







  • 1




    You can consider $0<alpha<frac12$.
    – Yves Daoust
    Jul 23 at 8:26










  • Is the data noisy or smooth ? Do the initial terms match the model ?
    – Yves Daoust
    Jul 23 at 9:06










  • What are you after, an estimate of the asymptotic behavior or just a best fit to the million points ?
    – Yves Daoust
    Jul 23 at 9:33













  • 1




    You can consider $0<alpha<frac12$.
    – Yves Daoust
    Jul 23 at 8:26










  • Is the data noisy or smooth ? Do the initial terms match the model ?
    – Yves Daoust
    Jul 23 at 9:06










  • What are you after, an estimate of the asymptotic behavior or just a best fit to the million points ?
    – Yves Daoust
    Jul 23 at 9:33








1




1




You can consider $0<alpha<frac12$.
– Yves Daoust
Jul 23 at 8:26




You can consider $0<alpha<frac12$.
– Yves Daoust
Jul 23 at 8:26












Is the data noisy or smooth ? Do the initial terms match the model ?
– Yves Daoust
Jul 23 at 9:06




Is the data noisy or smooth ? Do the initial terms match the model ?
– Yves Daoust
Jul 23 at 9:06












What are you after, an estimate of the asymptotic behavior or just a best fit to the million points ?
– Yves Daoust
Jul 23 at 9:33





What are you after, an estimate of the asymptotic behavior or just a best fit to the million points ?
– Yves Daoust
Jul 23 at 9:33











3 Answers
3






active

oldest

votes

















up vote
2
down vote













Rather than a two-parameter least-squares, you might note that your formula implies
$$a_i+1-a_i = n_i+1^alpha - n_i^alpha + n_i+1^1-alpha - n_i^1-alpha$$
Find the $alpha$ that minimizes the sum of squared errors here, then for $b$ take the average of $a_i - (n_i^alpha + n_i^1-alpha)$.






share|cite|improve this answer





















  • For sure ! Much simpler than my junk !
    – Claude Leibovici
    Jul 23 at 8:38










  • If I am right, $n_i=i$, there is no need to index.
    – Yves Daoust
    Jul 23 at 8:40










  • @YvesDaoust. Not necessary since you can use, say, $n=10,20,30,cdots$.
    – Claude Leibovici
    Jul 23 at 8:53










  • @ClaudeLeibovici: right, but then the formula must be with $a_n_i$.
    – Yves Daoust
    Jul 23 at 8:56










  • Many thanks. I have tried your answer. I thought the data can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$. Actually, I have $5$ sequences, each should be modeled with $alpha = 1/1.9,1/2.0, 1/2.1, 1/2.2, 1/2.5$ respectively. However, the strange thing is the partial derivative of the squared errors achieves $0$ at $alpha = 0.5$ for all sequences. I'm still trying to figure out what happens in the data.
    – Jimmy Kang
    Jul 23 at 20:20

















up vote
2
down vote













If this helps:



If $a_0$ or $a_1-2$ are available and give reasonable estimates of $b$, for large $n$ we have



$$fraca_n-hat bsqrt nsim n^alpha-1/2+n^1/2-alpha=2coshleft(left(frac12-alpharight)log nright)$$



and



$$alphasimfrac12-dfractextarcoshdfraca_n-hat b2sqrt nlog n.$$



This can be used f.i. as a starting value of an iterative method.






share|cite|improve this answer























  • This seems to be a quite good idea, for sure !
    – Claude Leibovici
    Jul 23 at 9:17










  • Sorry for the late reply. Data is noisy. The initial terms do not match the model. My concern is just to estimate the asymptotic behavior. The data should be $O(n^max(alpha,1-alpha))$. I want to get the order of the sequence, i.e. $max(alpha,1-alpha)$.
    – Jimmy Kang
    Jul 23 at 20:01











  • @JimmyKang: why do you add a constant term ?
    – Yves Daoust
    Jul 23 at 20:41










  • I revised the problem just now. Sorry for the confusion caused.
    – Jimmy Kang
    Jul 23 at 21:21

















up vote
1
down vote













If we look at the problem as a regression, you have $p$ data points $(n_i,a_i)$ and you want to adjust the model
$$a=n^alpha+n^1-alpha+b$$ In the least square sense, you need to minimize
$$SSQ=sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)^2$$ Computing the partial derivatives,we have
$$fracpartial SSQpartial b=2sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)=0tag 1$$
$$fracpartial SSQpartial a=2sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)left(n_i^alpha-n_i^1-alpharight)log(n_i)=0 tag 2$$ which is hard to solve.



However, from $(1)$, you can get $b$ as a function of $alpha$
$$b(alpha)=-frac 1p sum_i=1^p left(n_i^alpha+n_i^1-alpha -a_iright)$$ and then $(2)$ is "just" an equation in $alpha$.



The simplest would be to plot equation $(2)$ for $0 leq alpha leq 0.5$ and see where it does cancels. Zoom more and more to have more accurate results; when your accuracy has been reached, recompute $b$.



All of that can be done using Excel.



Edit



For illustration purposes, let us use the following data set
$$left(
beginarrayccc
i & n_i & a_n_i \
1 & 10 & 10 \
2 & 20 & 14 \
3 & 30 & 16 \
4 & 40 & 19 \
5 & 50 & 21 \
6 & 60 & 23 \
7 & 70 & 25 \
8 & 80 & 26 \
9 & 90 & 28 \
10 & 100 & 30
endarray
right)$$ A first run, using $Delta alpha=0.05$ would give
$$left(
beginarraycc
alpha & (2) \
0.05 & -36241.2 \
0.10 & -19913.1 \
0.15 & -10372.8 \
0.20 & -4943.28 \
0.25 & -1981.30 \
0.30 & -483.658 \
0.35 & 158.953
endarray
right)$$ So, the solution is between $0.30$ and $0.35$. Repeat using $Delta alpha=0.005$ to get
$$left(
beginarraycc
alpha & (2) \
0.315 & -221.502 \
0.320 & -148.821 \
0.325 & -82.8318 \
0.330 & -23.1689 \
0.335 & 30.5170
endarray
right)$$
So, the solution is between $0.330$ and $0.335$. Repeat using $Delta alpha=0.0005$ to get
$$left(
beginarraycc
alpha & (2) \
0.3305 & -17.537 \
0.3310 & -11.9644 \
0.3315 & -6.45098 \
0.3320 & -0.996267 \
0.3325 & 4.40004
endarray
right)$$ You will finish with $alpha=0.332092$ to which correspond $b=3.54869$.



Update



If the model is instead
$$a=b_1,n^alpha+b_2,n^1-alpha+b_3$$ consider that $alpha$ is fixed at a given value. For this value, define two parameters $x_i=n_i^ alpha$, $y_i=n_i^1-alpha$ and so, you face, for a given $alpha$ a linear regression
$$a=b_1,x+b_2,y+b_3$$ which is easy to solve using matrix calculation or simpler the normal equations
$$sum_i=1^p a_i=b_1 sum_i=1^p x_i+b_2 sum_i=1^p y_i+b_3,p$$
$$sum_i=1^p a_ix_i=b_1 sum_i=1^p x_i^2+b_2 sum_i=1^p x_iy_i+b_3sum_i=1^p x_i$$
$$sum_i=1^p a_iy_i=b_1 sum_i=1^p x_iy_i+b_2 sum_i=1^p y_i^2+b_3sum_i=1^p y_i$$ or just multilinear regression.



The resulting parameters are $b_1(alpha), b_2(alpha), b_3(alpha)$ and now consider
$$SSQ(alpha)=sum_i=1^p left(b_1(alpha),n_i^alpha+b_2(alpha),n_i^1-alpha+b_3(alpha) -a_iright)^2$$ which needs to be minimized with respect to $alpha$. As before, plot to locate more or less the minimum and zoom more and more until you reach the desired accuracy.



For example, considering the data set
$$left(
beginarrayccc
i & n_i & a_n_i \
1 & 10 & 254 \
2 & 20 & 357 \
3 & 30 & 442 \
4 & 40 & 516 \
5 & 50 & 584 \
6 & 60 & 646 \
7 & 70 & 705 \
8 & 80 & 761 \
9 & 90 & 814 \
10 & 100 & 865 \
11 & 110 & 914 \
12 & 120 & 961 \
13 & 130 & 1007 \
14 & 140 & 1052 \
15 & 150 & 1096
endarray
right)$$ we should have
$$left(
beginarraycc
alpha & SSQ(alpha) \
0.0 & 13251.5 \
0.1 & 126.579 \
0.2 & 33.3071 \
0.3 & 3.82528 \
0.4 & 1.38533 \
0.5 & 1745.69
endarray
right)$$ Continue zooming in the area of the minimum and finixh with
$$a=10.3547 ,n^0.363391+40.1755, n^0.636609+55.9928$$ which will give as final results
$$left(
beginarraycccc
i & n_i & a_n_i & textpredicted \
1 & 10 & 254 & 253.908 \
2 & 20 & 357 & 357.274 \
3 & 30 & 442 & 441.825 \
4 & 40 & 516 & 516.136 \
5 & 50 & 584 & 583.675 \
6 & 60 & 646 & 646.276 \
7 & 70 & 705 & 705.055 \
8 & 80 & 761 & 760.751 \
9 & 90 & 814 & 813.890 \
10 & 100 & 865 & 864.856 \
11 & 110 & 914 & 913.946 \
12 & 120 & 961 & 961.392 \
13 & 130 & 1007 & 1007.38 \
14 & 140 & 1052 & 1052.06 \
15 & 150 & 1096 & 1095.57
endarray
right)$$






share|cite|improve this answer























  • This answer gives me the basic idea of how to do the estimation here, especially the zoom operation. Thanks a lot.
    – Jimmy Kang
    Jul 23 at 19:56










Your Answer




StackExchange.ifUsing("editor", function ()
return StackExchange.using("mathjaxEditing", function ()
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
);
);
, "mathjax-editing");

StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "69"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);

else
createEditor();

);

function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
convertImagesToLinks: true,
noModals: false,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);



);








 

draft saved


draft discarded


















StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2860091%2fhow-to-estimate-alpha-given-a-discrete-sequence-an-n-alpha-n1-alp%23new-answer', 'question_page');

);

Post as a guest






























3 Answers
3






active

oldest

votes








3 Answers
3






active

oldest

votes









active

oldest

votes






active

oldest

votes








up vote
2
down vote













Rather than a two-parameter least-squares, you might note that your formula implies
$$a_i+1-a_i = n_i+1^alpha - n_i^alpha + n_i+1^1-alpha - n_i^1-alpha$$
Find the $alpha$ that minimizes the sum of squared errors here, then for $b$ take the average of $a_i - (n_i^alpha + n_i^1-alpha)$.






share|cite|improve this answer





















  • For sure ! Much simpler than my junk !
    – Claude Leibovici
    Jul 23 at 8:38










  • If I am right, $n_i=i$, there is no need to index.
    – Yves Daoust
    Jul 23 at 8:40










  • @YvesDaoust. Not necessary since you can use, say, $n=10,20,30,cdots$.
    – Claude Leibovici
    Jul 23 at 8:53










  • @ClaudeLeibovici: right, but then the formula must be with $a_n_i$.
    – Yves Daoust
    Jul 23 at 8:56










  • Many thanks. I have tried your answer. I thought the data can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$. Actually, I have $5$ sequences, each should be modeled with $alpha = 1/1.9,1/2.0, 1/2.1, 1/2.2, 1/2.5$ respectively. However, the strange thing is the partial derivative of the squared errors achieves $0$ at $alpha = 0.5$ for all sequences. I'm still trying to figure out what happens in the data.
    – Jimmy Kang
    Jul 23 at 20:20














up vote
2
down vote













Rather than a two-parameter least-squares, you might note that your formula implies
$$a_i+1-a_i = n_i+1^alpha - n_i^alpha + n_i+1^1-alpha - n_i^1-alpha$$
Find the $alpha$ that minimizes the sum of squared errors here, then for $b$ take the average of $a_i - (n_i^alpha + n_i^1-alpha)$.






share|cite|improve this answer





















  • For sure ! Much simpler than my junk !
    – Claude Leibovici
    Jul 23 at 8:38










  • If I am right, $n_i=i$, there is no need to index.
    – Yves Daoust
    Jul 23 at 8:40










  • @YvesDaoust. Not necessary since you can use, say, $n=10,20,30,cdots$.
    – Claude Leibovici
    Jul 23 at 8:53










  • @ClaudeLeibovici: right, but then the formula must be with $a_n_i$.
    – Yves Daoust
    Jul 23 at 8:56










  • Many thanks. I have tried your answer. I thought the data can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$. Actually, I have $5$ sequences, each should be modeled with $alpha = 1/1.9,1/2.0, 1/2.1, 1/2.2, 1/2.5$ respectively. However, the strange thing is the partial derivative of the squared errors achieves $0$ at $alpha = 0.5$ for all sequences. I'm still trying to figure out what happens in the data.
    – Jimmy Kang
    Jul 23 at 20:20












up vote
2
down vote










up vote
2
down vote









Rather than a two-parameter least-squares, you might note that your formula implies
$$a_i+1-a_i = n_i+1^alpha - n_i^alpha + n_i+1^1-alpha - n_i^1-alpha$$
Find the $alpha$ that minimizes the sum of squared errors here, then for $b$ take the average of $a_i - (n_i^alpha + n_i^1-alpha)$.






share|cite|improve this answer













Rather than a two-parameter least-squares, you might note that your formula implies
$$a_i+1-a_i = n_i+1^alpha - n_i^alpha + n_i+1^1-alpha - n_i^1-alpha$$
Find the $alpha$ that minimizes the sum of squared errors here, then for $b$ take the average of $a_i - (n_i^alpha + n_i^1-alpha)$.







share|cite|improve this answer













share|cite|improve this answer



share|cite|improve this answer











answered Jul 23 at 8:30









Robert Israel

304k22201441




304k22201441











  • For sure ! Much simpler than my junk !
    – Claude Leibovici
    Jul 23 at 8:38










  • If I am right, $n_i=i$, there is no need to index.
    – Yves Daoust
    Jul 23 at 8:40










  • @YvesDaoust. Not necessary since you can use, say, $n=10,20,30,cdots$.
    – Claude Leibovici
    Jul 23 at 8:53










  • @ClaudeLeibovici: right, but then the formula must be with $a_n_i$.
    – Yves Daoust
    Jul 23 at 8:56










  • Many thanks. I have tried your answer. I thought the data can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$. Actually, I have $5$ sequences, each should be modeled with $alpha = 1/1.9,1/2.0, 1/2.1, 1/2.2, 1/2.5$ respectively. However, the strange thing is the partial derivative of the squared errors achieves $0$ at $alpha = 0.5$ for all sequences. I'm still trying to figure out what happens in the data.
    – Jimmy Kang
    Jul 23 at 20:20
















  • For sure ! Much simpler than my junk !
    – Claude Leibovici
    Jul 23 at 8:38










  • If I am right, $n_i=i$, there is no need to index.
    – Yves Daoust
    Jul 23 at 8:40










  • @YvesDaoust. Not necessary since you can use, say, $n=10,20,30,cdots$.
    – Claude Leibovici
    Jul 23 at 8:53










  • @ClaudeLeibovici: right, but then the formula must be with $a_n_i$.
    – Yves Daoust
    Jul 23 at 8:56










  • Many thanks. I have tried your answer. I thought the data can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$. Actually, I have $5$ sequences, each should be modeled with $alpha = 1/1.9,1/2.0, 1/2.1, 1/2.2, 1/2.5$ respectively. However, the strange thing is the partial derivative of the squared errors achieves $0$ at $alpha = 0.5$ for all sequences. I'm still trying to figure out what happens in the data.
    – Jimmy Kang
    Jul 23 at 20:20















For sure ! Much simpler than my junk !
– Claude Leibovici
Jul 23 at 8:38




For sure ! Much simpler than my junk !
– Claude Leibovici
Jul 23 at 8:38












If I am right, $n_i=i$, there is no need to index.
– Yves Daoust
Jul 23 at 8:40




If I am right, $n_i=i$, there is no need to index.
– Yves Daoust
Jul 23 at 8:40












@YvesDaoust. Not necessary since you can use, say, $n=10,20,30,cdots$.
– Claude Leibovici
Jul 23 at 8:53




@YvesDaoust. Not necessary since you can use, say, $n=10,20,30,cdots$.
– Claude Leibovici
Jul 23 at 8:53












@ClaudeLeibovici: right, but then the formula must be with $a_n_i$.
– Yves Daoust
Jul 23 at 8:56




@ClaudeLeibovici: right, but then the formula must be with $a_n_i$.
– Yves Daoust
Jul 23 at 8:56












Many thanks. I have tried your answer. I thought the data can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$. Actually, I have $5$ sequences, each should be modeled with $alpha = 1/1.9,1/2.0, 1/2.1, 1/2.2, 1/2.5$ respectively. However, the strange thing is the partial derivative of the squared errors achieves $0$ at $alpha = 0.5$ for all sequences. I'm still trying to figure out what happens in the data.
– Jimmy Kang
Jul 23 at 20:20




Many thanks. I have tried your answer. I thought the data can be approximately modeled as $a(n) = n^alpha+ n^1-alpha+b$. Actually, I have $5$ sequences, each should be modeled with $alpha = 1/1.9,1/2.0, 1/2.1, 1/2.2, 1/2.5$ respectively. However, the strange thing is the partial derivative of the squared errors achieves $0$ at $alpha = 0.5$ for all sequences. I'm still trying to figure out what happens in the data.
– Jimmy Kang
Jul 23 at 20:20










up vote
2
down vote













If this helps:



If $a_0$ or $a_1-2$ are available and give reasonable estimates of $b$, for large $n$ we have



$$fraca_n-hat bsqrt nsim n^alpha-1/2+n^1/2-alpha=2coshleft(left(frac12-alpharight)log nright)$$



and



$$alphasimfrac12-dfractextarcoshdfraca_n-hat b2sqrt nlog n.$$



This can be used f.i. as a starting value of an iterative method.






share|cite|improve this answer























  • This seems to be a quite good idea, for sure !
    – Claude Leibovici
    Jul 23 at 9:17










  • Sorry for the late reply. Data is noisy. The initial terms do not match the model. My concern is just to estimate the asymptotic behavior. The data should be $O(n^max(alpha,1-alpha))$. I want to get the order of the sequence, i.e. $max(alpha,1-alpha)$.
    – Jimmy Kang
    Jul 23 at 20:01











  • @JimmyKang: why do you add a constant term ?
    – Yves Daoust
    Jul 23 at 20:41










  • I revised the problem just now. Sorry for the confusion caused.
    – Jimmy Kang
    Jul 23 at 21:21














up vote
2
down vote













If this helps:



If $a_0$ or $a_1-2$ are available and give reasonable estimates of $b$, for large $n$ we have



$$fraca_n-hat bsqrt nsim n^alpha-1/2+n^1/2-alpha=2coshleft(left(frac12-alpharight)log nright)$$



and



$$alphasimfrac12-dfractextarcoshdfraca_n-hat b2sqrt nlog n.$$



This can be used f.i. as a starting value of an iterative method.






share|cite|improve this answer























  • This seems to be a quite good idea, for sure !
    – Claude Leibovici
    Jul 23 at 9:17










  • Sorry for the late reply. Data is noisy. The initial terms do not match the model. My concern is just to estimate the asymptotic behavior. The data should be $O(n^max(alpha,1-alpha))$. I want to get the order of the sequence, i.e. $max(alpha,1-alpha)$.
    – Jimmy Kang
    Jul 23 at 20:01











  • @JimmyKang: why do you add a constant term ?
    – Yves Daoust
    Jul 23 at 20:41










  • I revised the problem just now. Sorry for the confusion caused.
    – Jimmy Kang
    Jul 23 at 21:21












up vote
2
down vote










up vote
2
down vote









If this helps:



If $a_0$ or $a_1-2$ are available and give reasonable estimates of $b$, for large $n$ we have



$$fraca_n-hat bsqrt nsim n^alpha-1/2+n^1/2-alpha=2coshleft(left(frac12-alpharight)log nright)$$



and



$$alphasimfrac12-dfractextarcoshdfraca_n-hat b2sqrt nlog n.$$



This can be used f.i. as a starting value of an iterative method.






share|cite|improve this answer















If this helps:



If $a_0$ or $a_1-2$ are available and give reasonable estimates of $b$, for large $n$ we have



$$fraca_n-hat bsqrt nsim n^alpha-1/2+n^1/2-alpha=2coshleft(left(frac12-alpharight)log nright)$$



and



$$alphasimfrac12-dfractextarcoshdfraca_n-hat b2sqrt nlog n.$$



This can be used f.i. as a starting value of an iterative method.







share|cite|improve this answer















share|cite|improve this answer



share|cite|improve this answer








edited Jul 23 at 9:08


























answered Jul 23 at 9:00









Yves Daoust

111k665203




111k665203











  • This seems to be a quite good idea, for sure !
    – Claude Leibovici
    Jul 23 at 9:17










  • Sorry for the late reply. Data is noisy. The initial terms do not match the model. My concern is just to estimate the asymptotic behavior. The data should be $O(n^max(alpha,1-alpha))$. I want to get the order of the sequence, i.e. $max(alpha,1-alpha)$.
    – Jimmy Kang
    Jul 23 at 20:01











  • @JimmyKang: why do you add a constant term ?
    – Yves Daoust
    Jul 23 at 20:41










  • I revised the problem just now. Sorry for the confusion caused.
    – Jimmy Kang
    Jul 23 at 21:21
















  • This seems to be a quite good idea, for sure !
    – Claude Leibovici
    Jul 23 at 9:17










  • Sorry for the late reply. Data is noisy. The initial terms do not match the model. My concern is just to estimate the asymptotic behavior. The data should be $O(n^max(alpha,1-alpha))$. I want to get the order of the sequence, i.e. $max(alpha,1-alpha)$.
    – Jimmy Kang
    Jul 23 at 20:01











  • @JimmyKang: why do you add a constant term ?
    – Yves Daoust
    Jul 23 at 20:41










  • I revised the problem just now. Sorry for the confusion caused.
    – Jimmy Kang
    Jul 23 at 21:21















This seems to be a quite good idea, for sure !
– Claude Leibovici
Jul 23 at 9:17




This seems to be a quite good idea, for sure !
– Claude Leibovici
Jul 23 at 9:17












Sorry for the late reply. Data is noisy. The initial terms do not match the model. My concern is just to estimate the asymptotic behavior. The data should be $O(n^max(alpha,1-alpha))$. I want to get the order of the sequence, i.e. $max(alpha,1-alpha)$.
– Jimmy Kang
Jul 23 at 20:01





Sorry for the late reply. Data is noisy. The initial terms do not match the model. My concern is just to estimate the asymptotic behavior. The data should be $O(n^max(alpha,1-alpha))$. I want to get the order of the sequence, i.e. $max(alpha,1-alpha)$.
– Jimmy Kang
Jul 23 at 20:01













@JimmyKang: why do you add a constant term ?
– Yves Daoust
Jul 23 at 20:41




@JimmyKang: why do you add a constant term ?
– Yves Daoust
Jul 23 at 20:41












I revised the problem just now. Sorry for the confusion caused.
– Jimmy Kang
Jul 23 at 21:21




I revised the problem just now. Sorry for the confusion caused.
– Jimmy Kang
Jul 23 at 21:21










up vote
1
down vote













If we look at the problem as a regression, you have $p$ data points $(n_i,a_i)$ and you want to adjust the model
$$a=n^alpha+n^1-alpha+b$$ In the least square sense, you need to minimize
$$SSQ=sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)^2$$ Computing the partial derivatives,we have
$$fracpartial SSQpartial b=2sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)=0tag 1$$
$$fracpartial SSQpartial a=2sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)left(n_i^alpha-n_i^1-alpharight)log(n_i)=0 tag 2$$ which is hard to solve.



However, from $(1)$, you can get $b$ as a function of $alpha$
$$b(alpha)=-frac 1p sum_i=1^p left(n_i^alpha+n_i^1-alpha -a_iright)$$ and then $(2)$ is "just" an equation in $alpha$.



The simplest would be to plot equation $(2)$ for $0 leq alpha leq 0.5$ and see where it does cancels. Zoom more and more to have more accurate results; when your accuracy has been reached, recompute $b$.



All of that can be done using Excel.



Edit



For illustration purposes, let us use the following data set
$$left(
beginarrayccc
i & n_i & a_n_i \
1 & 10 & 10 \
2 & 20 & 14 \
3 & 30 & 16 \
4 & 40 & 19 \
5 & 50 & 21 \
6 & 60 & 23 \
7 & 70 & 25 \
8 & 80 & 26 \
9 & 90 & 28 \
10 & 100 & 30
endarray
right)$$ A first run, using $Delta alpha=0.05$ would give
$$left(
beginarraycc
alpha & (2) \
0.05 & -36241.2 \
0.10 & -19913.1 \
0.15 & -10372.8 \
0.20 & -4943.28 \
0.25 & -1981.30 \
0.30 & -483.658 \
0.35 & 158.953
endarray
right)$$ So, the solution is between $0.30$ and $0.35$. Repeat using $Delta alpha=0.005$ to get
$$left(
beginarraycc
alpha & (2) \
0.315 & -221.502 \
0.320 & -148.821 \
0.325 & -82.8318 \
0.330 & -23.1689 \
0.335 & 30.5170
endarray
right)$$
So, the solution is between $0.330$ and $0.335$. Repeat using $Delta alpha=0.0005$ to get
$$left(
beginarraycc
alpha & (2) \
0.3305 & -17.537 \
0.3310 & -11.9644 \
0.3315 & -6.45098 \
0.3320 & -0.996267 \
0.3325 & 4.40004
endarray
right)$$ You will finish with $alpha=0.332092$ to which correspond $b=3.54869$.



Update



If the model is instead
$$a=b_1,n^alpha+b_2,n^1-alpha+b_3$$ consider that $alpha$ is fixed at a given value. For this value, define two parameters $x_i=n_i^ alpha$, $y_i=n_i^1-alpha$ and so, you face, for a given $alpha$ a linear regression
$$a=b_1,x+b_2,y+b_3$$ which is easy to solve using matrix calculation or simpler the normal equations
$$sum_i=1^p a_i=b_1 sum_i=1^p x_i+b_2 sum_i=1^p y_i+b_3,p$$
$$sum_i=1^p a_ix_i=b_1 sum_i=1^p x_i^2+b_2 sum_i=1^p x_iy_i+b_3sum_i=1^p x_i$$
$$sum_i=1^p a_iy_i=b_1 sum_i=1^p x_iy_i+b_2 sum_i=1^p y_i^2+b_3sum_i=1^p y_i$$ or just multilinear regression.



The resulting parameters are $b_1(alpha), b_2(alpha), b_3(alpha)$ and now consider
$$SSQ(alpha)=sum_i=1^p left(b_1(alpha),n_i^alpha+b_2(alpha),n_i^1-alpha+b_3(alpha) -a_iright)^2$$ which needs to be minimized with respect to $alpha$. As before, plot to locate more or less the minimum and zoom more and more until you reach the desired accuracy.



For example, considering the data set
$$left(
beginarrayccc
i & n_i & a_n_i \
1 & 10 & 254 \
2 & 20 & 357 \
3 & 30 & 442 \
4 & 40 & 516 \
5 & 50 & 584 \
6 & 60 & 646 \
7 & 70 & 705 \
8 & 80 & 761 \
9 & 90 & 814 \
10 & 100 & 865 \
11 & 110 & 914 \
12 & 120 & 961 \
13 & 130 & 1007 \
14 & 140 & 1052 \
15 & 150 & 1096
endarray
right)$$ we should have
$$left(
beginarraycc
alpha & SSQ(alpha) \
0.0 & 13251.5 \
0.1 & 126.579 \
0.2 & 33.3071 \
0.3 & 3.82528 \
0.4 & 1.38533 \
0.5 & 1745.69
endarray
right)$$ Continue zooming in the area of the minimum and finixh with
$$a=10.3547 ,n^0.363391+40.1755, n^0.636609+55.9928$$ which will give as final results
$$left(
beginarraycccc
i & n_i & a_n_i & textpredicted \
1 & 10 & 254 & 253.908 \
2 & 20 & 357 & 357.274 \
3 & 30 & 442 & 441.825 \
4 & 40 & 516 & 516.136 \
5 & 50 & 584 & 583.675 \
6 & 60 & 646 & 646.276 \
7 & 70 & 705 & 705.055 \
8 & 80 & 761 & 760.751 \
9 & 90 & 814 & 813.890 \
10 & 100 & 865 & 864.856 \
11 & 110 & 914 & 913.946 \
12 & 120 & 961 & 961.392 \
13 & 130 & 1007 & 1007.38 \
14 & 140 & 1052 & 1052.06 \
15 & 150 & 1096 & 1095.57
endarray
right)$$






share|cite|improve this answer























  • This answer gives me the basic idea of how to do the estimation here, especially the zoom operation. Thanks a lot.
    – Jimmy Kang
    Jul 23 at 19:56














up vote
1
down vote













If we look at the problem as a regression, you have $p$ data points $(n_i,a_i)$ and you want to adjust the model
$$a=n^alpha+n^1-alpha+b$$ In the least square sense, you need to minimize
$$SSQ=sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)^2$$ Computing the partial derivatives,we have
$$fracpartial SSQpartial b=2sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)=0tag 1$$
$$fracpartial SSQpartial a=2sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)left(n_i^alpha-n_i^1-alpharight)log(n_i)=0 tag 2$$ which is hard to solve.



However, from $(1)$, you can get $b$ as a function of $alpha$
$$b(alpha)=-frac 1p sum_i=1^p left(n_i^alpha+n_i^1-alpha -a_iright)$$ and then $(2)$ is "just" an equation in $alpha$.



The simplest would be to plot equation $(2)$ for $0 leq alpha leq 0.5$ and see where it does cancels. Zoom more and more to have more accurate results; when your accuracy has been reached, recompute $b$.



All of that can be done using Excel.



Edit



For illustration purposes, let us use the following data set
$$left(
beginarrayccc
i & n_i & a_n_i \
1 & 10 & 10 \
2 & 20 & 14 \
3 & 30 & 16 \
4 & 40 & 19 \
5 & 50 & 21 \
6 & 60 & 23 \
7 & 70 & 25 \
8 & 80 & 26 \
9 & 90 & 28 \
10 & 100 & 30
endarray
right)$$ A first run, using $Delta alpha=0.05$ would give
$$left(
beginarraycc
alpha & (2) \
0.05 & -36241.2 \
0.10 & -19913.1 \
0.15 & -10372.8 \
0.20 & -4943.28 \
0.25 & -1981.30 \
0.30 & -483.658 \
0.35 & 158.953
endarray
right)$$ So, the solution is between $0.30$ and $0.35$. Repeat using $Delta alpha=0.005$ to get
$$left(
beginarraycc
alpha & (2) \
0.315 & -221.502 \
0.320 & -148.821 \
0.325 & -82.8318 \
0.330 & -23.1689 \
0.335 & 30.5170
endarray
right)$$
So, the solution is between $0.330$ and $0.335$. Repeat using $Delta alpha=0.0005$ to get
$$left(
beginarraycc
alpha & (2) \
0.3305 & -17.537 \
0.3310 & -11.9644 \
0.3315 & -6.45098 \
0.3320 & -0.996267 \
0.3325 & 4.40004
endarray
right)$$ You will finish with $alpha=0.332092$ to which correspond $b=3.54869$.



Update



If the model is instead
$$a=b_1,n^alpha+b_2,n^1-alpha+b_3$$ consider that $alpha$ is fixed at a given value. For this value, define two parameters $x_i=n_i^ alpha$, $y_i=n_i^1-alpha$ and so, you face, for a given $alpha$ a linear regression
$$a=b_1,x+b_2,y+b_3$$ which is easy to solve using matrix calculation or simpler the normal equations
$$sum_i=1^p a_i=b_1 sum_i=1^p x_i+b_2 sum_i=1^p y_i+b_3,p$$
$$sum_i=1^p a_ix_i=b_1 sum_i=1^p x_i^2+b_2 sum_i=1^p x_iy_i+b_3sum_i=1^p x_i$$
$$sum_i=1^p a_iy_i=b_1 sum_i=1^p x_iy_i+b_2 sum_i=1^p y_i^2+b_3sum_i=1^p y_i$$ or just multilinear regression.



The resulting parameters are $b_1(alpha), b_2(alpha), b_3(alpha)$ and now consider
$$SSQ(alpha)=sum_i=1^p left(b_1(alpha),n_i^alpha+b_2(alpha),n_i^1-alpha+b_3(alpha) -a_iright)^2$$ which needs to be minimized with respect to $alpha$. As before, plot to locate more or less the minimum and zoom more and more until you reach the desired accuracy.



For example, considering the data set
$$left(
beginarrayccc
i & n_i & a_n_i \
1 & 10 & 254 \
2 & 20 & 357 \
3 & 30 & 442 \
4 & 40 & 516 \
5 & 50 & 584 \
6 & 60 & 646 \
7 & 70 & 705 \
8 & 80 & 761 \
9 & 90 & 814 \
10 & 100 & 865 \
11 & 110 & 914 \
12 & 120 & 961 \
13 & 130 & 1007 \
14 & 140 & 1052 \
15 & 150 & 1096
endarray
right)$$ we should have
$$left(
beginarraycc
alpha & SSQ(alpha) \
0.0 & 13251.5 \
0.1 & 126.579 \
0.2 & 33.3071 \
0.3 & 3.82528 \
0.4 & 1.38533 \
0.5 & 1745.69
endarray
right)$$ Continue zooming in the area of the minimum and finixh with
$$a=10.3547 ,n^0.363391+40.1755, n^0.636609+55.9928$$ which will give as final results
$$left(
beginarraycccc
i & n_i & a_n_i & textpredicted \
1 & 10 & 254 & 253.908 \
2 & 20 & 357 & 357.274 \
3 & 30 & 442 & 441.825 \
4 & 40 & 516 & 516.136 \
5 & 50 & 584 & 583.675 \
6 & 60 & 646 & 646.276 \
7 & 70 & 705 & 705.055 \
8 & 80 & 761 & 760.751 \
9 & 90 & 814 & 813.890 \
10 & 100 & 865 & 864.856 \
11 & 110 & 914 & 913.946 \
12 & 120 & 961 & 961.392 \
13 & 130 & 1007 & 1007.38 \
14 & 140 & 1052 & 1052.06 \
15 & 150 & 1096 & 1095.57
endarray
right)$$






share|cite|improve this answer























  • This answer gives me the basic idea of how to do the estimation here, especially the zoom operation. Thanks a lot.
    – Jimmy Kang
    Jul 23 at 19:56












up vote
1
down vote










up vote
1
down vote









If we look at the problem as a regression, you have $p$ data points $(n_i,a_i)$ and you want to adjust the model
$$a=n^alpha+n^1-alpha+b$$ In the least square sense, you need to minimize
$$SSQ=sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)^2$$ Computing the partial derivatives,we have
$$fracpartial SSQpartial b=2sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)=0tag 1$$
$$fracpartial SSQpartial a=2sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)left(n_i^alpha-n_i^1-alpharight)log(n_i)=0 tag 2$$ which is hard to solve.



However, from $(1)$, you can get $b$ as a function of $alpha$
$$b(alpha)=-frac 1p sum_i=1^p left(n_i^alpha+n_i^1-alpha -a_iright)$$ and then $(2)$ is "just" an equation in $alpha$.



The simplest would be to plot equation $(2)$ for $0 leq alpha leq 0.5$ and see where it does cancels. Zoom more and more to have more accurate results; when your accuracy has been reached, recompute $b$.



All of that can be done using Excel.



Edit



For illustration purposes, let us use the following data set
$$left(
beginarrayccc
i & n_i & a_n_i \
1 & 10 & 10 \
2 & 20 & 14 \
3 & 30 & 16 \
4 & 40 & 19 \
5 & 50 & 21 \
6 & 60 & 23 \
7 & 70 & 25 \
8 & 80 & 26 \
9 & 90 & 28 \
10 & 100 & 30
endarray
right)$$ A first run, using $Delta alpha=0.05$ would give
$$left(
beginarraycc
alpha & (2) \
0.05 & -36241.2 \
0.10 & -19913.1 \
0.15 & -10372.8 \
0.20 & -4943.28 \
0.25 & -1981.30 \
0.30 & -483.658 \
0.35 & 158.953
endarray
right)$$ So, the solution is between $0.30$ and $0.35$. Repeat using $Delta alpha=0.005$ to get
$$left(
beginarraycc
alpha & (2) \
0.315 & -221.502 \
0.320 & -148.821 \
0.325 & -82.8318 \
0.330 & -23.1689 \
0.335 & 30.5170
endarray
right)$$
So, the solution is between $0.330$ and $0.335$. Repeat using $Delta alpha=0.0005$ to get
$$left(
beginarraycc
alpha & (2) \
0.3305 & -17.537 \
0.3310 & -11.9644 \
0.3315 & -6.45098 \
0.3320 & -0.996267 \
0.3325 & 4.40004
endarray
right)$$ You will finish with $alpha=0.332092$ to which correspond $b=3.54869$.



Update



If the model is instead
$$a=b_1,n^alpha+b_2,n^1-alpha+b_3$$ consider that $alpha$ is fixed at a given value. For this value, define two parameters $x_i=n_i^ alpha$, $y_i=n_i^1-alpha$ and so, you face, for a given $alpha$ a linear regression
$$a=b_1,x+b_2,y+b_3$$ which is easy to solve using matrix calculation or simpler the normal equations
$$sum_i=1^p a_i=b_1 sum_i=1^p x_i+b_2 sum_i=1^p y_i+b_3,p$$
$$sum_i=1^p a_ix_i=b_1 sum_i=1^p x_i^2+b_2 sum_i=1^p x_iy_i+b_3sum_i=1^p x_i$$
$$sum_i=1^p a_iy_i=b_1 sum_i=1^p x_iy_i+b_2 sum_i=1^p y_i^2+b_3sum_i=1^p y_i$$ or just multilinear regression.



The resulting parameters are $b_1(alpha), b_2(alpha), b_3(alpha)$ and now consider
$$SSQ(alpha)=sum_i=1^p left(b_1(alpha),n_i^alpha+b_2(alpha),n_i^1-alpha+b_3(alpha) -a_iright)^2$$ which needs to be minimized with respect to $alpha$. As before, plot to locate more or less the minimum and zoom more and more until you reach the desired accuracy.



For example, considering the data set
$$left(
beginarrayccc
i & n_i & a_n_i \
1 & 10 & 254 \
2 & 20 & 357 \
3 & 30 & 442 \
4 & 40 & 516 \
5 & 50 & 584 \
6 & 60 & 646 \
7 & 70 & 705 \
8 & 80 & 761 \
9 & 90 & 814 \
10 & 100 & 865 \
11 & 110 & 914 \
12 & 120 & 961 \
13 & 130 & 1007 \
14 & 140 & 1052 \
15 & 150 & 1096
endarray
right)$$ we should have
$$left(
beginarraycc
alpha & SSQ(alpha) \
0.0 & 13251.5 \
0.1 & 126.579 \
0.2 & 33.3071 \
0.3 & 3.82528 \
0.4 & 1.38533 \
0.5 & 1745.69
endarray
right)$$ Continue zooming in the area of the minimum and finixh with
$$a=10.3547 ,n^0.363391+40.1755, n^0.636609+55.9928$$ which will give as final results
$$left(
beginarraycccc
i & n_i & a_n_i & textpredicted \
1 & 10 & 254 & 253.908 \
2 & 20 & 357 & 357.274 \
3 & 30 & 442 & 441.825 \
4 & 40 & 516 & 516.136 \
5 & 50 & 584 & 583.675 \
6 & 60 & 646 & 646.276 \
7 & 70 & 705 & 705.055 \
8 & 80 & 761 & 760.751 \
9 & 90 & 814 & 813.890 \
10 & 100 & 865 & 864.856 \
11 & 110 & 914 & 913.946 \
12 & 120 & 961 & 961.392 \
13 & 130 & 1007 & 1007.38 \
14 & 140 & 1052 & 1052.06 \
15 & 150 & 1096 & 1095.57
endarray
right)$$






share|cite|improve this answer















If we look at the problem as a regression, you have $p$ data points $(n_i,a_i)$ and you want to adjust the model
$$a=n^alpha+n^1-alpha+b$$ In the least square sense, you need to minimize
$$SSQ=sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)^2$$ Computing the partial derivatives,we have
$$fracpartial SSQpartial b=2sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)=0tag 1$$
$$fracpartial SSQpartial a=2sum_i=1^p left(n_i^alpha+n_i^1-alpha+b -a_iright)left(n_i^alpha-n_i^1-alpharight)log(n_i)=0 tag 2$$ which is hard to solve.



However, from $(1)$, you can get $b$ as a function of $alpha$
$$b(alpha)=-frac 1p sum_i=1^p left(n_i^alpha+n_i^1-alpha -a_iright)$$ and then $(2)$ is "just" an equation in $alpha$.



The simplest would be to plot equation $(2)$ for $0 leq alpha leq 0.5$ and see where it does cancels. Zoom more and more to have more accurate results; when your accuracy has been reached, recompute $b$.



All of that can be done using Excel.



Edit



For illustration purposes, let us use the following data set
$$left(
beginarrayccc
i & n_i & a_n_i \
1 & 10 & 10 \
2 & 20 & 14 \
3 & 30 & 16 \
4 & 40 & 19 \
5 & 50 & 21 \
6 & 60 & 23 \
7 & 70 & 25 \
8 & 80 & 26 \
9 & 90 & 28 \
10 & 100 & 30
endarray
right)$$ A first run, using $Delta alpha=0.05$ would give
$$left(
beginarraycc
alpha & (2) \
0.05 & -36241.2 \
0.10 & -19913.1 \
0.15 & -10372.8 \
0.20 & -4943.28 \
0.25 & -1981.30 \
0.30 & -483.658 \
0.35 & 158.953
endarray
right)$$ So, the solution is between $0.30$ and $0.35$. Repeat using $Delta alpha=0.005$ to get
$$left(
beginarraycc
alpha & (2) \
0.315 & -221.502 \
0.320 & -148.821 \
0.325 & -82.8318 \
0.330 & -23.1689 \
0.335 & 30.5170
endarray
right)$$
So, the solution is between $0.330$ and $0.335$. Repeat using $Delta alpha=0.0005$ to get
$$left(
beginarraycc
alpha & (2) \
0.3305 & -17.537 \
0.3310 & -11.9644 \
0.3315 & -6.45098 \
0.3320 & -0.996267 \
0.3325 & 4.40004
endarray
right)$$ You will finish with $alpha=0.332092$ to which correspond $b=3.54869$.



Update



If the model is instead
$$a=b_1,n^alpha+b_2,n^1-alpha+b_3$$ consider that $alpha$ is fixed at a given value. For this value, define two parameters $x_i=n_i^ alpha$, $y_i=n_i^1-alpha$ and so, you face, for a given $alpha$ a linear regression
$$a=b_1,x+b_2,y+b_3$$ which is easy to solve using matrix calculation or simpler the normal equations
$$sum_i=1^p a_i=b_1 sum_i=1^p x_i+b_2 sum_i=1^p y_i+b_3,p$$
$$sum_i=1^p a_ix_i=b_1 sum_i=1^p x_i^2+b_2 sum_i=1^p x_iy_i+b_3sum_i=1^p x_i$$
$$sum_i=1^p a_iy_i=b_1 sum_i=1^p x_iy_i+b_2 sum_i=1^p y_i^2+b_3sum_i=1^p y_i$$ or just multilinear regression.



The resulting parameters are $b_1(alpha), b_2(alpha), b_3(alpha)$ and now consider
$$SSQ(alpha)=sum_i=1^p left(b_1(alpha),n_i^alpha+b_2(alpha),n_i^1-alpha+b_3(alpha) -a_iright)^2$$ which needs to be minimized with respect to $alpha$. As before, plot to locate more or less the minimum and zoom more and more until you reach the desired accuracy.



For example, considering the data set
$$left(
beginarrayccc
i & n_i & a_n_i \
1 & 10 & 254 \
2 & 20 & 357 \
3 & 30 & 442 \
4 & 40 & 516 \
5 & 50 & 584 \
6 & 60 & 646 \
7 & 70 & 705 \
8 & 80 & 761 \
9 & 90 & 814 \
10 & 100 & 865 \
11 & 110 & 914 \
12 & 120 & 961 \
13 & 130 & 1007 \
14 & 140 & 1052 \
15 & 150 & 1096
endarray
right)$$ we should have
$$left(
beginarraycc
alpha & SSQ(alpha) \
0.0 & 13251.5 \
0.1 & 126.579 \
0.2 & 33.3071 \
0.3 & 3.82528 \
0.4 & 1.38533 \
0.5 & 1745.69
endarray
right)$$ Continue zooming in the area of the minimum and finixh with
$$a=10.3547 ,n^0.363391+40.1755, n^0.636609+55.9928$$ which will give as final results
$$left(
beginarraycccc
i & n_i & a_n_i & textpredicted \
1 & 10 & 254 & 253.908 \
2 & 20 & 357 & 357.274 \
3 & 30 & 442 & 441.825 \
4 & 40 & 516 & 516.136 \
5 & 50 & 584 & 583.675 \
6 & 60 & 646 & 646.276 \
7 & 70 & 705 & 705.055 \
8 & 80 & 761 & 760.751 \
9 & 90 & 814 & 813.890 \
10 & 100 & 865 & 864.856 \
11 & 110 & 914 & 913.946 \
12 & 120 & 961 & 961.392 \
13 & 130 & 1007 & 1007.38 \
14 & 140 & 1052 & 1052.06 \
15 & 150 & 1096 & 1095.57
endarray
right)$$







share|cite|improve this answer















share|cite|improve this answer



share|cite|improve this answer








edited Jul 24 at 3:52


























answered Jul 23 at 7:53









Claude Leibovici

111k1055126




111k1055126











  • This answer gives me the basic idea of how to do the estimation here, especially the zoom operation. Thanks a lot.
    – Jimmy Kang
    Jul 23 at 19:56
















  • This answer gives me the basic idea of how to do the estimation here, especially the zoom operation. Thanks a lot.
    – Jimmy Kang
    Jul 23 at 19:56















This answer gives me the basic idea of how to do the estimation here, especially the zoom operation. Thanks a lot.
– Jimmy Kang
Jul 23 at 19:56




This answer gives me the basic idea of how to do the estimation here, especially the zoom operation. Thanks a lot.
– Jimmy Kang
Jul 23 at 19:56












 

draft saved


draft discarded


























 


draft saved


draft discarded














StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2860091%2fhow-to-estimate-alpha-given-a-discrete-sequence-an-n-alpha-n1-alp%23new-answer', 'question_page');

);

Post as a guest













































































Comments

Popular posts from this blog

What is the equation of a 3D cone with generalised tilt?

Color the edges and diagonals of a regular polygon

Relationship between determinant of matrix and determinant of adjoint?