ImplicitDemands Posted May 10 Posted May 10 (edited) As far I can discern when inventing combinatorics I also found the product rule for differentiation and derivatives. Not "Isaac Newton" or some other name that I can't confirm ever existed. When pertaining to the scientific method, empirical evidence is based on observation which I don't rely on as heavily as mathematics for separating fact from fiction. I know I don't get anything out of sharing an arithmetic that can literally be used to earn a person money for cracking a password combination or solving any type of Rubiks cube. I should be retired for this alone. I only share it because I know some things that are potentially worth a lot more. Just realize whenever I post one of these there's an implied expectation somewhere in there for doing so. Edited May 10 by ImplicitDemands
exchemist Posted May 10 Posted May 10 (edited) 2 hours ago, ImplicitDemands said: As far I can discern when inventing combinatorics I also found the product rule for differentiation and derivatives. Not "Isaac Newton" or some other name that I can't confirm ever existed. When pertaining to the scientific method, empirical evidence is based on observation which I don't rely on as heavily as mathematics for separating fact from fiction. I know I don't get anything out of sharing an arithmetic that can literally be used to earn a person money for cracking a password combination or solving any type of Rubiks cube. I should be retired for this alone. I only share it because I know some things that are potentially worth a lot more. Just realize whenever I post one of these there's an implied expectation somewhere in there for doing so. Leibnitz and Newton are thought to have independently developed calculus at around the same time, though I think it was Leibnitz who published first. As so often with science and mathematics, the idea was germinating at the time and trying to determine who got there "first" is rather debatable and of limited value. Both men most certainly existed, though. I can't make sense of the rest of your post. Edited May 10 by exchemist
ImplicitDemands Posted May 10 Author Posted May 10 (edited) 8 hours ago, exchemist said: Leibnitz and Newton are thought to have independently developed calculus at around the same time, though I think it was Leibnitz who published first. As so often with science and mathematics, the idea was germinating at the time and trying to determine who got there "first" is rather debatable and of limited value. Both men most certainly existed, though. I can't make sense of the rest of your post. Well that doesn't tell me how they came up with the product rule. As for my post, I was showing how to break down all the 10 digit passwords below the obvious 10^10 possible combinations so one could work with them. I even made it more difficult of a task by having the number 1 repeated, that means there would have been two ones in any arrangement of 10 different digits. I had to take the zero out and find the 1-9 combinations. Then, I would have eliminated all the combinations with 9 in them, and I would have put 0 back in to replace the nine. Therefore, when computing for zero you'd have one set of 9^9 combinations where 9 is included and a second set of 9^9 combinations where 0 is included if 1 weren't repeating twice. So the 123456789 {0} is the equivalent of a square exponential, if you did the product rule it would be like you base^2, and then doing the process shown and swapping the 9 and the 0 would be like performing the product rule on base^2, that is (2 x base)^(2-1) = 2 x base. In essence, the difference between the 10^10 combinations of all the digits on your keyboard and the combinations you get when you sum up 123456789 {0} & 123456780 {9} is the same difference between base^2 and 2 x base (it's derivative). A Rubik's cube would be treated the same, with 54 digits instead of 10. Edited May 10 by ImplicitDemands
KJW Posted May 10 Posted May 10 (edited) 1 hour ago, ImplicitDemands said: Well that doesn't tell me how they came up with the product rule. I don't know precisely how Newton or Leibniz obtained the product rule of differential calculus, but it seems rather easy to obtain to me: [math]\text{By definition:}[/math] [math]\dfrac{df(x)}{dx} \buildrel \rm def \over = \displaystyle \lim_{h \to 0} \dfrac{f(x + h) - f(x)}{h}[/math] [math]\text{Therefore:}[/math] [math]\dfrac{df(x)g(x)}{dx} = \displaystyle \lim_{h \to 0} \dfrac{f(x + h) g(x + h) - f(x)g(x)}{h}[/math] [math]= \displaystyle \lim_{h \to 0} \dfrac{f(x + h) g(x + h) - f(x) g(x + h) + f(x) g(x + h) - f(x)g(x)}{h}[/math] [math]= \displaystyle \lim_{h \to 0} \dfrac{f(x + h) g(x + h) - f(x) g(x + h)}{h} + \displaystyle \lim_{h \to 0} \dfrac{f(x) g(x + h) - f(x)g(x)}{h}[/math] [math]= \displaystyle \lim_{h \to 0} \dfrac{f(x + h) g(x) - f(x) g(x)}{h} + \displaystyle \lim_{h \to 0} \dfrac{f(x) g(x + h) - f(x)g(x)}{h}[/math] [math]= \left(\displaystyle \lim_{h \to 0} \dfrac{f(x + h) - f(x)}{h}\right) g(x) + f(x) \left(\displaystyle \lim_{h \to 0} \dfrac{g(x + h) - g(x)}{h}\right)[/math] [math]= \dfrac{df(x)}{dx} g(x) + f(x) \dfrac{dg(x)}{dx}[/math] Edited May 10 by KJW 1
MigL Posted May 10 Posted May 10 I was just about to explain how calculus arose out of the limiting value for a function when the change approaches zero when you posted your excellent response; including the product rule, no less.
Genady Posted May 10 Posted May 10 36 minutes ago, KJW said: I don't know precisely how Newton or Leibniz obtained the product rule of differential calculus I didn't find it but since most Newton's derivations in Principia were geometrical, I assume that the product rule's derivation was geometrical as well. It could be as simple as this:
KJW Posted May 10 Posted May 10 I found this article titled "Calculus Before Newton and Leibniz - An in-depth article on the history of calculus". Here is the introductory section of the article: The Development of Calculus History has a way of focusing credit for any invention or discovery on one or two individuals in one time and place. The truth is not as neat. When we give the impression that Newton and Leibniz created calculus out of whole cloth, we do our students a disservice. Newton and Leibniz were brilliant, but even they weren’t capable of inventing or discovering calculus. The body of mathematics we know as calculus developed over many centuries in many different parts of the world, not just western Europe but also ancient Greece, the Middle East, India, China, and Japan. Newton and Leibniz drew on a vast body of knowledge about topics in both differential and integral calculus. The subject would continue to evolve and develop long after their deaths. What marks Newton and Leibniz is that they were the first to state, understand, and effectively use the Fundamental Theorem of Calculus. No two people have moved our understanding of calculus as far or as fast. But the problems that we study in calculus—areas and volumes, related rates, position/velocity/acceleration, infinite series, differential equations—had been solved before Newton or Leibniz was born. It took some 1,250 years to move from the integral of a quadratic to that of a fourth-degree polynomial. But awareness of this struggle can be a useful reminder for us. The grand sweeping results that solve so many problems so easily (integration of a polynomial being a prime example) hide a long conceptual struggle. When we jump too fast to the magical algorithm and fail to acknowledge the effort that went into its creation, we risk dragging our students past that conceptual understanding. This article explores the history of calculus before Newton and Leibniz: the people, problems, and places that are part of the rich story of calculus.
ImplicitDemands Posted May 10 Author Posted May 10 (edited) In my experience the content in the OP was learned before the power rule which is the basis of both the product rule and derivatives. 1 hour ago, KJW said: integration I found that through limits prior to learning (a/(b+1))^(b+1). You'd think looking at the power rule a^b; f(a)=(ba)^(b-1) that an integral would be f(a)=(a/b)^(b+1) where does that a/b+1 come from. Before learning integration I learned R=r+1 to work on a miss-written derivation problem: The diameter of the garden was actually 8", the radius was for 4" on the bottom lift picture there. Anyway the problem was admittedly miss-written. I learned why the integral adds one to the dividend which had the same value as the exponent from an error in the construction of a word problem. If I had to guess I'd assume I was in a simulation where a learning computer feeds off what I figure out to create more arithmetic. https://en.wikipedia.org/wiki/Intelligence_amplification Edited May 10 by ImplicitDemands
ImplicitDemands Posted May 17 Author Posted May 17 On 5/10/2024 at 10:28 AM, ImplicitDemands said: I had to take the zero out and find the 1-9 combinations. Then, I would have eliminated all the combinations with 9 in them, and I would have put 0 back in to replace the nine. Therefore, when computing for zero you'd have one set of 9^9 combinations where 9 is included and a second set of 9^9 combinations where 0 is included if 1 weren't repeating twice. So the 123456789 {0} is the equivalent of a square exponential, if you did the product rule it would be like you base^2, and then doing the process shown and swapping the 9 and the 0 would be like performing the product rule on base^2, that is (2 x base)^(2-1) = 2 x base. In essence, the difference between the 10^10 combinations of all the digits on your keyboard and the combinations you get when you sum up 1123456789 {0} & 1123456780 {9} is the same difference between base^2 and 2 x base (it's derivative). Let's take 4 unique digits for example: 1234 (1123, 2113, 2311, 1231, 1132, 3112, 3211, 1321) {4) + (1124, 2114, 2411, 1241, 1142, 4112, 4211, 1421) {3) = 16 different combinations. 4^4 unique combinations has derivative of 16^3, 16^3/4^4 = 16
ImplicitDemands Posted May 18 Author Posted May 18 (edited) I mean the power rule has to be a derivative of the combinatorics above, where there is only one repeating number that repeats once, in that combinatorics you have to leave one number out to find the number of combinations and then swap it for one other number. For 4 digits you could have 4 repeating once, twice, thrice, or all numbers could be 4, same for 3,2,1; where altogether including the 4 x 24 value, you have a total of 4^4 possible figurations. But when only one of those repeats once, it's only 16 possible configurations. The derivative 16^3=16 x 4^4. You can clearly see that the formula where the power rule comes into play is logn(x), where x=c times d, where if the total number of possible configurations is d=a^b and n=a times b. Finally, c=total number of possible configurations with one repeating number that repeats once as shown above. From that you can recognize the power rule a^b = (a times b)^(b-1), and its integral a^b=(a/(b+1))^(b+1) which has nothing to do with the economic optimization problem I had written out about the garden other than it using a the power rule one time. So if you had up all those functions in the OP picture you should get 10^18/10^10=10^8, log100(10^18)=9; (10^10)'=100^9 Edited May 18 by ImplicitDemands
ImplicitDemands Posted May 18 Author Posted May 18 (edited) 3 hours ago, ImplicitDemands said: I mean the power rule has to be a derivative of the combinatorics above, where there is only one repeating number that repeats once, in that combinatorics you have to leave one number out to find the number of combinations and then swap it for one other number. For 4 digits you could have 4 repeating once, twice, thrice, or all numbers could be 4, same for 3,2,1; where altogether including the 4 x 24 value, you have a total of 4^4 possible figurations. But when only one of those repeats once, it's only 16 possible configurations. The derivative 16^3=16 x 4^4. You can clearly see that the formula where the power rule comes into play is logn(x), where x=c times d, where if the total number of possible configurations is d=a^b and n=a times b. Finally, c=total number of possible configurations with one repeating number that repeats once as shown above. From that you can recognize the power rule a^b = (a times b)^(b-1), and its integral a^b=(a/(b+1))^(b+1) which has nothing to do with the economic optimization problem I had written out about the garden other than it using a the power rule one time. So if you had up all those functions in the OP picture you should get 10^18/10^10=10^8, log100(10^18)=9; (10^10)'=100^9 Obviously the OP was wrong about calculating the configurations but, given 10 x ((25 x 10!)/4!), 1123456789 {0} is like 1/100th of the possible combinations of repeating number so in theory there's plenty of room left for the rest of repeating configurations. And this is why the derivative of an exponent maximizes its dimensions. Edited May 18 by ImplicitDemands
ImplicitDemands Posted May 19 Author Posted May 19 (edited) Obviously if anyone 5 digits they would have seen why it is an incompatible way to proof the power rule. 6's log links to 2. Let's try it for next lower even number, 2 digits. 11 {2}, has a value of 1 configurations, x2 is 2. ((2x2)^(2-1))/2^2 = 1. If I did 8 digits (64^7/8^8=262144), I'm betting that since log64(262144)=3 you wouldn't have to do it with 8 again either, because 8's log links to 3. If you had 3 digits 113 131 311 = 3 and (9^2)/(3^3)=3 Let me see if I didn't do 4 wrong: 1123 {4} 1. 1123 2. 1132 3. 1231 4. 1321 5. 2311 6. 3211 7. 3112 8. 2113 9. 1213 10. 1312 11. 3121 12. 2131 1124 {3} 13. 1124 etc... 12 x 2 = 24 so yeah it wouldn't work like I said with 1 repeating it had 4 extra configurations in it. And even if I hadn't multiplied by 2, 16^3/4^4=16, so 4^4 x 12 wouldn't do it. It seems that Newton could have come up with the power rule by using one repeating digit in a combinatorics equation where the logarithms link, like 3 and 8, or 2 and 6. Edited May 19 by ImplicitDemands
ImplicitDemands Posted May 19 Author Posted May 19 (edited) 100^9/10^10=100000000->log100100000000=4 Yeah I'm betting this method wouldn't work for 10 digits since it links to 4 which didn't work. Maybe 12 digits? 144^11/12^12=61917364224->log14461917364224=5 Nope! 14? 196^13/14^14=56693912375296->log19656693912375296=6 ✓! So 2, 3, 6, 8, & 14 all check out for this method of placing one repeating digit into a combinatorics problem which I'm betting is where the power rule came from. Edited May 19 by ImplicitDemands
KJW Posted May 19 Posted May 19 (edited) On 5/11/2024 at 5:28 AM, ImplicitDemands said: I found that through limits prior to learning (a/(b+1))^(b+1). You'd think looking at the power rule a^b; f(a)=(ba)^(b-1) that an integral would be f(a)=(a/b)^(b+1) where does that a/b+1 come from. One thing about the integral of [math]x^n[/math] that I find interesting is the case of [math]n = -1[/math]: [math]\displaystyle \int x^n\, dx = \begin{cases}\ \dfrac{x^{n+1}}{n+1} + C & \text{if } n \neq -1 \\ \\ \ \log(x) + C &\text{if } n = -1 \end{cases}[/math] Note that: [math]\displaystyle \int x^{-1 + \varepsilon} \ dx = \dfrac{x^{\varepsilon}}{\varepsilon} + C[/math] for all [math]\varepsilon \neq 0[/math] regardless of how small [math]\varepsilon[/math] is. Furthermore, note that [math]x^{-1 - \varepsilon}[/math] can be deformed to [math]x^{-1 + \varepsilon}[/math] without discontinuity at [math]x^{-1}[/math]. Therefore, one would expect that: [math]\displaystyle \int x^{-1 - \varepsilon} \, dx[/math] can be deformed to: [math]\displaystyle \int x^{-1 + \varepsilon} \ dx[/math] without discontinuity at: [math]\displaystyle \int x^{-1} \ dx[/math] even though the above formula seems to indicate that this is not the case. But let's consider the definite integral: [math]\displaystyle \lim_{\varepsilon \to 0} \displaystyle \int_{1}^{x} u^{-1 + \varepsilon} \ du[/math] [math]= \displaystyle \lim_{\varepsilon \to 0} \dfrac{x^{\varepsilon} - 1}{\varepsilon}[/math] [math]= \log(x)[/math] Thus, it can be seen that the definite integral of [math]x^{-1 + \varepsilon}[/math] is continuous with respect to [math]\varepsilon[/math] at [math]x^{-1}[/math]. Interestingly, this notion can be extended to the definite integral of [math]\log(x)[/math] as follows: [math]\displaystyle \int_{1}^{x} \log(v) \ dv[/math] [math]= x \log(x) - x + 1[/math] And: [math]\displaystyle \lim_{\varepsilon \to 0} \displaystyle \int_{1}^{x} \displaystyle \int_{1}^{v} u^{-1 + \varepsilon} \ du \ dv[/math] [math]= \displaystyle \lim_{\varepsilon \to 0} \displaystyle \int_{1}^{x} \dfrac{v^{\varepsilon} - 1}{\varepsilon} \ dv[/math] [math]= \displaystyle \lim_{\varepsilon \to 0} \dfrac{x^{\varepsilon + 1}}{\varepsilon (\varepsilon + 1)} - \dfrac{x}{\varepsilon} - \dfrac{1}{\varepsilon (\varepsilon + 1)} + \dfrac{1}{\varepsilon}[/math] [math]= \displaystyle \lim_{\varepsilon \to 0} \dfrac{x^{\varepsilon + 1}}{\varepsilon (\varepsilon + 1)} - \dfrac{x (\varepsilon + 1)}{\varepsilon (\varepsilon + 1)} - \dfrac{1}{\varepsilon (\varepsilon + 1)} + \dfrac{(\varepsilon + 1)}{\varepsilon (\varepsilon + 1)}[/math] [math]= \displaystyle \lim_{\varepsilon \to 0} \dfrac{x^{\varepsilon + 1}}{\varepsilon (\varepsilon + 1)} - \dfrac{x \varepsilon}{\varepsilon (\varepsilon + 1)} - \dfrac{x}{\varepsilon (\varepsilon + 1)} - \dfrac{1}{\varepsilon (\varepsilon + 1)} + \dfrac{\varepsilon}{\varepsilon (\varepsilon + 1)} + \dfrac{1}{\varepsilon (\varepsilon + 1)}[/math] [math]= \displaystyle \lim_{\varepsilon \to 0} \dfrac{x^{\varepsilon + 1}}{\varepsilon (\varepsilon + 1)} - \dfrac{x \varepsilon}{\varepsilon (\varepsilon + 1)} - \dfrac{x}{\varepsilon (\varepsilon + 1)} + \dfrac{\varepsilon}{\varepsilon (\varepsilon + 1)}[/math] [math]= \displaystyle \lim_{\varepsilon \to 0} \dfrac{x^{\varepsilon + 1}}{\varepsilon} - \dfrac{x \varepsilon}{\varepsilon} - \dfrac{x}{\varepsilon} + \dfrac{\varepsilon}{\varepsilon}[/math] [math]= \displaystyle \lim_{\varepsilon \to 0} \dfrac{x^{\varepsilon + 1}}{\varepsilon} - x - \dfrac{x}{\varepsilon} + 1[/math] [math]= x \Big(\displaystyle \lim_{\varepsilon \to 0} \dfrac{x^{\varepsilon} - 1}{\varepsilon}\Big) - x + 1[/math] [math]= x \log(x) - x + 1[/math] However, if one starts with [math]x^{\varepsilon}[/math] and form the derivative: [math]\displaystyle \lim_{\varepsilon \to 0} \dfrac{dx^{\varepsilon}}{dx}[/math] [math]= \displaystyle \lim_{\varepsilon \to 0} \varepsilon x^{\varepsilon - 1}[/math] [math]= 0[/math] If we consider [math]\varepsilon[/math] to be small but not infinitesimal, then for the integral, we start with [math]x^{\varepsilon - 1}[/math] and end with [math]\dfrac{x^{\varepsilon}}{\varepsilon}[/math], whereas for the derivative, we start with [math]x^{\varepsilon}[/math] and end with [math]\varepsilon x^{\varepsilon - 1}[/math]. That is, the derivative is smaller than the integral by factor of [math]\varepsilon[/math], becoming zero in the limit. Thus, although repeated integration starting from [math]x^{\varepsilon - 1}[/math] can use the power function integration formula, the resulting sequence of functions are distinct from power functions obtained by starting from, for example, [math]x^0[/math]. Edited May 19 by KJW
ImplicitDemands Posted May 19 Author Posted May 19 (edited) 8 hours ago, KJW said: One thing about the integral of xn that I find interesting is the case of n=−1 : ∫xndx=⎧⎩⎨⎪⎪⎪⎪⎪⎪ xn+1n+1+C log(x)+Cif n≠−1if n=−1 Note that: ∫x−1+ε dx=xεε+C for all ε≠0 regardless of how small ε is. Furthermore, note that x−1−ε can be deformed to x−1+ε without discontinuity at x−1 . Therefore, one would expect that: ∫x−1−εdx can be deformed to: ∫x−1+ε dx without discontinuity at: ∫x−1 dx even though the above formula seems to indicate that this is not the case. But let's consider the definite integral: limε→0∫x1u−1+ε du =limε→0xε−1ε =log(x) Thus, it can be seen that the definite integral of x−1+ε is continuous with respect to ε at x−1 . Interestingly, this notion can be extended to the definite integral of log(x) as follows: ∫x1log(v) dv =xlog(x)−x+1 And: limε→0∫x1∫v1u−1+ε du dv =limε→0∫x1vε−1ε dv =limε→0xε+1ε(ε+1)−xε−1ε(ε+1)+1ε =limε→0xε+1ε(ε+1)−x(ε+1)ε(ε+1)−1ε(ε+1)+(ε+1)ε(ε+1) =limε→0xε+1ε(ε+1)−xεε(ε+1)−xε(ε+1)−1ε(ε+1)+εε(ε+1)+1ε(ε+1) =limε→0xε+1ε(ε+1)−xεε(ε+1)−xε(ε+1)+εε(ε+1) =limε→0xε+1ε−xεε−xε+εε =limε→0xε+1ε−x−xε+1 =x(limε→0xε−1ε)−x+1 =xlog(x)−x+1 However, if one starts with xε and form the derivative: limε→0dxεdx =limε→0εxε−1 =0 If we consider ε to be small but not infinitesimal, then for the integral, we start with xε−1 and end with xεε , whereas for the derivative, we start with xε and end with εxε−1 . That is, the derivative is smaller than the integral by factor of ε , becoming zero in the limit. Thus, although repeated integration starting from xε−1 can use the power function integration formula, the resulting sequence of functions are distinct from power functions obtained by starting from, for example, x0 . I misspoke, bottom line is there is no difference between the integral and the derivative, the derivative a^b multiplies b and subtracts it by one in the exponent, the integral doesn't change the original b value that gets subtracted in the derivation which is why it has to divide by b+1. Bottom line is I falsely claimed it had something to do with limits. Really it is a product of combinatorics with one repeated digit. In my previous post I write that it doesn't work for the combinatorial of 4 where you have one digit repeat, I believe this is because 4 is a square. It worked for 2, 3, and 6. But I also believe it should work for 5, 8, 12 and 14 digits because none of those are squares. It probably wouldn't work for 9. Bottom line is the number of configurations where there is one repeated digit is exactly equal to dividend of the derivative of a number of digits and that same number of digits if that number is not a square, this has been repeatedly proved in this topic and is so uncanny that I picked one repeated digit to start examining password combinations just before taking calculus using math and that's what worked with the power rule, without which there would be no calculus, that I'm seriously questioning the nature of my reality. Edited May 19 by ImplicitDemands
ImplicitDemands Posted May 21 Author Posted May 21 (edited) On 5/10/2024 at 2:28 PM, ImplicitDemands said: In my experience the content in the OP was learned before the power rule which is the basis of both the product rule and derivatives. I found that through limits prior to learning (a/(b+1))^(b+1). You'd think looking at the power rule a^b; f(a)=(ba)^(b-1) that an integral would be f(a)=(a/b)^(b+1) where does that a/b+1 come from. Before learning integration I learned R=r+1 to work on a miss-written derivation problem: The diameter of the garden was actually 8", the radius was for 4" on the bottom lift picture there. Anyway the problem was admittedly miss-written. I learned why the integral adds one to the dividend which had the same value as the exponent from an error in the construction of a word problem. If I had to guess I'd assume I was in a simulation where a learning computer feeds off what I figure out to create more arithmetic. https://en.wikipedia.org/wiki/Intelligence_amplification It was this after all, nothing to do with combinatorics. I realized this when the number of configurations I got for 6 digits turned out to be wrong. Anyway the full limits proof from which the power rule was derived from is as follows: The calculator is telling me that it doesn't work (i.e. is "true") for ((r+1)(r-1)^2)/((r-1)(r+1)^2) or its inverse ((r-1)(r+1)^2)/((r+1)(r-1)^2)), so I arrived at ((r-1)^3)/(r+1)(r-1)^2 through sheer trial and error. So if R=r+1, looking at that (r-1)^3/(r(r-1)^2); if r=3, it looks exactly like an integral since 6^2=36 and 2^3=8->8/36=2^3/6^2 because if 6^2 is the derivative of 2^3 then 2^3 is the integral of 6^2 which is (6/(2+1))^(2+1) = 2^3 it's almost plain as day how Newton found the formula from limits here. Any questions? We're done here. Edited May 21 by ImplicitDemands
ImplicitDemands Posted May 22 Author Posted May 22 (edited) Correction, it was (r+1)^3 for the power rule. I see it now. r^3+3r^2+3r+1 If you look at this polynomial it shows f(r^3)''' or all the derivations of r^3 until it is 1. (3 x r)^(3-1) = 3r^2 (1 x 3)r^(2-1) = 3r (1 x 3)r^(1-1)=3r^0 = 1 Edited May 22 by ImplicitDemands
HopDavid Posted October 13 Posted October 13 On 5/10/2024 at 12:33 AM, exchemist said: Leibnitz and Newton are thought to have independently developed calculus at around the same time, though I think it was Leibnitz who published first. As so often with science and mathematics, the idea was germinating at the time and trying to determine who got there "first" is rather debatable and of limited value. Both men most certainly existed, though. I can't make sense of the rest of your post. So far as I know Fermat was the first to use methods of exhaustion to find the slope of a tangent to a curve. So I would call Fermat the father of differential calculus. I am told Descartes also did this. And I'd argue Cavalieri is the father of integral calculus. See Cavalieri's quadrature: Link. Cavalieri's quadrature is familiar to most 1st year calculus students. The fundamental theorem of calculus has been discovered by James Gregory and Isaac Barrow. Isaac Barrow was Newton's older colleague at Cambridge. Also helping lay the foundations of calculus in the generation prior to Newton and Leibniz was Descartes, Kepler, Wallis and others. Historian Thony Christie argues that inventing (or discovering) calculus was the collaborative effort of many people over many years. See his essay The Wrong Question: Link.
Recommended Posts
Create an account or sign in to comment
You need to be a member in order to leave a comment
Create an account
Sign up for a new account in our community. It's easy!
Register a new accountSign in
Already have an account? Sign in here.
Sign In Now