• @IAm_A_Complete_Idiot@sh.itjust.works
    link
    fedilink
    7
    edit-2
    8 months ago

    According to the benchmark in the article it’s already way faster at n = 1000. I think you’re overestimating the cost of multiplication relative to just cutting down n logarithmically.

    log_2(1000) = roughly a growth factor of 10. 2000 would be 11, and 4000 would be 12. Logs are crazy.

    • @cbarrick@lemmy.world
      cake
      link
      fedilink
      English
      28 months ago

      The article is comparing to the dynamic programming algorithm, which requires reading and writing to an array or hash table (the article uses a hash table, which is slower).

      The naive algorithm is way faster than the DP algorithm.

      • @t_veor@sopuli.xyz
        link
        fedilink
        English
        28 months ago

        It’s not that hard to check yourself. Running the following code on my machine, I get that the linear algebra algorithm is already faster than the naive algorithm at around n = 100 or so. I’ve written a more optimised version of the naive algorithm, which is beaten somewhere between n = 200 and n = 500.

        Try running this Python code on your machine and see what you get:

        import timeit
        
        def fib_naive(n):
            a = 0
            b = 1
            while 0 < n:
                b = a + b
                a = b - a
                n = n - 1
            return a
        
        def fib_naive_opt(n):
            a, b = 0, 1
            for _ in range(n):
                a, b = b + a, b
            return a
        
        def matmul(a, b):
            return (
                (a[0][0] * b[0][0] + a[0][1] * b[1][0], a[0][0] * b[0][1] + a[0][1] * b[1][1]),
                (a[1][0] * b[0][0] + a[1][1] * b[1][0], a[1][0] * b[0][1] + a[1][1] * b[1][1]),
            )
        
        def fib_linear_alg(n):
            z = ((1, 1), (1, 0))
            y = ((1, 0), (0, 1))
            while n > 0:
                if n % 2 == 1:
                    y = matmul(y, z)
                z = matmul(z, z)
                n //= 2
        
            return y[0][0]
        
        def time(func, n):
            times = timeit.Timer(lambda: func(n)).repeat(repeat=5, number=10000)
            return min(times)
        
        for n in (50, 100, 200, 500, 1000):
            print("========")
            print(f"n = {n}")
            print(f"fib_naive:\t{time(fib_naive, n):.3g}")
            print(f"fib_naive_opt:\t{time(fib_naive_opt, n):.3g}")
            print(f"fib_linear_alg:\t{time(fib_linear_alg, n):.3g}")
        

        Here’s what it prints on my machine:

        ========
        n = 50
        fib_naive:      0.0296
        fib_naive_opt:  0.0145
        fib_linear_alg: 0.0701
        ========
        n = 100
        fib_naive:      0.0652
        fib_naive_opt:  0.0263
        fib_linear_alg: 0.0609
        ========
        n = 200
        fib_naive:      0.135
        fib_naive_opt:  0.0507
        fib_linear_alg: 0.0734
        ========
        n = 500
        fib_naive:      0.384
        fib_naive_opt:  0.156
        fib_linear_alg: 0.112
        ========
        n = 1000
        fib_naive:      0.9
        fib_naive_opt:  0.347
        fib_linear_alg: 0.152