I see your perspective and I think you kinda miss my perspective which I am to blame for.
I don’t say there weren’t improvements. I am saying that given the uncertainty of “goodness”. Maybe we shouldn’t idolize it. You can appreciate the attempt of creating memory safe code through a programing language without thinking the bare metal code should be written in that language. You can like a typeless easy to write language like Js without thinking desktop app should be written in it. You can like the idea behind functional programming while believing that any application is in the end about side effects and therefore a purely functional application impossible.
You can approach the whole topic as an area of study and possible technological advances instead of a dogma.
You can like the idea behind functional programming while believing that any application is in the end about side effects and therefore a purely functional application impossible.
It’s a bit of a tangent, but if you’re doing something completely deterministic and non-interactive, like computing a digit of pi, it’s great in practice as well. I use Haskell semi-regularly for that kind of thing.
You could argue printing the output is a side effect, but is a side effect followed by termination really “side”?
I see your perspective and I think you kinda miss my perspective which I am to blame for.
I don’t say there weren’t improvements. I am saying that given the uncertainty of “goodness”. Maybe we shouldn’t idolize it. You can appreciate the attempt of creating memory safe code through a programing language without thinking the bare metal code should be written in that language. You can like a typeless easy to write language like Js without thinking desktop app should be written in it. You can like the idea behind functional programming while believing that any application is in the end about side effects and therefore a purely functional application impossible.
You can approach the whole topic as an area of study and possible technological advances instead of a dogma.
Oh, well I can agree with that.
It’s a bit of a tangent, but if you’re doing something completely deterministic and non-interactive, like computing a digit of pi, it’s great in practice as well. I use Haskell semi-regularly for that kind of thing.
You could argue printing the output is a side effect, but is a side effect followed by termination really “side”?