Is the amount of heat produced by a computer mining bitcoin the same as a traditional heater, if they have the same input power?
Yes$^1$. In fact all heat is the same. There is only one heat. Be it the thermal energy stored in a mass by virtue of its temperature or be it the heat being generated at the core of stars. In fact, all forms of energy are same and in theory, inter-convertible. They differ only in the manner they are generated.
I know of course about the law of conservation of energy, but it feels that some "work" has been done on the information in the system, and thus the waste heat energy produced would be less.
The amount of waste energy is not dependent on what the input power is being used for. A traditional heater of the same effective power and efficiency as a mining computer would also do the same work as the computer: the difference is that it does work to produce thermal energy (separate form the waste, err, thermal energy) while the computer does work to process "information".
I would like to point out here, that the waste heat generated in a computer, during mining or any other regular task is mostly not the heat generated during erasure of information (which coincidentally, also generated heat but is far less than what current architecture efficiencies already loose) but rather the joule heating of circuits, leakage currents and other losses.
I know this mining is not work in the sense W=Fd
It is. All work is ultimately, at the fundamental level, $W=\vec{F}.\vec{d}$ since that is the definition of work. Be it work done to heat stuff, work done in muscles, work done in computers. Reasoning how exactly that happens if its not evident at the face of the system, is another matter altogether.
but it does seem to be decreasing the entropy in the universe by organizing information. It feels intuitive that some energy should be used up in the very computation itself.
I commend your intuition. The fact that information erasure must necessarily accompany an energy cost is in fact true, at least in classical thermodynamics. Erasure of information is an irreversible process. Such erasures are necessary in any device using boolean logic since classical gates are irreversible. This loss then sets the lower limit on how much energy must be wasted in irreversible information erasures. As mentioned before, real-world devices waste a lot more than this.
If not, then wouldn't computation be free and hence it could be used to decrease entropy (by organizing information) without using energy?
The fact that irreversible information erasure must accompany energy expenditure and net entropy increase are related as both deal with counting the states of the system. However, I am not clear on the specifics.
So, finally
Does bitcoin mining take work waste energy ?
Yes, but not in any way different from any other heat engine. From an information perspective, the only unique thing about a compute process, is the energy loss from information erasure, which in the real world is minuscule relative to other losses.
$^1$ if they have the same efficiency but that's understood I guess.