This question is motivated by me trying to finish my answer to this question.
In the 2019 SI redefinition the kg was redefined in terms of Planck's constant (and the second and meter):
$$ 1 \text{ kg}= (h) \times (1 \text{ s}) \times (1 \text{ m})^{-2} \times (6.62607015\times 10^{34}) $$
This means to realize the meter you must have a measurement of mass that is directly related, in some way, to Planck's constant. This is currently realized at precision metrology labs using the Kibble balance. The crude idea of the Kibble balance is you use electromagnets to generate a force that cancels the weight of a mass (you independently measure local acceleration due to gravity so you can convert weight to mass). You do some tricks and measure some voltages and currents related to the magnetic force you are leveraging. You actually measure those electrical parameters using calibrated electrical standards based on superconducting Josephson junctions. Because Josephson junctions rely on the quantum Hall effect it turns out something about them is quantized in units of Planck's constant $h$. When you express the mass of the object in terms of those voltages and currents you can proceed to express it in terms of $h$ as needed for a primary mass standard.
A major downside of this method, while establishing a non-artifact based SI standard, is that it requires monetarily and technically expensive superconducting Josephson junctions.
My question is, is there any way (even if it is inaccurate and imprecise) to realize a primary mass standard based on the current SI definition of the kg, that does not rely on superconducting Josephson junctions? Or more generally, that does not rely on the quantum Hall effect or any device that requires advanced fabrication techniques?
As a starting point, this youtube video explains how a voltage can pretty easily be calibrated to Planck's constant and the electron charge. Basically, an LED with wavelength $\lambda = c/\nu$ emits photons with energy $h\nu$. This means it won't turn on until the voltage driving it is $V > e h \nu = e h c/\lambda$. If we assume the meter has been realized, then you create a diffraction grating setup and use that to measure $\lambda$. The problems are, even though this voltage is calibrated to microscopic quantities, (1) the voltage depends on $e$ as well as $h$ and (2) the Kibble balance requires a precisely measured voltage AND current. This method does not produce a calibrated current. (One could validly argue the LEDs are "devices that require advanced fabrication techniques" but I'd rebut that there is a qualitative difference with them compared to quantum Hall devices in that I can go to the store and get an LED for $0.05 [or get one out of some electronic box in my house] and they don't require cyrogenic cooling to operate.)
A shorter way to ask this question might be: Is there a way, without the quantum Hall effect, to refer a current measurement to Planck's constant $h$?
update: I had a recollection about an ultracold atomic physics experiment I had heard about which involves kicking a single atom with a photon with a known wavelength and watching how the photon recoils as way to measure mass in a way that is related to Plancks constant. See Lan et. al., Science Vol 339 Issue 6119 (2013). Apologies for the paywall. Whiles this approach gives an affirmative answer to the titular question, it still requires a very complex apparatus including many lasers, ultracold atoms, and an ultra-high vacuum chamber. The question of a "simple" primary standard realization of the kg remains.