What's new inHYPERFINITE FACTORS, PADIC LENGTH SCALE HYPOTHESIS, AND DARK MATTER HIERARCHYNote: Newest contributions are at the top! 
Year 2015 
Could electrolysis involve dark matter and new physics?During years I have many times tried to understand what happens in electrolysis and every time I have been forced to to admit that I do not! Very embarrassing observation. I have tried to gain wisdom from an old chemistry book with 1000 pages again and again but always in vain. This is especially embarrassing because a unified theory builder to be taken seriously is expected to build brave new brane worlds in 11 or 12 dimensions to possibly explain a possible detected particle at mass 750 GeV at LHC instead of trying to understand age old little problems solved aeons ago. The waucoefficient of chemistry is zero as compared to the awesome 10^{500} of Mtheory. Energetics has been my personal problem (besides funding). I learn from chemistry book that an electric field  say voltage of 2 V per 1 mm splits molecules to ions. The bond energies of molecules are in few eV range. For instance, OH bond has 5 eV energy. V= 2V/mm electric field corresponds to electrostatic energy E=eVd∼ 2^{10} eV energy gain for a unit charge moving from the end of the bond to the other one. This is incredibly small energy and to my understanding should have absolutely no effect to the state molecule. Except that it has! A heretic thought: could it be that chemists have just accepted this fact (very reasonable!) and built their models as mathematical parameterizations without any attempt to understand what really happens? Could the infinite vanity of theoretical physicists have prevented them from lowering themselves to the intellectual level of chemists and prevented them from seeing that electrolysis is not at all understood? In order that this kind of energy would have so drastic effect as splitting molecule to pieces, the system molecule + yet unidentified "something" must be in critical state. Something at the top of hill so that even slightest perturbation makes it fall down. The technical term is criticality or even quantum criticality.
Entire spectrum of values of h_{eff}/h is possible. For laser pulse induced fusion (see the article ) assumed to induce longitudinal compression one would have h_{eff}/h≈ 10^{3}. Dark nuclear physics with nonstandard values of Planck constant would be a crucial element of electrolysis. Condensed matter physics and nuclear physics would not live in totally separate compartments and dark matter an ordinary matter would interact! How humiliating for theoreticians! I do not hear the derisive laughter of superstring theoreticians anymore! Ordinary electrolysis would thus produce dark nuclei. The problem is that most of them would leak out from the system along dark flux tubes and potentially available nuclear energy is lost! As also various elements so badly needed by modern technosociety! For instance, in the splitting of water to hydrogen, the flux tubes assignable to the beam containing hydrogen would take the dark nuclei away. Could one transform dark nuclei to ordinary ones?
See the article Cold Fusion Again or the chapter with the same title. 
Cold fusion againDuring years I have developed two models of cold fusion and in this chapter these models are combined together. The basic idea of TGD based model of cold is that cold fusion occurs in two steps. First dark nuclei (large h_{eff}=n× h) with much lower binding energy than ordinary nuclei are formed at magnetic flux tubes possibly carrying monopole flux. These nuclei can leak out the system along magnetic flux tubes. Under some circumstances these dark nuclei can transform to ordinary nuclei and give rise to detectable fusion products. An essential additional condition is that the dark protons can decay to neutrons rapidly enough by exchanges of dark weak bosons effectively massless below atomic length scale. This allows to overcome the Coulomb wall and explains why final state nuclei are stable and the decay to ordinary nuclei does not yield only protons. Thus it seems that this model combined with the TGD variant of WidomLarsen model could explain nicely the existing data. I will describe the steps leading to the TGD inspired model for cold fusion combining the earlier TGD variant of WidomLarsen model with the model inspired by the TGD inspired model of Pollack's fourth phase of water using as input data findings from laser pulse induced cold fusion discovered by Leif Holmlid and collaborators. I consider briefly also alternative options (models assuming surface plasma polariton and heavy electron). After that I apply TGD inspired model in some cases (PonsFleischman effect, bubble fusion, and LeClair effect). The model explains the strange findings about cold fusion  in particular the fact that only stable nuclei are produced  and suggests that also ordinary nuclear reactions might have more fundamental description in terms of similar model. See the article Cold Fusion Again or the chapter with the same title. 
Analogs of quantum matrix groups from finite measurement resolution?The notion of quantum group replaces ordinary matrices with matrices with noncommutative elements. This notion is physically very interesting, and in TGD framework I have proposed that it should relate to the inclusions of von Neumann algebras allowing to describe mathematically the notion of finite measurement resolution (see this). These ideas have developed slowly through various side tracks. In it is interesting to consider the notion of quantum matrix inspired by recent view about quantum TGD. It turns out that under some additional conditions this approach provides a concrete representation and physical interpretation of quantum groups in terms of finite measurement resolution.
Quantum matrices define a more general structure than quantum group but provide a concrete representation for them in terms of finite measurement resolution if q is a root of unity. For q=+/ 1 (BoseEinstein or FermiDirac statistics) one obtains quantum matrices for which the determinant is apart from possible change by sign factor invariant under the permutations of both rows and columns. One can also understand the recursive fractal structure of inclusion sequences of hyperfinite factors resulting by replacing operators appearing as matrix elements with quantum matrices and a concrete connection with quantum groups emerges. In Zero Energy Ontology (ZEO) Mmatrix serving as the basic building brick of unitary Umatrix and identified as a hermitian square root of density matrix provides a possible application for this vision. Especially fascinating is the possibility of hierarchies of measurement resolutions represented as inclusion sequences realized as recursive construction of Mmatrices. Quantization would emerge already at the level of complex numbers appearing as Mmatrix elements. This approach might allow to unify various ideas behind TGD. For instance, Yangian algebras emerging naturally in twistor approach are examples of quantum algebras. The hierarchy of Planck constants should have a close relationship with inclusions and fractal hierarchy of subalgebras of supersymplectic and other conformal algebras. See the article Analogs of quantum matrix groups from finite measurement resolution? or the chapter Evolution of Ideas about Hyperfinite Factors in TGD. 
Evidence for macroscopic quantum coherence of fluid flow at criticalityEvidence for the hierarchy of Planck constants implying macroscopic quantum coherence in quantum critical systems is rapidly accumulating. Also people having the courage to refer to TGD in their articles are gradually emerging. The most recent fluid dynamics experiment providing this kind of evidence is performed by Yves Couder and Emmanuel Fort (see for instance the article Single particle diffraction in macroscopic scale). Mathematician John W. M. Bush has commented these findings in the Proceedings of National Academy of Sciences and the article provides references to a series of papers by Couder and collaborators. The system studied consist of a tray containing water at a surface, which is oscillating. The intensity of vibration is just below the critical value inducing so called Faraday waves at the surface of water. Although the water surface is calm, water droplet begins to bounce and generates waves propagating along the water surface  "walkers". Walkers behave like classical particles at Bohr orbits. As they pass through a pair of slits they behave they choose random slit but several experiments produce interference pattern. Walkers exhibit an effect analogous to quantum tunneling and even the analogs of quantum mechanical bound states of walkers realized as circular orbits emerge as the water tray rotates! The proposed interpretation of the findings is in terms of Bohm's theory. Personally I find it very difficult to believe in this since Bohm's theory has profound mathematical difficulties. Bohm's theory was inspired by Einstein's belief on classical determinism and the idea that quantum nondeterminism is not actual but reduces to the presence of hidden variables. Unfortunately, this idea led to no progress. TGD is analogous to Bohm's theory in that classical theory is exact but quantum theory is now only an exact classical correlate: there is no attempt to eliminate quantum nondeterminism. Quantum jumps are between superpositions of entire classical time evolutions rather than their time=constant snapshots: this solves the basic paradox of Copenhagen interpretation. A more refined formulation is in terms of zero energy ontology, which in turn forces to generalize quantum measurement theory to a theory of consciousness. Macroscopic quantum coherence associated with the behavior of droplets bouncing on the surface of water is suggested by the experiments. For instance, quantum measurement theory seems to apply to the behavior of single droplet as it passes through slit. In TGD the prerequisite for macroscopic quantum coherence would be quantum criticality at which large h_{eff}=n×h is possible. There indeed is an external oscillation of the tray containing water with an amplitude just below the criticality for the generation of Faraday waves at the surface of water. Quantum classical correspondence states that the quantum behavior should have a classical correlate. The basic structure of classical TGD is that of hydrodynamics in the sense that dynamics reduces to conservation laws plus conditions expressing the vanishing of an infinite number of so called supersymplectic charges  the conditions guarantee strong form of holography and express quantum criticality. The generic solution of classical field equations could reduce to Frobenius integrability conditions guaranteing that the conserved isometry currents are integrable and thus define global coordinates varying along the flow lines. One should be of course very cautious. For ordinary Schrödinger equation the system is closed. Now the system is open. This is not a problem if the only function of external vibration is to induce quantum criticality. The experiment brings in mind the old vision of Frölich about external vibrations as induced of what looks like quantum coherence. In TGD framework this coherence would be forced coherence at the level of visible matter but the oscillation itself would correspond to genuine macroscopic quantum coherence and large value of h_{eff}. A standard example are penduli, which gradually start to oscillate in unisono in presence of weak synchronizing signal. In brain neurons would start to oscillator synchronously by the presence of dark photons with large h_{eff}. See the chapter Criticality and dark matter. 
TGD inspired view about blackholes and Hawking radiation: part IThe most recent revealation of Hawking was in Hawking radiation conference held in KTH Royal Institute of Technology in Stockholm. The title of the posting of Bee telling about what might have been revealed is "Hawking proposes new idea for how information might escape from black holes". Also Lubos has  a rather aggressive  blog post about the talk. A collaboration of Hawking, Andrew Strominger and Malcom Perry is behind the claim and the work should be published within few months.The first part of posting gives a critical discussion of the existing approach to black holes and Hawking gravitation. The intention is to demonstrate that a pseudo problem following from the failure of General Relativity below black hole horizon is in question. In the second past of posting I will discuss TGD view about blackholes and Hawking radiation. There are several new elements involved but concerning black holes the most relevant new element is the assignment of Euclidian spacetime regions as lines of generalized Feynman diagrams implying that also blackhole interiors correspond to this kind of regions. Negentropy Maximization Principle is also an important element and predicts that number theoretically defined black hole negentropy can only increase. The real surprise was that the temperature of the variant of Hawking radiation at the flux tubes of proton Sun system is room temperature! Could TGD variant of Hawking radiation be a key player in quantum biology? Is information lost or not in blackhole collapse? The basic problem is that classically the collapse to blackhole seems to destroy all information about the matter collapsing to the blackhole. The outcome is just infinitely dense mass point. There is also a theorem of classical GRT stating that blackhole has no hair: blachole is characterized only by few conserved charges. Hawking has predicted that blackhole loses its mass by generating radiation, which looks like thermal. As blackhole radiates its mass away, all information about the material which entered to the blackhole seems to be lost. If one believes in standard quantum theory and unitary evolution preserving the information, and also forgets the standard quantum theory's prediction that state function reductions destroy information, one has a problem. Does the information really disappear? Or is the GRT description incapable to cope with the situation? Could information find a new representation? Superstring models and AdS/CFT correspondence have inspired the proposal that a hologram results at the horizon and this hologram somehow catches the information by defining the hair of the blackhole. Since the radius of horizon is proportional to the mass of blackhole, one can however wonder what happens to this information as the radius shrinks to zero when all mass is Hawking radiated out. What Hawking suggests is that a new kind of symmetry known as supertranslations  a notion originally introduced by Bondi and Metzner  could somehow save the situation. Andrew Strominger has recently discussed the notion. The information would be "stored to supertranslations". Unfortunately this statement says nothing to me nor did not say to Bee and New Scientist reporter. The idea however seems to be that the information carried by Hawking radiation emanating from the blackhole interior would be caught by the hologram defined by the blackhole horizon. Supertranslation symmetry acts at the surface of a sphere with infinite radius in asymptotically flat spacetimes looking like empty Minkowski space in very distant regions. The action would be translations along sphere plus Poincare transformations. What comes in mind in TGD framework is conformal transformations of the boundary of 4D lightcone, which act as scalings of the radius of sphere and conformal transformations of the sphere. Translations however translate the tip of the lightcone and Lorentz transformations transform the sphere to an ellipsoid so that one should restrict to rotation subgroup of Lorentz group. Besides this TGD allows huge group of symplectic transformations of δ CD× CP_{2} acting as isometries of WCW and having structure of conformal algebra with generators labelled by conformal weights. Sharpening of the argument of Hawking There is now a popular article explaining the intuitive picture behind Hawking's proposal. The blackhole horizon would involve tangential flow of light and particles of the infalling matter would induce supertranslations on the pattern of this light thus coding information about their properties to this light. After that this light would be radiated away as analog of Hawking radiation and carry out this information. The objection would be that in GRT horizon is no way special  it is just a coordinate singularity. Curvature tensor does not diverge either and Einstein tensor and Ricci scalar vanish. This argument has been used in the firewall debates to claim that nothing special should occur as horizon is traversed. Why light would rotate around it? No reason for this! The answer in TGD framework would be obvious: horizon is replaced for TGD analog of blackhole with a lightlike 3surface at which the induced metric becomes Euclidian. Horizon becomes analogous to light front carrying not only photons but all kinds of elementary particles. Particles do not fall inside this surface but remain at it! What are the problems? My fate is to be an aggressive dissident listened by noone, and I find it natural to continue in the role of angry old man. Be cautious, I am arrogant, I can bite, and my bite is poisonous!
See the chapter Criticality and dark matter or the article TGD view about black holes and Hawking radiation. 
TGD inspired view about blackholes and Hawking radiation: part IITGD view about black holes and Hawking radiation: part IIIn the second part of posting I discuss TGD view about blackholes and Hawking radiation. There are several new elements involved but concerning black holes the most relevant new element is the assignment of Euclidian spacetime regions as lines of generalized Feynman diagrams implying that also blackhole interiors correspond to this kind of regions. Negentropy Maximization Principle is also an important element and predicts that number theoretically defined black hole negentropy can only increase. The real surprise was that the temperature of the variant of Hawking radiation at the flux tubes of proton Sun system is room temperature! Could TGD variant of Hawking radiation be a key player in quantum biology? The basic ideas of TGD relevant for blackhole concept My own basic strategy is to not assume anything not necessitated by experiment or not implied by general theoretical assumptions  these of course represent the subjective element. The basic assumptions/predictions of TGD relevant for the recent discussion are following.
The analogs of blackholes in TGD Could blackholes have any analog in TGD? What about Hawking radiation? The following speculations are inspired by the above general vision.
The reduction sequences consist of life cycles at fixed boundary and the size of blackhole like state as of any state is expected to increase in discrete steps if it participates to cosmic expansion in average sense. This requires that the mass of blackhole like object gradually increases. The interpretation is that ordinary matter gradually transforms to dark matter and increases dark mass M= R/G. Cosmic expansion is not observed for the sizes of individual astrophysical objects, which only comove. The solution of the paradox is that they suddenly increase their size in state function reductions. This hypothesis allows to realize Expanding Earth hypothesis in TGD framework (see this). Number theoretically preferred scalings of blackhole radius come as powers of 2 and this would be the scaling associated with Expanding Earth hypothesis. See the chapter Criticality and dark matter or the article TGD view about black holes and Hawking radiation. 
About negentropic entanglement as an analog of error correcting codeIn classical computation, the simplest manner to control errors is to take several copies of the bit sequences. In quantum case nocloning theorem prevents this. Error correcting codes code n information qubits to the entanglement of N>n physical qubits. Additional contraints represents the subspace of nqubits as a lowerdimensional subspace of N qubits. This redundant representation is analogous to the use of parity bits. The failure of the constraint to be satisfied tells that the error is present and also the character of error. This makes possible the automatic correction of the error is simple enough  such as the change of the phase of spin state or or spin flip.Negentropic entanglement (NE) obviously gives rise to a strong reduction in the number of states of tensor product. Consider a system consisting of two entangled systems consisting of N_{1} and N_{2} spins. Without any constraints the number of states in state basis is 2^{N1}× 2^{N2} and one as N_{1}+N_{2} qubits. The elements of entanglement matrix can be written as E_{A,B}, A== ⊗_{i=1}^{N1} (m_{i},s_{i}), B== ⊗_{k=1}^{N2} (m_{k},s_{k}) in order to make manifest the tensor product structure. For simplicity one can consider the situation N_{1}=N_{2}=N. The unnormalized general entanglement matrix is parametrized by 2× 2^{2N} independent real numbers with each spin contributing two degrees of freedom. Unitary entanglement matrix is characterized by 2^{2N} real numbers. One might perhaps say that one has 2N real bits instead of almost 2N+1 real qubits. If the time evolution according to ZEO respects the negentropic character of entanglement, the sources of errors are reduced dramatically. The challenge is to understand what kind of errors NE eliminates and how the information bits are coded by it. NE is respected if the errors act as unitary transformations E→ UEU^{†} of the unitary entanglement matrix. One can consider two interpretations.
Gauge invariance has turned out to be a fundamental symmetry principle, and one can ask whether unitary entanglement matrices assuming that only the eigenvalues matter, could give rise to a simulation of discrete gauge theories. The reduction of the information to that provided by the diagonal form be interpreted as an analog of gauge invariance?

Quantum measurement and quantum computation in TGD UniverseDuring years I have been thinking how quantum computation could be carried out in TGD Universe (see this). There are considerable deviations from the standard view. Zero Energy Ontology (ZEO), weak form of NMP dictating the dynamics of state function reduction, negentropic entanglement (NE), and hierarchy of Planck constants define the basic differences between TGD based and standard quantum measurement theory. TGD suggests also the importance of topological quantum computation (TQC) like processes with braids represented as magnetic flux tubes/strings along them.The natural question that popped in my mind was how NMP and Zero Energy Ontology (ZEO) could affect the existing view about TQC. The outcome was a more precise view about TQC. The basic observation is that the phase transition to dark matter phase reduces dramatically the noise affecting quantum quits. This together with robustness of braiding as TQC program raises excellent hopes about TQC in TGD Universe. The restriction to negentropic spacelike entanglement (NE) defined by a unitary matrix is something new but does not seem to have any fatal consequences as the study of Shor's algorithm shows. NMP strongly suggests that when a pair of systems  the ends of braid  suffer state function reduction, the NE must be transferred somehow from the system. How? The model for quantum teleportation allows to identify a possible mechanism allowing to achieve this. This mechanism could be fundamental mechanism of information transfer also in living matter and phosphorylation could represent the transfer of NE according to this mechanism: the transfer of metabolic energy would be at deeper level transfer of negentropy. Quantum measurements could be actually seen as transfer of negentropy at deeper level. For details see the chapter Negentropy Maximization Principleor the article Quantum Measurement and Quantum Computation in TGD Universe. 
Can bacteria induce superfluidity?Claims about strange experimental findings providing support for TGD have started to accumulate in accelerating pace. During about week I have learned about four anomalies! The identification of the dark matter as h_{eff} phases is the common denominator of the explanations of these findings.
As the number of bacteria (E. coli) was increased, the viscosity associated with shear stress (the viscous force parallel to the surface) dropped: this in accordance with theoretical expectations. Adding about 6 billion cells (the fluid volume is not mentioned but it seems that the effect occurs above critical density of bacteria), the apparent viscosity dropped to zero  or more precisely, below the experimental resolution. The superfluid like behavior was preserved above the critical concentration. What is important that this did not happen for dead bacteria: bacteria play an active role in the reduction of viscosity. Researchers are not able to identify the mechanism leading to the superfluidlike behavior but some kind of collective effect is believed to be in question. The findings suggest that the flagellae  kind of spinning hairs used by the bacteria to propel themselves  should play an essential part in the phenomenon. As bacteria swim, they fight against current, decreasing the local forces between molecules that determine the fluid's viscosity. Above critical density the local effects would somehow become global. Cates et al have proposed this kind of phenomenon: see the article "Shearing Active Gels Close to the IsotropicNematic Transition". The authors speak in the abstract about zero apparent viscosity.
See the chapter Criticality and dark matter". 
Topological order in Quantum TGDTopological order is a rather advanced concept of condensed matter physics. There are several motivations for the notion of topological order in TGD.
In the article Topological order and Quantum TGD topological order and its category theoretical description are considered from TGD point of view  category theoretical notions are indeed very natural in TGD framework. The basic finding is that the concepts developed in condensed matter physics (topological order, rough description of states as tangles (graphs imbedded in 3D space), ground state degeneracy, surface states protected by symmetry or topology) fit very nicely to TGD framework and has interpretation in terms of the new spacetime concept. This promises applications also in the conventional areas of condensed matter physics such as more precise description of solid, liquid, and gas phases. See the chapter Criticality and dark matter" or the article Topological order in Quantum TGD . 
Deconstruction and reconstruction in quantum physics and conscious experienceDeconstruction means roughly putting something into pieces. Often deconstruction is thought to involve also the reconstruction. This process is applied in deconstructivistic architecture as one can learn by going to Wikipedia and also cubism brings in mind this kind of approach. Reconstruction organizes typical features of existing styles in new  one might even say "crazy" manner. There can be even a kind of "social interaction" between buildings: as if they were communicating by exchanging features. Similar recombination of elements from various styles have appeared also in music  neoclassicism comes in mind immediately. Postmodernism is a closely related movement and claims that truths are social constructs: great narratives are dead. Nothing could irritate more the physicist who has learned how much mistakes, wrong tracks, and hard work are needed to distill the truth! Everything does not go! On the other hand, one can argue that the recent state of stagnation in the frontier of theoretical physics suggests that postmodernists are right. Superstrings and multiverse are definitely highly social constructs: superstrings were the only game in the town for decades but now American Mathematical Society is worried that super string theoreticians are spoiling the public image of science. Multiverse was in fashion only few years. Certainly one great narrative  the story or reductionism and materialism thought to find its final culmination as Mtheory  is dead. It is however nonsense to claim that all great narratives are dead. That telling alternative great narratives in respected journals is impossible does not mean that they are dead! But the association of deconstruction with postmodernism does not justify throwing away the ideas of deconstruction and reconstruction. Rather, one can ask whether they could be made part of a new great narrative about physical world and consciousness. 1. Deconstruction and reconstruction in perception, condensed matter physics and in TGD inspired theory of consciousness Deconstruction and reconstruction appear in the construction of percepts, in condensed matter physics, and are also part of TGD inspired theory of consciousness. 1.1 Perception The very idea of deconstruction in architectural sense is highly interesting from the perspective of both quantum physics and consciousness. I was astonished as I learned for about 35 years ago that the buildup of our perception involves very concretely what I would now call deconstruction and reconstruction and I could not understand why this. First the sensory input is decomposed into features. Edges, corners, positions, motions analyzed to direction and velocity, colors,... Objects are replaced with collections of attributes: position, motion, shape, surface texture, color,.... Deconstruction occurs at lower cortical layers. After this reconstruction takes place: various kinds of features are combined together through a mysterious looking process of binding  and the outcome is a percept. Reconstruction can occur also in "wrong" manner. This occurs in hallucinations, delusions, and dreams. Humour is based on association of "wrong" things together, making intentional category errors. Synesthesia involves association between different sensory modalities: note with a given pitch has characteristic color or numbers correspond to colors or shapes. I remember an article telling about how subject persons in hypnosis can experience what circle with four corners looks like. Some attribute can be lacking from the reconstruction: person can perceive the car as object but not its motion. The car is there now and a moment later it is here. Nothing between. Also non  standard reconstructions are possible. Could these nonstandard reconstructions define a key aspect of creativity? Could reconstruction create in some lucky situations new idea rather than hallucination or delusion? For few years ago I listened a radio document about a professional, who builds soundscapes to movies and learned that the construction of soundscape is deconstruction followed by reconstruction. One starts from natural sounds but as such they are not very impressive: driving by car over someone does not create any dramatic sound effect  just "splat"  nothing else. This is so nondramatic that it has been be used to create black humour. In order to cure the situation the real sounds are analyzed to features and then reconstructed by amplifying some features and by throwing away the unessential ones. The fictive output sounds much more real than the real input. Actors are masters of this technique and this is why videos about ordinary people doing something funny is like looking autistic ghosts. And if you look at the collection of modules of video game you see modules with name "Aargh", "Auch", "Bangggg", etc.. Association is the neuroscientist's key notion allowing to get an idea about what happens in reconstruction. Reconstruction involves association of various features to form percepts. First this process occurs for various sensory modalities. These intermediate sensory percepts are then combined to full percept in association regions. But what associations are at deeper level? What features are? Heretic could ask whether they could correspond to conscious experiences not conscious to us but conscious at lower subconscious level. Reader perhaps noticed that deconstruction and reconstruction took place here: the student is not supposed to ask this question since most theories of consciousness for some funny reason  maybe a pure accident  make the assumption that consciousness has no structure  no selves with subselves with subselves with... For physicist this kind deconstruction of consciousness is very natural. How do these features bind to our conscious percepts? Neuroscience alone cannot tell much about this since it is based on physicalism: "hard problem" articulates this dead end. The following considerations represent deconstructions and reconstructions, and I will not explicitly mention when this happens  just warning. 1.2 Condensed matter physics One must bring in some basic notions of quantum theory if one wants to reduce de and reconstruction to quantum physics. The key mathematical fact is that in quantum theory each particle in manyparticle state corresponds to a tensor factor of state space of the entire system. This notion is very difficult to explain without actually having a lecture series about quantum theory and I prove in the following that this is indeed the case.
Now comes the surprise: condensed matter physicists have discovered deconstruction long time ago)! Condensed matter electron can be deconstructed under some circumstances.
It is however true that 1dimensional systems are very special physically. Braid statistics replaces ordinary statistics bringing in a lot of new effects. Furthermore, 2D integrable gauge theories allow to model interactions as permutations of quantum numbers and lead to elegant models describing deconstructed degrees of fields as quantum fields in 2D Minkowski space with interactions reducing to 2particle interactions decribable in terms of Rmatrix satisfying the YangBaxter equations. It is difficult to say how much the association of deconstruction to 1D systems is due the fact that they are mathematically easier to handle than higherD ones and there is existing machinery. The rise of superstring models certainly was to a high degree due to this technical easiness. As I tried to tell about 3surfaces replacing strings as fundamental dynamical objects, the almost reflect like debunking of this idea was to say that superconformal invariance of superstring models is lost and the theory is not calculable and does not even exist  period. It took indeed a long time to realize that superconformal symmetry allows a huge generalization, when spacetime is 4D and imbedding space has Minkowski space as its Cartesian factor. Twistorial considerations fixed the imbedding space uniquely to M^{4}× CP_{2}. The lesson is clear: theoretician should be patient and realize that theory building is much more than going to math library and digging the needed mathematics. Maybe colleagues are mature to learn this lesson some day. 1.3 TGD inspired theory of consciousness The believer in quantum consciousness of course wonders what could be the quantum counterparts of de and reconstruction as mechanism of perception. It would seem that analysis and synthesis of the sensory input deconstructs the mental image associated with it to features  perhaps simpler fundamental mental images  and reconstruct from these the percept as mental image. What does this correspond at the level of physics? Before one can really answer one must understand what the quantum physical correlates of mental image are. How mental images die and are born? What features are as mental images? What their binding to sensory percepts does mean physically? Here I can answer only on my own behalf and to do it I must introduce the basic notions and ideas of TGD inspired theory of consciousness. I will not go into details here because I have done this so many times and just suggest that the reading of some basic stuff about TGD inspired theory of consciousness. Suffice it to list just the basic ideas and notions.
What about reconstruction in this framework?
2. Could condensed matter physics and consciousness theory have something to share? Magnetic bodies are present in all scales and one can ask whether consciousness theory and condensed matter physics might have something in common. Could the proposed picture of matter as consisting of selves with subselves with.... defining analogs of quasiparticles and collective excitations make sense even at the level of condensed matter? Could construction and reconstruction of mental images identifiable as subselves take place already at this level and have interpretation in terms of primitive information processing building standardized primitive mental images? Deconstruction need not be restricted to electron and velocity could be replaced by oscillation frequency for various fields: at quantum level there is not actually real distinction since in quantum theory velocity defines wave vector. Also more complex objects, atoms, molecules, etc. could be deconstructed and the process could occur at the level of magnetic bodies and involve in essential manner reconnection and other "motor actions" of flux tubes. The notions of quasiparticle and collective excitation would generalized dramatically and the general vision about basic mechanism might help to understand this zoo of exotics. Future condensed matter theorists might also consider the possibility of reconstruction in new manner giving rise to the analogs of synesthesia. Could features from different objects be recombined to form exotic quasiobjects having parts all around. Could dark matter in TGD sense be involved in an essential manner? Could cyclotron resonance or its absence serve as a correlate for the binding? Note that the disjoint regions of space would be in welldefined sense near to each other in the reconstructed state. Topology would be different: effective padic topology could provide a natural description for a situation: in padic topology systems at infinite distance in real sense can be infinitesimally close to each other padically. See the chapter Criticality and dark matter" or the article Deconstruction and reconstruction in quantum physics and conscious experience 
A new control mechanism of TGD inspired quantum biologyThe idea that TGD Universe is quantum critical, is the key idea of quantum TGD and fixes the theory more or less uniquely since the only coupling constant parameter of the theory  Kähler coupling strength  is analogous to critical temperature. Also more than one basic parameters are in principle possible  maximal quantum criticality fixes the values of all of them  but it seems that only Kähler coupling strength is needed. TGD Universe is a quantum critical fractal: like a ball at the top of hill at the top of hill at.... Quantum criticality allows to avoid the fine tuning problems plaguing as a rule various unified theories. Quantum criticality The meaning of quantum criticality at the level of dynamics has become only gradually clearer. The development of several apparently independent ideas generated for about decade ago have led to the realization that quantum criticality is behind all of them. Behind quantum criticality are in turn number theoretic vision and strong forms of general coordinate invariance and holography.
Quantum criticality and TGD inspired quantum biology In TGD inspired quantum biology quantum criticality is in crucial role. First some background.
A new mechanism of quantum criticality Consider now the mechanisms of quantum criticality. The TGD based model (see this) for the recent paradoxical looking finding (see this) that topological insulators can behave like conductors in external magnetic field led to a discovery of a highly interesting mechanism of criticality, which could play a key role in living matter.
A new mechanism of quantum criticality and biocontrol The quantum criticality of the process in which new electron orbit emerges near Fermi surface suggests a new mechanism of quantum biocontrol by generation of super currents or its reversal.
See the chapter Criticality and Dark Matter or the article A new control mechanism of TGD inspired quantum biology . 
Does the physics of SmB_{6} make the fundamental dynamics of TGD directly visible?The group of Suchitra Sebastian has discovered very unconventional condensed matter system which seems to be simultaneously both insulator and conductor of electricity in presence of magnetic field. Science article is entitled "Unconventional Fermi surface in an insulating state". There is also a popular article "Paradoxical Crystal Baffles Physicists" in Quanta Magazine summarizing the findings. I learned about the finding first from the blog posting of Lubos (I want to make absolutely clear that I do not share the racistic attitudes of Lubos towards Greeks. I find the discussions between Lubos and same minded blog visitor barbarians about the situation in Greece disgusting). Observations The crystal studied at superlow temperatures was Samarium hexaboride  briefly SmB_{6}. The high resistance implies that electron cannot move more that one atom's width in any direction. Sebastian et al however observed electrons traversing over a distance of millions of atoms a distance of orde 10^{4} m, the size of a large neuron. So high mobility is expected only in conductors. SmB_{6} is neither metal or insulator or is both of them! The finding is described by Sebastian as a "big schock and as a "magnificent paradox by condensed matter theorists Jan Zaanen. Theoreticians have started to make guesses about what might be involved but according to Zaanen there is no even remotely credible hypothesis has appeared yet. On basis of its electronic structure SmB_{6} should be a conductor of electricity and it indeed is at room temperature: the average number conduction electrons per SmB_{6} is one half. At low temperatures situation however changes. At low temperatures electrons behave collectivly. In superconductors resistance drops to zero as a consequence. In SmB_{6} just the opposite happens. Each Sm nucleus has the average 5.5 electrons bound to it at tight orbits. Below 223 degrees of Celsius the conduction electrons of SmB_{6} are thought to "hybridize" around samarium nuclei so that the system becomes an insulator. Various signatures demonstrate that SmB_{6} indeed behaves like an insulator. During last five years it has been learned that SmB_{6} is not only an insulator but also so called topological insulator. The interior of SmB_{6} is insulator but the surface acts as a conductor. In their experiments Sebastian et al hoped to find additional evidence for the topological insulator property and attempted to measure quantum oscillations in the electrical resistance of their crystal sample. The variation of quantum oscillations as sample is rotated can be used to map out the Fermi surface of the crystal. No quantum oscillations were seen. The next step was to add magnetic field and just see whether something interesting happens and could save the project. Suddenly the expected signal was there! It was possible to detect quantum oscillations deep in the interior of the sample and map the Fermi surface! The electrons in the interior travelled 1 million times faster than the electrical resistance would suggest. Fermi surface was like that in copper, silver or gold. A further surprise was that the growth of the amplitude of quantum oscillations as temperature was decreased, was very different from the predictions of the universal LifshitzKosevich formula for the conventional metals. Could TGD help to understand the strange behavior of SmB_{6}? There are several indications that the paradoxical effect might reveal the underlying dynamics of quantum TGD. The mechanism of conduction must represent new physics and magnetic field must play a key role by making conductivity possible by somehow providing the "current wires". How? The TGD based answer is completely obvious: magnetic flux tubes. One should also understand topological insulator property at deeper level, that is the conduction along the boundaries of topological insulator. One should understand why the current runs along 2D surfaces. In fact, many exotic condensed matter systems are 2dimensional in good approximation. In the models of integer and fractional quantum Hall effect electrons form a 2D system with braid statistics possible only in 2D system. High temperature superconductivity is also an effectively 2D phenomenon.One should also understand topological insulator property at deeper level, that is the conduction along the boundaries of topological insulator.
Quantum criticality is the crucial aspect and corresponds to the situation in which the magnetic field attains a value for which a new orbit emerges/disappears at the surface of the flux tube: in this situation dark electron phase with nonstandard value of h_{eff} can be generated. This mechanism is expected to apply also in bio superconductivity and to provide a general control tool for magnetic body.
What about quantum oscillations in TGD framework?
See the chapter Criticality and Dark Matter or the article Does the physics of SmB_{6} make the fundamental dynamics of TGD directly visible? 
Discretization and quantum group description as different aspects of finite measurement resolutionIn Thinking Allowed Original there was a link to a very interesting article with title "Designing Curved Blocks of Quantum SpaceTime...Or how to build quantum geometry from curved tetrahedra in loop quantum gravity" telling about the work of Etera Livine working at LPENSL (I let the reader to learn what this means;). The idea of the article The popular article mentions a highly interesting mathematical result relevant for TGD. The idea is to build 3geometry  not by putting together flat tetrahedra or more general polyhedra along their boundaries  but by using curved hyperbolic tetrahedra or more generally polygons) defined in 3D hyperbolic space  negative constant curvature space with Lorentz group acting as isometries  cosmic time=constant section of standard cosmology. As a special case one obtains tesselation of 3D hyperbolic space H^{3}. This is somewhat trivial outcome so that one performs a "twisting". Some words about tesselations/lattices/crystals are in order first.
Back to the article and its message! The condition for tetrahedron property stating in flat case that the sum of the 4 normal vectors vanishes generalizes, and is formulated in group SU(2) rather than in group E^{3} (Euclidian 3space). The popular article states that deformation of sum to product of SU(2) elements is equivalent with a condition defining classical qdeformation for gauge group. If this is true, a connection between "quantum quantum mechanics" and hyperbolic geometries therefore might exist and would correspond to a transition from flat E^{3} to hyperbolic H^{3}. Let loop gravity skeptic talk first This looks amazing but it is better to remain skeptic since the work relates to loop quantum gravity and involves specific assumptions and different motivations.
The notion of finite measurement resolution The notion of finite measurement resolution emerged first in TGD through the realization that von Neumann algebras known as hyperfinite factors of type I_{1} (perhaps also of type III_{1}) emerge naturally in TGD framework. The spinors of "world of classical worlds" (WCW) identifiable in terms of fermionic Fock space provide a canonical realization for them. The inclusions of hyperfinite factors provide a natural description of finite measurement resolution with included factor defining the subalgebra, whose action generates states not distinguishable from the original ones. The inclusions are labelled by quantum phases coming as roots of unity and labelling also quantum groups. Hence the idea that quantum groups could allow to describe the quantal aspects of finite measurement resolution whereas discretization would define its classical aspects. pAdic sectors of TGD define a correlate for cognition in TGD Universe and cognitive resolution is forced by number theory. Indeed, one cannot define the notion of angle in padic context but one can define phases in terms of algebraic extensions of padic numbers defined by roots of unity: hence a finite cognitive resolution is unavoidable and might have a correlate also at the level of real physics. The discrete algebraic extensions of rationals forming a cognitive and evolutionary hierarchy induce extensions of padic numbers appearing in corresponding adeles and for them quantum groups should be a necessary ingredient of description. The following arguments support this view and make it more concrete. Quantum groups and discretization as two manners to describe finite measurement resolution in TGD framework What about quantum groups in TGD framework? I have also proposed that qdeformations could represent finite measurement resolution. There might be a connection between discretizing and quantum groups as different aspects of finite measurement resolution. For instance, quantum group SU(2)_{q} allows only a finite number of representations (maximum value for angular momentum): this conforms with finite angular resolution implying a discretization in angle variable. At the level of padic number fields the discretization of phases exp(iφ) as roots U_{n}=exp(i2π/n) of unity is unavoidable for number theoretical reasons and makes possible discrete Fourier analysis for algebraic extension. There exist actually a much stronger hint that discretization and quantum groups related to each other. This hint leads actually to a concrete proposal how discretization is described in terms of quantum group concept.

Two kinds of negentropic entanglementsThe most general view is that negentropic entanglement NE corresponds to algebraic entanglement with entanglement coefficients in some algebraic extension of rationals. The condition that the outcome of state function reduction is eigenspace of density matrix fixes the density matrix of the final state to be a projector with identical eigenvalues defining the probabilities of various states. But what if the eigenvalues and thus also eigenvectors of the density matrix, which are algebraic numbers, do not belong to the algebraic extensions involved. Can state function reduction reduction occur at all so that this kind of NE would be stable? The following argument suggests that also more general algebraic entanglement could be reasonably stable against NMP, namely the entanglement for which the eigenvalues of the density matrix and eigenvectors are outside the algebraic extension associated with the parameters characterizing string world sheets and partonic 2surfaces as spacetime genes. The restriction to a particular extension of rationals  a central piece of the number theoretical vision about quantum TGD  implies that density matrix need not allow diagonalization. In eigen state basis one would have has algebraic extension defined by the characteristic polynomial of the density matrix and its roots define the needed extension which could be quite well larger than the original extension. This would make state stable against state function reduction. If this entanglement is algebraic, one can assign to it a negative number theoretic entropy. This negentropic entanglement is stable against NMP unless the algebraic extension associated with the parameters characterizing the parameters of string world sheets and partonic surfaces defining spacetime genes is allowed to become larger in a state function reduction to the opposite boundary of CD generating reincarnated self and producing eigenstates involving algebraic numbers in a larger algebraic extension of rationals. Could this kind of extension be an eureka experience meaning a step forwards in cognitive evolution? If this picture makes sense, one would have both the unitary NE with a density matrix, which is projector and the algebraic NE with eigen values and NE for which the eigenstates of density matrix outside the algebraic extension associated with the spacetime genes. Note that the unitary entanglement is "meditative" in the sense that any state basis is possible and therefore in this state of consciousness it is not possible to make distinctions. This strongly brings in mind koans of Zen buddhism and enlightment experience. The more general irreducible algebraic entanglement could represent abstractions as rules in which the state pairs in the superposition represent the various instances of the rule. For details see the chapter Negentropy Maximization Principle or the article Impressions created by TSC2015 conference. 
Quantitative model of high T_{c} superconductivity and biosuperconductivityI have developed already earlier a rough model for high T_{c} super conductivity. The members of Cooper pairs are assigned with parallel flux tubes carrying fluxes which have either same or opposite directions. The essential element of the model is hierarchy of Planck constants defining a hierarchy of dark matters.
For details see the chapter SuperConductivity in ManySheeted SpaceTime or the article Quantitative model of high T_{c} superconductivity and biosuperconductivity. 
Updated Negentropy Maximization PrincipleQuantum TGD involves "holy trinity" of time developments. There is the geometric time development dictated by the preferred extremal of Kähler action crucial for the realization of General Coordinate Invariance and analogous to Bohr orbit. There is what I originally called unitary "time development" U: Ψ_{i}→ UΨ_{i}→ Ψ_{f}, associated with each quantum jump. This would be the counterpart of the Schrödinger time evolution U(t,t→ ∞). Quantum jump sequence itself defines what might be called subjective time development. Concerning U, there is certainly no actual Schrödinger equation involved: situation is in practice same also in quantum field theories. It is now clear that in Zero Energy Ontology (ZEO) U can be actually identified as a sequence of basic steps such that single step involves a unitary evolution inducing delocalization in the moduli space of causal diamonds CDs) followed by a localization in this moduli space selecting from a superposition of CDs single CD. This sequence replaces a sequence of repeated state function reductions leaving state invariant in ordinary QM. Now it leaves in variant second boundary of CD (to be called passive boundary) and also the parts of zero energy states at this boundary. There is now a very attractive vision about the construction of transition amplitudes for a given CD, and it remains to be see whether it allows an extension so that also transitions involving change of the CD moduli characterizing the nonfixed boundary of CD. A dynamical principle governing subjective time evolution should exist and explain state function reduction with the characteristic oneone correlation between macroscopic measurement variables and quantum degrees of freedom and state preparation process. Negentropy Maximization Principle is the candidate for this principle. In its recent form it brings in only a single little but overall important modification: state function reductions occurs also now to an eigenspace of projector but the projector can now have dimension which is larger than one. Self has free will to choose beides the maximal possible dimension for this subspace also lower dimension so that one can speak of weak form of NMP so that negentropy gain can be also below the maximal possible: we do not live in the best possible world. Second important ingredient is the notion of negentropic entanglement relying on padic norm. The evolution of ideas related to NMP has been slow and tortuous process characterized by misinterpretations, overgeneralizations, and unnecessarily strong assumptions, and has been basically evolution of ideas related to the anatomy of quantum jump and of quantum TGD itself. Quantum measurement theory is generalized to theory of consciousness in TGD framework by replacing the notion of observer as outsider of the physical world with the notion of self. Hence it is not surprising that several new key notions are involved.

Individual nucleons inside nuclei do not behave according to predictionsQuantum TGD involves "holy trinity" of time developments. There is the geometric time development dictated by the preferred extremal of Kähler action crucial for the realization of General Coordinate Invariance and analogous to Bohr orbit. There is what I originally called unitary "time development" U: Ψ_{i}→ UΨ_{i}→ Ψ_{f}, associated with each quantum jump. This would be the counterpart of the Schrödinger time evolution U(t,t→ ∞). Quantum jump sequence itself defines what might be called subjective time development. Concerning U, there is certainly no actual Schrödinger equation involved: situation is in practice same also in quantum field theories. It is now clear that in Zero Energy Ontology (ZEO) U can be actually identified as a sequence of basic steps such that single step involves a unitary evolution inducing delocalization in the moduli space of causal diamonds CDs) followed by a localization in this moduli space selecting from a superposition of CDs single CD. This sequence replaces a sequence of repeated state function reductions leaving state invariant in ordinary QM. Now it leaves in variant second boundary of CD (to be called passive boundary) and also the parts of zero energy states at this boundary. There is now a very attractive vision about the construction of transition amplitudes for a given CD, and it remains to be see whether it allows an extension so that also transitions involving change of the CD moduli characterizing the nonfixed boundary of CD. A dynamical principle governing subjective time evolution should exist and explain state function reduction with the characteristic oneone correlation between macroscopic measurement variables and quantum degrees of freedom and state preparation process. Negentropy Maximization Principle is the candidate for this principle. In its recent form it brings in only a single little but overall important modification: state function reductions occurs also now to an eigenspace of projector but the projector can now have dimension which is larger than one. Self has free will to choose beides the maximal possible dimension for this subspace also lower dimension so that one can speak of weak form of NMP so that negentropy gain can be also below the maximal possible: we do not live in the best possible world. Second important ingredient is the notion of negentropic entanglement relying on padic norm. The evolution of ideas related to NMP has been slow and tortuous process characterized by misinterpretations, overgeneralizations, and unnecessarily strong assumptions, and has been basically evolution of ideas related to the anatomy of quantum jump and of quantum TGD itself. Quantum measurement theory is generalized to theory of consciousness in TGD framework by replacing the notion of observer as outsider of the physical world with the notion of self. Hence it is not surprising that several new key notions are involved.

Individual nucleons inside nuclei do not behave according to predictionsIndividual nucleons do not behave in nuclei as the existing theory predicts (see the popular article). This is a conclusion reached by an international team of scientists which has published their findings as article article in Phys. Rev. Letters). I am not a nuclear physicists but have proposed what I call nuclear string model. Despite this I to try to understand what has been found and what nuclear string model can say about the findings. Background and results There are many models of atomic nuclei and each of them explains some aspects of nucleus. Nucleus can be modelled rigid body or as a kind of quantum liquid. In the prevailing average field approach the presence of other nucleons is described in terms of a potential function and calculates the states of individual nucleons in this potential using Schrödinger equation. It is essential that nucleons are assumed to be independent. The model taking potential function to be that of harmonic oscillator is surprisingly successful but one must introduce corrections such as spinorbit coupling in order to understand the energy spectrum. In this approach the notion of nuclear shell emerges. In atomic physics and chemistry the closed shells do not participate to the interaction and the outermost shell characterized by valence dictates to a higher degree the chemical properties of atom. Valence is positive if outer shell contains particles. Valence if negative if some of them are lacking. Something similar is to be expected also now. In this case full shells correspond to magic numbers for protons and neutrons separately (note that protons and neutrons seem to behave rather independently, something highly nontrivial!). The nuclei with valence +1 or 1 would correspond to almost magic nuclei. One generally accepted correction to the harmonic oscillator model is inspired by the assumption that heavier nuclei can be described as a kind of blob of condensed matter obeying equation of state allowing to introduce notions like acoustic waves and surface waves. The nucleon at the unfilled shell would reside at the surface of this blob. The blob has vibrational excitations characterized by multipolarity (spherical harmonic characterized by angular momentum quantum numbers and the radial part of the oscillation amplitude. These excitations give rise to analogs of surface waves in water. Valence nucleons interact with the oscillations and affect the energy levels of the valence nucleons. The predictions of this model are calculable. The team has studied almost doubly magic nuclei with valence equal to plus or 1 and calculated the effects on the energy levels of the nucleon and found that the observed effects are signficantly smaller than the predicted ones. This finding challenges both the mean field approach or the idea that nucleus can be idealized as a condensed matter like system or both. Nuclear string model In TGD framework ordinary model of nucleus is replaced with what I call nuclear string model. The basic ideas of the nuclear string model are following.
How does nuclear string model relate to the shell model? In the mean field approximation particles move independently in a potential describing the effects of the other nucleons. The basis for Nnucleon wave functions can be constructed as products of those associated with individual nucleons. Under what conditions nuclear string model is consistent with independent particle approach?
Semiclassical considerations One can consider the situation also semiclassically.
Could magic numbers be understood in terms of Platonic solids? Harmonic oscillator model predicts the numbers of nucleons for magic nuclei as sums of numbers of nucleons for the full shells involved but the predictions are not quite correct. One can however modify the model to get the observed magic numbers. Could these numbers be consistent with the idea that a full shell corresponds to a Platonic solid such that closed nuclear string, which can connect only neighboring vertices goes through its vertices without intersecting itself?
Could one understand the numbers n in terms of Platonic solids?
These findings would suggest that the independent particle model is not a good approximation for light nuclei for which a model as a molecule like entity with rather rigid position of nucleons can be considered if Platonic solids are taken as metric objects. The experimental findings from TGD point of view? On basis of the experimental findings it is far from clear whether one can model nuclei as objects characterized by continuous nucleon densities and obeying some thermodynamical equation of state from which the dynamics describing the oscillations of nucleon densities can be deduced.
See chapter Nuclear string model or the article Individual nucleons inside nuclei do not behave according to predictions. 
Quantization of conductance in neutral matter as evidence for manysheeted spacetime?We are living really interesting times. Experimental sciences are producing with accelerating pace new discoveries challenging the existing theories and it is difficult to avoid the impression that a revolution is going on in physics and also in biology and neuroscience. It is a pity that colleagues do not seem to even realize what is going on. Ulla's Facebook page Quantum Biology, coherence and decoherence contained this morning a link to and article published in Nature.
The article tells about quantized conductance in neutral matter. In quantum Hall effect conductances is quantized in multiples of e^{2}/h. Now the however is in multiples of 1/h. Looks strange! This is due to the fact that voltage is not present now: particles are neutral and electric field is replaced with the gradient of chemical potential and electric current with particle current. Hence elementary charge e is replaced with the unit for particle number which is just 1 rather than e. Hence the quantisation as multiples of 1/h but in complete analogy with Quantum Hall Effect (QHE). What comes to my innocent in mind is that the effect is mathematically like QHE and that there is also fractional variant of it as in the case of QHE. In QHE magnetic field and cyclotron states at flux quanta of this field are in key role. But in the situation considered they are not present if we live in the standard model world. What is the situation in TGD?
If this approach is on the correct track then the thermodynamical description in terms of chemical potential cannot be fundamental (the gradient of the chemical potential replaces that of electric potential in this description). Leaving the realm of standard model, one could however wonder whether the thermodynamical description using chemical potentials (chemistry is by definition effective theory!) is really fundamental in quantum regime and whether it could reduce to something more fundamental which standard model can describe only phenomenologically.
The most obvious objection that the quantum of conductivity for neutral particles is 1/h rather than g^{2}/h, where g is appropriate weak coupling strength does not bite. Experimentalists measure particle currents rather than Z^{0} currents (j= j_{Z}/g_{Z}) and use gradient of chemical potential instead of Z^{0} potentials μ= g_{Z}E_{Z}). j_{Z}= σ E_{Z} implies that the quantization of the conductance is in multiples of 1/h. For details and references see the chapter Quantum Hall effect and hierarchy of Planck constants or the article Quantization of conductance in neutral matter as evidence for manysheeted spacetime?. 