On Saturday, January 12, 2019 at 7:09:29 PM UTC-6, Lawrence Crowell wrote:
>
> On Saturday, January 12, 2019 at 4:17:56 PM UTC-6, Brent wrote:
>>
>>
>>
>> On 1/12/2019 2:51 AM, Philip Thrift wrote:
>>
>>
>>
>> On Friday, January 11, 2019 at 7:19:06 PM UTC-6, Brent wrote: 
>>>
>>>
>>>
>>> On 1/11/2019 1:57 PM, Philip Thrift wrote:
>>>
>>>
>>>
>>> On Friday, January 11, 2019 at 2:46:35 PM UTC-6, Brent wrote: 
>>>>
>>>>
>>>>
>>>> On 1/11/2019 6:01 AM, John Clark wrote:
>>>>
>>>> On Thu, Jan 10, 2019 at 8:18 PM Brent Meeker <[email protected]> 
>>>> wrote:
>>>>
>>>> * > The fine structure constant is e^2/hbar*c.  Those three values are 
>>>>> measured independent of any Feynman diagrams*
>>>>>
>>>>
>>>> Absolutely correct. So if you use Feynman diagrams to predict what some 
>>>> physical system is going to do, such as a physical system of 2 electrons 
>>>> being hit by a photon of light with a wavelength small enough to contain 
>>>> enough energy to prevent the electrons repulsion, then you'd better get a 
>>>> number very close to the Fine Structure Constant. If you don't then 
>>>> Feynman 
>>>> Diagrams aren't any good. 
>>>>
>>>> They didn't use 12,672 Feynman Diagrams because they wanted to know 
>>>> what the Fine Structure Constant was, they already knew what that 
>>>> number was to many decimal places from exparament, they used 12,672 
>>>> Feynman Diagrams because they wanted to see if Feynman Diagrams 
>>>> worked. And it turned out they worked spectacularly well in that 
>>>> situation, 
>>>> and that gives scientists great confidence they can use Feynman Diagrams 
>>>> in 
>>>> other situations to calculate what other physical systems will do that 
>>>> involve the Electromagnetic Force.
>>>>
>>>>
>>>> There's always an interplay between theory and experiment.  It's 
>>>> completely analogous to Maxwell's discovery that light is EM waves. There 
>>>> were already experimental values of the permittivity and permeability of 
>>>> the vacuum and there were values for the speed of light.  Maxwell showed 
>>>> that his theory of EM predicted waves and using the permittivity and 
>>>> permeability values the speed of the waves matched that of light.  Now the 
>>>> speed of light is a defined constant and so are the permittivity and 
>>>> permeability of the vacuum.  So the connecting of the three values by a 
>>>> theory allows their values to be defined.  In the case of the anomalous 
>>>> magnetic moment of the electron, hbar and c are already defined constants. 
>>>>  
>>>> So quantum field theory (for which Feynman diagrams are just a 
>>>> calculational tool) linked them and e to g.
>>>>
>>>> Brent
>>>>
>>>>
>>>
>>>
>>> If Feynman Diagrams (tools) are sufficient (to match experimental data) 
>>> then Quantum Field Theory can be thrown in the wastebasket.
>>>
>>>
>>> ?? Feynman Diagrams are just a mathematical trick for summing up terms 
>>> to approximate the propagator of QFT.  
>>>
>>> Brent
>>>
>>
>>
>> You just make Feynman Diagrams the fundamental elements of the theory, 
>> and propagators derived from them.
>>
>>
>> How many diagrams?  The propagator has a clear interpretation as 
>> connecting the field at x with the field at y.  Feynman showed that his 
>> diagrams provided a good mnemonic for the infinite number of terms that 
>> would sum to the propagator.  If you take the diagrams as fundamental you 
>> then need to specify how many.
>>
>>
>> Just like histories are made fundamental, and Hilbert Spaces are derived 
>> from them.
>>
>>
>> Hilbert spaces are infinite dimensional vector spaces.  So you have the 
>> same problem: How many histories?
>>
>> Brent
>>
>  
> The number of diagrams grows exponentially. As I recall the QED industry 
> is up to 12 orders of radiative corrections and renormalization orders. The 
> number of diagrams to evaluate and sum is in the millions if not billions. 
> This stuff is done on supercomputers these days. People do not really 
> evaluate Feynman diagrams, they write computer programs.
>
> LC
>


Supercomputers are the future of theoretical physics it seems, like the one 
at LSU, SuperMike-II.

http://www.hpc.lsu.edu/docs/guides.php?system=SuperMike2

*SuperMike-II is a 146 TFlops Peak Performance 440 compute node cluster 
running the Red Hat Enterprise Linux 6 operating system. Each node contains 
two 8-Core Sandy Bridge Xeon 64-bit processors operating at a core 
frequency of 2.6 GHz. Fifty of the compute nodes also have two NVIDIA M2090 
GPUs that provide an additional 66 Tflops total Peak performance.*

use in LQG:
https://www.lsu.edu/mediacenter/news/2018/12/20physastro_singh_prl.php 


- pt

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to