You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I was wondering if the far detuned dipole beams of a FORT / ODT lead to an A.C. Stark shift of the optical transitions used for a MOT. For the example of a 1064nm dipole trap that is briefly superimposed with a red Sr-MOT, I tried to calculate the light shift according to 7.7 Foot (see PDF attached). As you can see, the result I get is pretty much zero but I'm confused by the sentence "Normally light shifts are most important at large frequency detuning [...]" in 7.7 Foot.
What do you think - is my rough estimation of the magnitude of the light shift correct in principle or have I missed something crucial in my approach? Does someone have practical experience with MOT-->ODT transitions and knows if that effect becomes relevant in real life?
Why I am needing this: I want to simulate MOT --> ODT transitions with AtomECS and I want to know if an a.c. Stark shift is something we'd have to implement for this to work correctly. Stark-shift_ODT.pdf
The text was updated successfully, but these errors were encountered:
I'm on holiday so I won't add more detail now - but I'll comment that 'shift' could refer to either a transition or the energy levels themselves. Large frequency detuning = dipole trap = large AC stark shift used to confine atoms. May or not produce a large shift in transition frequencies, depending on the shift of the ground and excited states.
Thanks, important remark, I am aware of that... that's why I calculate \Delta \omega... but we can discuss that another time. So far, I wish you nice holidays. :)
My current plan to implement this is using a CoolingTransitionStarkShift component, which will hold the susceptibility of the atom's cooling transition frequency to DipoleTrap light. This can then be accounted for when we add in doppler + zeeman shifts.
I was wondering if the far detuned dipole beams of a FORT / ODT lead to an A.C. Stark shift of the optical transitions used for a MOT. For the example of a 1064nm dipole trap that is briefly superimposed with a red Sr-MOT, I tried to calculate the light shift according to 7.7 Foot (see PDF attached). As you can see, the result I get is pretty much zero but I'm confused by the sentence "Normally light shifts are most important at large frequency detuning [...]" in 7.7 Foot.
What do you think - is my rough estimation of the magnitude of the light shift correct in principle or have I missed something crucial in my approach? Does someone have practical experience with MOT-->ODT transitions and knows if that effect becomes relevant in real life?
Why I am needing this: I want to simulate MOT --> ODT transitions with AtomECS and I want to know if an a.c. Stark shift is something we'd have to implement for this to work correctly.
Stark-shift_ODT.pdf
The text was updated successfully, but these errors were encountered: