• JimKassonpanorama_fish_eye
    1738 posts
    2 years ago

    It’s 1.3 um on the object side. So 13 um on the image side, whatever the metric is

    The lens I have is 1500 lp/mm on the object side, for a line width of around 400 nm.

  • DonaldBpanorama_fish_eye
    2378 posts
    2 years ago

    i dont know the maths , i can only gauge my lens performance on bug bits 🤨

    so is that why they say 5 um to 11um for best results ?

  • DonaldBpanorama_fish_eye
    2378 posts
    2 years ago

    what does this mean to a layman 🤔

  • DonaldBpanorama_fish_eye
    2378 posts
    2 years ago

    but isnt that for a projection of 100x100mm image arnt you going to squash the projection for your MF sensor

  • JACShelp_outline
    878 posts
    2 years ago

    [deleted]

  • JimKassonpanorama_fish_eye
    1738 posts
    2 years ago

    At the recommended 436 nm spectral line and magnification ratio, the lens is supposed to be diffraction limited. But note I’m talking about using the lens reversed, so the sensor side sees only 150 lp/mm.

    The lens was designed for stepping semiconductor lithography.

  • JimKassonpanorama_fish_eye
    1738 posts
    2 years ago

    My sensor is well under 100x100 mm.

  • DonaldBpanorama_fish_eye
    2378 posts
    2 years ago

    yes thats right but your lens was designed to project a transparency of 100x100mm onto a micro chip !
    but arnt you only projecting the image onto mf sensor in reverse

  • DonaldBpanorama_fish_eye
    2378 posts
    2 years ago

    if this is the case from back in the 80s 40 years ago then how are we making circuits so small today ?

  • DonaldBpanorama_fish_eye
    2378 posts
    2 years ago

    did some googling. very interesting. much more advanced to when i used to make my own PCBs when i was into electronics 45 years ago.

  • IliahBorgpanorama_fish_eye
    976 posts
    2 years ago

    That, given you have 10x/0.25 objective, for green 550 μm wavelength you need at least a 5.5 μm pixel, and depending on other factors, smaller pixels may be better. Look up the Rayleigh criterion.

  • JimKassonpanorama_fish_eye
    1738 posts
    2 years ago

    The lens was designed for 1:10. In that usage the wafer side image circle diameter was 14mm and the mask side image circle was 140mm. So as a 10x lens on a 33x44 mm sensor, the object side field of view will be 3.3x4.4mm.

  • JimKassonpanorama_fish_eye
    1738 posts
    2 years ago

    Shorter wavelengths, some tricky optical schemes.

    www.asml.com/en/products/euv-lithography-systems

  • DonaldBpanorama_fish_eye
    2378 posts
    2 years ago

    truly amazing

  • Ghundredpanorama_fish_eye
    758 posts
    2 years ago

    For sure. To get a bit of an idea of scale, if you enlarged the diameter of a human hair of 0.1mm up to the height of the Empire state building, one nm would be less than 6mm high on that scale. And we take for granted our computer chips running 3-5nm circuitry. For want of a better word. Truly mind boggling.

  • DonaldBpanorama_fish_eye
    2378 posts
    2 years ago

    did you read where the computers that control the lasers have deep learning AI built in and adjust the control accuracy at 50,000 times per sec , we think the sony a1 is amazing at 1/400 sec 🤔 i suppose you would expect that with 6 billion dollars spent on development 🤨

  • JimKassonpanorama_fish_eye
    1738 posts
    2 years ago

    What I read is that the lasers that are aimed at the tin droplets have a pulse repetition rate of 50kHz.

    The wafer positioning control system operates at 20kHz, with 250 picometer repeatability.

    Note that their first systems operated at 436 nm, which is the wavelength for which my lens was designed.

  • JimKassonpanorama_fish_eye
    1738 posts
    2 years ago

    I didn’t see in write up, but you can’t use refractive lenses at that wavelength. It’s all done with mirrors. The optics come from Zeiss, with defect correction at the molecular level. The optical path needs to be in a vacuum, since air absorbs EUV.

    Even the mask making is exotic.