We describe a model evaluating changes in the optical isolation of a Faraday isolator when passing from air to vacuum in terms of different thermal effects in the crystal. The changes are particularly significant in the crystal thermal lensing (refraction index and thermal expansion) and in its Verdet constant and can be ascribed to the less efficient convection cooling of the magneto-optic crystal of the Faraday isolator. An isolation decrease by a factor of 10 is experimentally observed in a Faraday isolator that is used in a gravitational wave experiment (Virgo) with a 10W input laser when going from air to vacuum. A finite element model simulation reproduces with a great accuracy the experimental data measured on Virgo and on a test bench. A first set of measurements of the thermal lensing has been used to characterize the losses of the crystal, which depend on the sample. The isolation factor measured on Virgo confirms the simulation model and the absorption losses of 0:0016 ± 0:0002=cm for the TGG magneto-optic crystal used in the Faraday isolator.