Excuse the cheekiness, but you "know your onions" - could you recommend undervoltage settings for a 1080 and a 1080ti (different computers)
Every sample is a bit different, but the same basic procedure applies to any of the last several generations of NVIDIA hardware.
Get MSI AB, and instead of using any clock or voltage offsets, open up the frequency/voltage curve editor with CTRL+F. Pick a clock speed and voltage, then use shift+left click and drag to highight the entire curve after the voltage you've decided on as your limit, then click on of the points on the highlighted section of the graph and drag that entire section to some point below the clock speed of your chose voltage and 'apply'. Now you just adjust the points at your chosen voltage and lower.
For the 1080 Ti, 900mV @ 1900MHz is a good starting point (~90% of GP102 parts are sable here). As long as the highest point on the curve is 1900MHz, and that point is at 900mV, the card will not exceed that voltage or clock speed. Most 1080 Tis will be stable or close to stable here, but you'll need to do some stress testing to be sure (3DMark TimeSpy tends to crash out sooner than most tests and it's worth the five dollars so that you can loop it), and may have a bit more room to increase clocks, or reduce voltage...or you may need to pull back on clocks or allow the same 1900MHz to take a bit more voltage.
Do note that the curve offsets are heavily temperature dependent on these parts. If you adjust the curve while the card is cold, you will see lower clocks under load. Ideally, you tune the curve near peak temperature, with windowed 3D load running in the background to keep temps up and provide immediate feedback for performance/stability changes (I recommend Unigine Valley, for convience and ability to run at full speed when not in focus, it's not a particularly great stress test though).
The 1080 is close enough that the same numbers may also work, though it's a smaller GPU that will tend to clock a little higher at the same voltage.
A similar procedure will apply to every NVIDIA GPU from the 1xxx, 2xxx, or 3xxx lines (including the OP's 1660), though the specific clocks and voltages that are likely to be stable can vary.
According to Nvidia, NVENC is a dedicated part of the GPU that isn't used for anything else. If that is true, it makes perfect sense that using NVENC adds to the GPU's total heat output.
This is true, but it's power consumption is positively tiny relative to the rest of the GPU, at least at the high-end, and without overriding the reduced power state it forces the memory into, it's probably less than the power consumption of the last 500MHz of my GDDR5/6X OCs on my 1080 Ti or 3080. Same goes for AMDs VCN equivalent.
None of the GPUs I've personally tested appreciably increase in power consumption when encoding with their hardware encoders while gaming, even at fairly extreme settings (4k60 ~100Mbps HEVC, or H264 in the case of my older Pascal parts). it's just so small an addition relative to the rest of the GPU.
Slower card with a capped frame rate...it will be proportionally much more relevant, but the power/temp would still have had to have been fairly borderline for NVENC to be the final straw.