Hi Dhruv - agreed, waiting 15 seconds until your lights turn on is not an ideal scenario, so maybe that wasn't the best example ;D
I only have started playing with the new serverless inference, so I'm not sure yet how we can optimise the setup to reduce model loading time. Once I find more answers I will be happy to share my findings :)