Hexa PA has been updated.
v0.3.0 is a major update, that comes with fixes after an API update broke it, available models updated to the latest gpt-3.5-turbo-0125 and gpt-4-turbo-preview, and the context management has been bufeed.
When set to GPT 3.5 Turbo, Max Input Tokens = 0 (unlimited), Max Context Messages > 0, Max Output Tokens > 0 it switches seemlessly between 4k(latest) and 16k(legacy) models depending on how long the messages are.
It's no longer an even split between input an output tokens, it can now take advantage of GPT-4 Turbo's 128K context even though it only does 4k output. (Successfully tested with more then 4k context.)