I'm using chatgpt's API to discuss book topics. In order for chatgpt to understand the whole story I had to add context.
This means that all user questions and chatgpt replies are sent with the same request. Thus very quickly reaching the maximum support token limit. and usage fees also increase rapidly.
Please show me a short way to reduce the amount of tokens sent, thereby reducing costs.
Below is the example I chatgpt request