Hey everyone! I've been using Claude Pro for analyzing philosophy books, but I'm finding that I'm really running out of my usage limit fast. I think it might be because the conversations are getting too long and accumulating a lot of previous context. Is that what's causing me to hit the limit so quickly? Any tips on how to manage my usage better?
8 Answers
Exactly! When you engage in a long conversation, Claude has to read everything that's been said, which eats up your token count quickly. A trick you could use is to ask Claude to summarize the conversation whenever you see a warning about token exhaustion. This could help prevent that buildup!
Yes, that's definitely a factor! The longer the context, the more each prompt costs in terms of token usage. Try to be more selective with what you're including. You might also want to look into using a vector database for retrieval-augmented generation (RAG) through the MCP server on desktop Claude.
Which MCP would you recommend?
Honestly, I think Anthropic is dropping the ball with Claude Pro since they released Max. You might be better off switching to another large language model; Claude could be more trouble than it's worth right now.
Which model are you switching to?
For analyzing philosophy books, have you checked out NotebookLM from Google? It might suit your needs better without hitting those limits as quickly.
Interesting! Is NotebookLM user-friendly? How does it compare to Claude for philosophical texts?
If your project doesn't require internet access, consider using Haiku for lower costs per token. It could really help reduce your usage!
Depending on your needs and machine setup, you could go for a local model. I’ve had good results with Mistral-small and DeepSeek. Also, ChatGPT is generally more generous with limits and works well for philosophy texts, so that could be a good option too.
Absolutely! You might want to use Gemini for the initial info gathering, and then ask Claude to expand on those insights. Think of Claude as the wise monk—only ask for the important stuff—but Gemini acts like the helpful assistant to fetch the info first.
Could you give some examples of using a vector database for RAG?