service_context = ServiceContext.from_defaults(llm='local', chunk_size_limit=3000)
File "/Library/Frameworks/Python.framework/Versions/3.11/lib/python3.11/site-packages/llama_cpp/llama.py", line 900, in _create_completion raise ValueError( ValueError: Requested tokens (3993) exceed context window of 3900