The post contains information about loading a GPT-J model, including details about the model's size and memory requirements. Community members discuss the challenges of running the model, noting that it requires a large amount of video RAM (at least 12 GB) to run efficiently on a GPU. Some community members suggest trying to run the model on a CPU instead, but note that this may not be as efficient. One community member mentions having access to an A6000 GPU and plans to experiment with running the model on it.
The GPT-J model is quite big - the compact version of the model uses 16-bit floating point representation of the weights and is still 12 GB big. This means that in order to run inference on your computer, you would need to have a video card with at least 12 GB of video RAM. Alternatively, you can try to run the python implementations on the CPU, but that would probably not be very efficient as they are primarily optimized for running on a GPU (or at least this is my guess - I don't have much experience with python).