text-generation-inference causal-lm int8 PyTorch PostTrainingStatic IntelĀ® Neural Compressor neural-compressor

INT8 GPT-J 6B

GPT-J 6B is a transformer model trained using Ben Wang's Mesh Transformer JAX. "GPT-J" refers to the class of model, while "6B" represents the number of trainable parameters.

This int8 PyTorch model is generated by neural-compressor.