The model was trained using text databases from the Internet. This included a whopping 570 GB of data obtained from books, web texts, Wikipedia, articles and others written on the Internet. To be even more exact, 300 billion words were introduced into the system. ChatGPT has been optimized based on GPT-3.5, a language model designed to produce text.
ChatGPT was optimized for dialogue using reinforcement learning with human feedback (RLHF), a method that uses human demonstrations and comparison of preferences to guide the model to the desired behavior. ChatGPT is an extrapolation of a class of machine learning natural language processing models known as large language models (LLM). LLMs digest enormous amounts of text data and infer relationships between words within the text. These models have grown in recent years as we have seen advances in computing power.
LLMs increase their capacity as they increase the size of their input data sets and the parameter space.