

Or if you have good hardware that doesn’t need the transcoding. If I was loading up h265 video on my server, I’d need to convert it to h264 or something else compatible if I wanted to use it with my iPad, since it’s old enough it doesn’t support doing anything but software decoding of that codec, and it doesn’t have the strongest processor.






It’s something of the law of averages. At their core, an LLM is a sophisticated text prediction algorithm, that boils down the entire corpus of human language into numeric tokens, that it averages out, and creates entire sentences by determining the next most likely word to fill the space.
Given enough data, and you need a tremendous amount of it for an LLM, patterns start to come about, and many of those end up the ones that we see in LLMs.