An Analysis of DeepMind's 'Language Modeling Is Compression' Paper
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearAB
    abhi9u
    Now 100%

    Interesting. I'm just thinking aloud to understand this.

    In this case, the models are looking at a few sequence of bytes in their context and are able to predict the next byte(s) with good accuracy, which allows efficient encoding. Most of our memories are associative, i.e. we associate them with some concept/name/idea. So, do you mean, our brain uses the concept to predict a token which gets decoded in the form of a memory?

    4
  • An Analysis of DeepMind's 'Language Modeling Is Compression' Paper
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearAB
    abhi9u
    Now 100%

    Yes. They also mention that using such large models for compression is not practical because their size thwarts any amount of data you might want to compress. But this result gives a good picture into how generalized such large models are, and how well they are able to predict the next tokens for image/audio data at a high accuracy.

    2
  • An Analysis of DeepMind's 'Language Modeling Is Compression' Paper
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearAB
    abhi9u
    Now 100%

    Do you mean the number of tokens in the LLM's tokenizer, or the dictionary size of the compression algorithm?

    The vocab size of the pretrained models is not mentioned anywhere in the paper. Although, they did conduct an experiment where they measured compression performance while using tokenizers of different vocabulary sizes.

    If you meant the dictionary size of the compression algorithm, then there was no dictionary because they only used arithmetic coding to do the compression which doesn't use dictionaries.

    3
  • How to build a computer using origami
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearAB
    abhi9u
    Now 100%

    I don't know. I have found that the folks on Technology community appreciate many of my computer science posts. But a dedicated Comp Science community which is active, will be awesome.

    2
  • Why and How Does Python Use Bloom Filters in String Processing?
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearAB
    abhi9u
    Now 100%

    Thank you! That's helpful. I spent quite some time trying to understand the difference between UTF-8 and Python's representation and arrived at the same understanding as you wrote. However, most of the external documents simply say that strings in Python are UTF-8 which made me conclude that perhaps I am missing something and it might be safer to write it as utf-8.

    I will look more in the code as you suggested.

    1
  • Why and How Does Python Use Bloom Filters in String Processing?
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearAB
    abhi9u
    Now 100%

    Hi @qwop, I am the author. Thank you for reading and the kind words. I would like to understand the error I made better so that I don't repeat in future, and if I can fix it. Could you please clarify?

    3
  • "Initials" by "Florian Körner", licensed under "CC0 1.0". / Remix of the original. - Created with dicebear.comInitialsFlorian Körnerhttps://github.com/dicebear/dicebearAB
    Now
    25 10

    abhi9u

    lemmy.world