On 2018/03/12 02:07, Keith Turner via Unicode wrote:
Yeah, it certainly results in larger utf8 strings. For example a sha256
hash is 112 bytes when encoded as Ecoji utf8. For base64, sha256 is 44
bytes.
Even though its more bytes, Ecoji has less visible characters than base64
for sha256.
Oh, let him have a little fun. At least he's using emoji for something
related to characters, instead of playing Mr. Potato Head.
Incidentally, more prior art on large-base encoding:
https://sites.google.com/site/markusicu/unicode/base16k
--
Doug Ewell | Thornton, CO, US | ewellic.org
On Sun, Mar 11, 2018 at 11:25 AM, Philippe Verdy wrote:
> Ideally, the purpose of such base-1024 encoding is to allow compacting
> arbitrary data into plain-text which can be safely preserved including by
> Unicode normalization and transforms by encoding like UTF-8.
> But
Ideally, the purpose of such base-1024 encoding is to allow compacting
arbitrary data into plain-text which can be safely preserved including by
Unicode normalization and transforms by encoding like UTF-8.
But then we have a way to do that is such a way that this minimizes the
UTF-8 string sizes
Neat! Prior art:
- https://github.com/watson/base64-emoji
- https://github.com/nate-parrott/emojicode
On Sun, Mar 11, 2018 at 6:04 AM, Keith Turner via Unicode <
unicode@unicode.org> wrote:
> I created a neat little project based on Unicode emojis. I thought
> some on this list may find
I created a neat little project based on Unicode emojis. I thought
some on this list may find it interesting. It encodes arbitrary data
as 1024 emojis. The project is called Ecoji and is hosted on github
at https://github.com/keith-turner/ecoji
Below are some examples of encoding and decoding.
6 matches
Mail list logo