Re: Reducing the overhead of NUMERIC data
| От | Tom Lane |
|---|---|
| Тема | Re: Reducing the overhead of NUMERIC data |
| Дата | |
| Msg-id | 8698.1131070663@sss.pgh.pa.us обсуждение исходный текст |
| Ответ на | Re: Reducing the overhead of NUMERIC data (Gregory Maxwell <gmaxwell@gmail.com>) |
| Ответы |
Re: Reducing the overhead of NUMERIC data
|
| Список | pgsql-hackers |
Gregory Maxwell <gmaxwell@gmail.com> writes:
> Another way to look at this is in the context of compression: With
> unicode, characters are really 32bit values... But only a small range
> of these values is common. So we store and work with them in a
> compressed format, UTF-8.
> As such it might be more interesting to ask some other questions like:
> are we using the best compression algorithm for the application, and,
> why do we sometimes stack two compression algorithms?
Actually, the real reason we use UTF-8 and not any of the
sorta-fixed-size representations of Unicode is that the backend is by
and large an ASCII, null-terminated-string engine. *All* of the
supported backend encodings are ASCII-superset codes. Making
everything null-safe in order to allow use of UCS2 or UCS4 would be
a huge amount of work, and the benefit is at best questionable.
regards, tom lane
В списке pgsql-hackers по дате отправления: