Re: Persist MVCC forever - retain history

Поиск
Список
Период
Сортировка
От Mark Dilger
Тема Re: Persist MVCC forever - retain history
Дата
Msg-id D809DE2C-FA2A-4C31-AA54-1B0F5E974BC3@enterprisedb.com
обсуждение исходный текст
Ответ на Re: Persist MVCC forever - retain history  (Mitar <mmitar@gmail.com>)
Ответы Re: Persist MVCC forever - retain history  (Mitar <mmitar@gmail.com>)
Список pgsql-hackers

> On Jul 2, 2020, at 5:58 PM, Mitar <mmitar@gmail.com> wrote:
>
>> Plus, wrap-around and freezing aren’t just nice-to-have features.
>
> Oh, I forgot about that. ctid is still just 32 bits? So then for such
> table with permanent MVCC this would have to be increased, to like 64
> bits or something. Then one would not have to do wrap-around
> protection, no?

I think what you propose is a huge undertaking, and would likely result in a fork of postgres not compatible with the
publicsources.  I do not recommend the project.  But in answer to your question.... 

Yes, the values stored in the tuple header are 32 bits.  Take a look in access/htup_details.h.  You'll notice that
HeapTupleHeaderDatahas a union: 

    union
    {
        HeapTupleFields t_heap;
        DatumTupleFields t_datum;
    }           t_choice;

If you check, HeapTupleFields and DatumTupleFields are the same size, each having three 32 bit values, though they mean
differentthings.  You may need to expand types TransactionId, CommandId, and Oid to 64 bits, expand varlena headers to
64bits, and typemods to 64 bits.  You may find that it is harder to just expand a subset of those, given the way these
fieldsoverlay in these unions.  There will be lot of busy work going through the code to adjust everything else to
match. Just updating printf style formatting in error messages may take a long time. 

If you do choose to expand only some of the types, say just TransactionId and CommandId, you'll have to deal with the
sizemismatch between HeapTupleFields and DatumTupleFields. 

Aborted transactions leave dead rows in your tables, and you may want to deal with that for performance reasons.  Even
ifyou don't intend to remove deleted rows, because you are just going to keep them around for time travel purposes, you
mightstill want to use vacuum to remove dead rows, those that never committed. 

You'll need to think about how to manage the growing clog if you don't intend to truncate it periodically.  Or if you
dointend to truncate clog periodically, you'll need to think about the fact that you have TransactionIds in your tables
olderthan what clog knows about. 

You may want to think about how keeping dead rows around affects index performance.

I expect these issues to be less than half what you would need to resolve, though much of the rest of it is less clear
tome. 

—
Mark Dilger
EnterpriseDB: http://www.enterprisedb.com
The Enterprise PostgreSQL Company






В списке pgsql-hackers по дате отправления:

Предыдущее
От: Fujii Masao
Дата:
Сообщение: Re: track_planning causing performance regression
Следующее
От: Etsuro Fujita
Дата:
Сообщение: Re: estimation problems for DISTINCT ON with FDW