Re: Getting Mysql data into Postgres: least painful methods?

Поиск
Список
Период
Сортировка
От Ken Tanzer
Тема Re: Getting Mysql data into Postgres: least painful methods?
Дата
Msg-id CAD3a31VFJ7-wU0sTTDUg5yjo9vEULhFGDq40+NB092=QhS7a+g@mail.gmail.com
обсуждение исходный текст
Ответ на Re: Getting Mysql data into Postgres: least painful methods?  (John R Pierce <pierce@hogranch.com>)
Ответы Re: Getting Mysql data into Postgres: least painful methods?  (Adrian Klaver <adrian.klaver@gmail.com>)
Список pgsql-general
Thanks for all the responses.  I would totally go with the FDW, were I not stuck using (someone else's) 8.4.

In this case it looks like 24 tables, with CSV-like import files totaling 7G.

Since there didn't seem to be a clean, simple and automated path from mysql to postgres, I'm back to skipping mysql entirely and just trying to modify the mysql files to feed directly into postgres.

To that end, they have to be transformed a bit, which I've written a bit of script to accomplish.  I'm wondering if there's a way to avoid creating another 7G of slightly-modified import files before feeding them to postgres.  Specifically, is there a way to do something like

\copy my_table FROM '`cat my_import_file | my_transform_script`'

My 2 goals here are to be somewhat efficient (by not duplicating the input files), and to keep this all within a transaction.  I could have the script transform each file separately and pipe it to postgres:

(echo 'copy mytable from stdin...' ; cat my_import_file | my_transform_script ) | psql

but I'm thinking that there's no way to group those all into a transaction.

Hopefully this makes sense, and any suggestions welcome.  Thanks.

Ken




On Fri, Jan 11, 2013 at 7:13 PM, John R Pierce <pierce@hogranch.com> wrote:
On 1/11/2013 3:54 PM, Ken Tanzer wrote:
Here's the fuller description of what I'm trying to do.  I've got a dataset (a UMLS Metathesaurus subset) that I need to get into a Postgres database.  It's all reference data, and so will be read-only.  There's no functions or logic involved. I anticipate having to update it at least quarterly, so I'd like to get to a well-grooved import process.


how many tables?  if its just one or a couple tables, can you get the data as CSV?   then it would be trivial to import into postgres, using the COPY command (or, \c from psql)...

another alternative, investigate "ETL" tools, these are general purpose data manglers that can connect to a source database (usually any of about 20 supported), extract data, transform it if needed, and load it into a destination database (from a list of 20 or so typically supported)






--
AGENCY Software  
A data system that puts you in control
(253) 245-3801

В списке pgsql-general по дате отправления:

Предыдущее
От: Edson Richter
Дата:
Сообщение: Re: Enhancement proposal - detect chain of triggers from inside the trigger
Следующее
От: Scott Marlowe
Дата:
Сообщение: Re: Linux Distribution Preferences?