Thread: Re: [GENERAL] Re: [PATCH] Contrib C source for casting MONEY to INT[248] and FLOAT[48]
Re: [GENERAL] Re: [PATCH] Contrib C source for casting MONEY to INT[248] and FLOAT[48]
From
ÀîÁ¢ÐÂ
Date:
Bruce Momjian: I am a begineer,The question is PgSQL support the full entrity integrity and refernece integerity.For example.does it support "Restricted Delete、NULLIFIES-delete,default-delete....",I read yourbook,But can not find detail.Where to find? >---------------------------(end of broadcast)--------------------------- >TIP 2: you can get off all lists at once with the unregister command > (send "unregister YourEmailAddressHere" to majordomo@postgresql.org) lilixin@cqu.edu.cn >---------------------------(end of broadcast)--------------------------- >TIP 2: you can get off all lists at once with the unregister command > (send "unregister YourEmailAddressHere" to majordomo@postgresql.org) 致 礼! 李立新 lilixin@cqu.edu.cn
* Naomi Walker <nwalker@eldocomp.com> [010706 17:57]: > If PostgreSQL is run on a system that has a file size limit (2 gig?), where > might cause us to hit the limit? PostgreSQL is smart, and breaks the table files up at ~1GB per each, so it's transparent to you. You shouldn't have to worry about it. LER > -- > Naomi Walker > Chief Information Officer > Eldorado Computing, Inc. > 602-604-3100 ext 242 > > > ---------------------------(end of broadcast)--------------------------- > TIP 4: Don't 'kill -9' the postmaster > -- Larry Rosenman http://www.lerctr.org/~ler Phone: +1 972-414-9812 E-Mail: ler@lerctr.org US Mail: 1905 Steamboat Springs Drive, Garland, TX 75044-6749
If PostgreSQL is run on a system that has a file size limit (2 gig?), where might cause us to hit the limit? -- Naomi Walker Chief Information Officer Eldorado Computing, Inc. 602-604-3100 ext 242
(This question was answered several days ago on this list; please check the list archives before posting. I believe it's also in the FAQ.) > If PostgreSQL is run on a system that has a file size limit (2 > gig?), where might cause us to hit the limit? Postgres will never internally use files (e.g. for tables, indexes, etc) larger than 1GB -- at that point, the file is split. However, you might run into problems when you export the data from Pg to another source, such as if you pg_dump the contents of a database > 2GB. In that case, filter pg_dump through gzip or bzip2 to reduce the size of the dump. If that's still not enough, you can dump individual tables (with -t) or use 'split' to divide the dump into several files. Cheers, Neil