Thread: Fix for large file support

Fix for large file support

From
Zdenek Kotala
Date:
Current version of postgres support only 1GB chunks. This limit is
defined in the pg_config_manual.h header file. However this setting
allows to have maximal 2GB chunks. Main problem is that md storage
manager and buffile use "long" data type (32bits) for offset instead
"off_t" defined in <sys/types.h>.

off_t is 32bits long on 32bits OS and 64bits long on 64bits OS or when
application is compiled with large file support.

Attached patch allow to setup bigger chunks than 4GB on OS with large
file support.

I tested it on 7GB table and it works.


Please, look on it and let me know your comments or if I miss something.


TODO/questions:

1) clean/update comments about limitation

2) Is there some doc for update?

3) I would like to add some check compare sizeof(off_t) and chunk size
setting and protect postgres with missconfigured chunk size. Is mdinit()
good place for this check?

4) I'm going to take bigger machine for test with really big table.


    with regards Zdenek


Index: src/backend/storage/file/buffile.c
===================================================================
RCS file: /projects/cvsroot/pgsql/src/backend/storage/file/buffile.c,v
retrieving revision 1.25
diff -c -r1.25 buffile.c
*** src/backend/storage/file/buffile.c    5 Jan 2007 22:19:37 -0000    1.25
--- src/backend/storage/file/buffile.c    6 Apr 2007 12:08:47 -0000
***************
*** 42,48 ****
   * Note we adhere to this limit whether or not LET_OS_MANAGE_FILESIZE
   * is defined, although md.c ignores it when that symbol is defined.
   */
! #define MAX_PHYSICAL_FILESIZE  (RELSEG_SIZE * BLCKSZ)

  /*
   * This data structure represents a buffered file that consists of one or
--- 42,48 ----
   * Note we adhere to this limit whether or not LET_OS_MANAGE_FILESIZE
   * is defined, although md.c ignores it when that symbol is defined.
   */
! #define MAX_PHYSICAL_FILESIZE  ((off_t)RELSEG_SIZE * BLCKSZ)

  /*
   * This data structure represents a buffered file that consists of one or
***************
*** 54,60 ****
      int            numFiles;        /* number of physical files in set */
      /* all files except the last have length exactly MAX_PHYSICAL_FILESIZE */
      File       *files;            /* palloc'd array with numFiles entries */
!     long       *offsets;        /* palloc'd array with numFiles entries */

      /*
       * offsets[i] is the current seek position of files[i].  We use this to
--- 54,60 ----
      int            numFiles;        /* number of physical files in set */
      /* all files except the last have length exactly MAX_PHYSICAL_FILESIZE */
      File       *files;            /* palloc'd array with numFiles entries */
!     off_t       *offsets;        /* palloc'd array with numFiles entries */

      /*
       * offsets[i] is the current seek position of files[i].  We use this to
***************
*** 70,76 ****
       * Position as seen by user of BufFile is (curFile, curOffset + pos).
       */
      int            curFile;        /* file index (0..n) part of current pos */
!     int            curOffset;        /* offset part of current pos */
      int            pos;            /* next read/write position in buffer */
      int            nbytes;            /* total # of valid bytes in buffer */
      char        buffer[BLCKSZ];
--- 70,76 ----
       * Position as seen by user of BufFile is (curFile, curOffset + pos).
       */
      int            curFile;        /* file index (0..n) part of current pos */
!     off_t        curOffset;        /* offset part of current pos */
      int            pos;            /* next read/write position in buffer */
      int            nbytes;            /* total # of valid bytes in buffer */
      char        buffer[BLCKSZ];
***************
*** 95,101 ****
      file->numFiles = 1;
      file->files = (File *) palloc(sizeof(File));
      file->files[0] = firstfile;
!     file->offsets = (long *) palloc(sizeof(long));
      file->offsets[0] = 0L;
      file->isTemp = false;
      file->dirty = false;
--- 95,101 ----
      file->numFiles = 1;
      file->files = (File *) palloc(sizeof(File));
      file->files[0] = firstfile;
!     file->offsets = (off_t *) palloc(sizeof(off_t));
      file->offsets[0] = 0L;
      file->isTemp = false;
      file->dirty = false;
***************
*** 121,128 ****

      file->files = (File *) repalloc(file->files,
                                      (file->numFiles + 1) * sizeof(File));
!     file->offsets = (long *) repalloc(file->offsets,
!                                       (file->numFiles + 1) * sizeof(long));
      file->files[file->numFiles] = pfile;
      file->offsets[file->numFiles] = 0L;
      file->numFiles++;
--- 121,128 ----

      file->files = (File *) repalloc(file->files,
                                      (file->numFiles + 1) * sizeof(File));
!     file->offsets = (off_t *) repalloc(file->offsets,
!                                       (file->numFiles + 1) * sizeof(off_t));
      file->files[file->numFiles] = pfile;
      file->offsets[file->numFiles] = 0L;
      file->numFiles++;
***************
*** 273,281 ****
          bytestowrite = file->nbytes - wpos;
          if (file->isTemp)
          {
!             long        availbytes = MAX_PHYSICAL_FILESIZE - file->curOffset;

!             if ((long) bytestowrite > availbytes)
                  bytestowrite = (int) availbytes;
          }

--- 273,281 ----
          bytestowrite = file->nbytes - wpos;
          if (file->isTemp)
          {
!             off_t        availbytes = MAX_PHYSICAL_FILESIZE - file->curOffset;

!             if ((off_t) bytestowrite > availbytes)
                  bytestowrite = (int) availbytes;
          }

***************
*** 445,454 ****
   * impossible seek is attempted.
   */
  int
! BufFileSeek(BufFile *file, int fileno, long offset, int whence)
  {
      int            newFile;
!     long        newOffset;

      switch (whence)
      {
--- 445,454 ----
   * impossible seek is attempted.
   */
  int
! BufFileSeek(BufFile *file, int fileno, off_t offset, int whence)
  {
      int            newFile;
!     off_t        newOffset;

      switch (whence)
      {
***************
*** 531,537 ****
  }

  void
! BufFileTell(BufFile *file, int *fileno, long *offset)
  {
      *fileno = file->curFile;
      *offset = file->curOffset + file->pos;
--- 530,536 ----
  }

  void
! BufFileTell(BufFile *file, int *fileno, off_t *offset)
  {
      *fileno = file->curFile;
      *offset = file->curOffset + file->pos;
***************
*** 544,559 ****
   * the file.  Note that users of this interface will fail if their files
   * exceed BLCKSZ * LONG_MAX bytes, but that is quite a lot; we don't work
   * with tables bigger than that, either...
   *
   * Result is 0 if OK, EOF if not.  Logical position is not moved if an
   * impossible seek is attempted.
   */
  int
! BufFileSeekBlock(BufFile *file, long blknum)
  {
      return BufFileSeek(file,
                         (int) (blknum / RELSEG_SIZE),
!                        (blknum % RELSEG_SIZE) * BLCKSZ,
                         SEEK_SET);
  }

--- 543,558 ----
   * the file.  Note that users of this interface will fail if their files
   * exceed BLCKSZ * LONG_MAX bytes, but that is quite a lot; we don't work
   * with tables bigger than that, either...
   *
   * Result is 0 if OK, EOF if not.  Logical position is not moved if an
   * impossible seek is attempted.
   */
  int
! BufFileSeekBlock(BufFile *file, BlockNumber blknum)
  {
      return BufFileSeek(file,
                         (int) (blknum / RELSEG_SIZE),
!                        ((off_t)blknum % RELSEG_SIZE) * BLCKSZ,
                         SEEK_SET);
  }

Index: src/backend/storage/file/fd.c
===================================================================
RCS file: /projects/cvsroot/pgsql/src/backend/storage/file/fd.c,v
retrieving revision 1.137
diff -c -r1.137 fd.c
*** src/backend/storage/file/fd.c    6 Mar 2007 02:06:14 -0000    1.137
--- src/backend/storage/file/fd.c    6 Apr 2007 12:08:47 -0000
***************
*** 128,134 ****
      File        nextFree;        /* link to next free VFD, if in freelist */
      File        lruMoreRecently;    /* doubly linked recency-of-use list */
      File        lruLessRecently;
!     long        seekPos;        /* current logical file position */
      char       *fileName;        /* name of file, or NULL for unused VFD */
      /* NB: fileName is malloc'd, and must be free'd when closing the VFD */
      int            fileFlags;        /* open(2) flags for (re)opening the file */
--- 128,134 ----
      File        nextFree;        /* link to next free VFD, if in freelist */
      File        lruMoreRecently;    /* doubly linked recency-of-use list */
      File        lruLessRecently;
!     off_t        seekPos;        /* current logical file position */
      char       *fileName;        /* name of file, or NULL for unused VFD */
      /* NB: fileName is malloc'd, and must be free'd when closing the VFD */
      int            fileFlags;        /* open(2) flags for (re)opening the file */
***************
*** 1136,1143 ****
      return pg_fsync(VfdCache[file].fd);
  }

! long
! FileSeek(File file, long offset, int whence)
  {
      int            returnCode;

--- 1136,1143 ----
      return pg_fsync(VfdCache[file].fd);
  }

! off_t
! FileSeek(File file, off_t offset, int whence)
  {
      int            returnCode;

***************
*** 1203,1209 ****
   * XXX not actually used but here for completeness
   */
  #ifdef NOT_USED
! long
  FileTell(File file)
  {
      Assert(FileIsValid(file));
--- 1203,1209 ----
   * XXX not actually used but here for completeness
   */
  #ifdef NOT_USED
! off_t
  FileTell(File file)
  {
      Assert(FileIsValid(file));
***************
*** 1214,1220 ****
  #endif

  int
! FileTruncate(File file, long offset)
  {
      int            returnCode;

--- 1214,1220 ----
  #endif

  int
! FileTruncate(File file, off_t offset)
  {
      int            returnCode;

***************
*** 1227,1233 ****
      if (returnCode < 0)
          return returnCode;

!     returnCode = ftruncate(VfdCache[file].fd, (size_t) offset);
      return returnCode;
  }

--- 1227,1233 ----
      if (returnCode < 0)
          return returnCode;

!     returnCode = ftruncate(VfdCache[file].fd, (off_t) offset);
      return returnCode;
  }

Index: src/backend/storage/smgr/md.c
===================================================================
RCS file: /projects/cvsroot/pgsql/src/backend/storage/smgr/md.c,v
retrieving revision 1.127
diff -c -r1.127 md.c
*** src/backend/storage/smgr/md.c    17 Jan 2007 16:25:01 -0000    1.127
--- src/backend/storage/smgr/md.c    6 Apr 2007 12:08:48 -0000
***************
*** 325,331 ****
  void
  mdextend(SMgrRelation reln, BlockNumber blocknum, char *buffer, bool isTemp)
  {
!     long        seekpos;
      int            nbytes;
      MdfdVec    *v;

--- 325,331 ----
  void
  mdextend(SMgrRelation reln, BlockNumber blocknum, char *buffer, bool isTemp)
  {
!     off_t        seekpos;
      int            nbytes;
      MdfdVec    *v;

***************
*** 351,360 ****
      v = _mdfd_getseg(reln, blocknum, isTemp, EXTENSION_CREATE);

  #ifndef LET_OS_MANAGE_FILESIZE
!     seekpos = (long) (BLCKSZ * (blocknum % ((BlockNumber) RELSEG_SIZE)));
!     Assert(seekpos < BLCKSZ * RELSEG_SIZE);
  #else
!     seekpos = (long) (BLCKSZ * (blocknum));
  #endif

      /*
--- 351,360 ----
      v = _mdfd_getseg(reln, blocknum, isTemp, EXTENSION_CREATE);

  #ifndef LET_OS_MANAGE_FILESIZE
!     seekpos =  (off_t)BLCKSZ * (blocknum % ((BlockNumber) RELSEG_SIZE));
!     Assert(seekpos < (off_t)BLCKSZ * RELSEG_SIZE);
  #else
!     seekpos =  (off_t)BLCKSZ * blocknum;
  #endif

      /*
***************
*** 507,523 ****
  void
  mdread(SMgrRelation reln, BlockNumber blocknum, char *buffer)
  {
!     long        seekpos;
      int            nbytes;
      MdfdVec    *v;

      v = _mdfd_getseg(reln, blocknum, false, EXTENSION_FAIL);

  #ifndef LET_OS_MANAGE_FILESIZE
!     seekpos = (long) (BLCKSZ * (blocknum % ((BlockNumber) RELSEG_SIZE)));
!     Assert(seekpos < BLCKSZ * RELSEG_SIZE);
  #else
!     seekpos = (long) (BLCKSZ * (blocknum));
  #endif

      if (FileSeek(v->mdfd_vfd, seekpos, SEEK_SET) != seekpos)
--- 507,523 ----
  void
  mdread(SMgrRelation reln, BlockNumber blocknum, char *buffer)
  {
!     off_t        seekpos;
      int            nbytes;
      MdfdVec    *v;

      v = _mdfd_getseg(reln, blocknum, false, EXTENSION_FAIL);

  #ifndef LET_OS_MANAGE_FILESIZE
!     seekpos = (off_t)BLCKSZ * (blocknum % ((BlockNumber) RELSEG_SIZE));
!     Assert(seekpos < (off_t)BLCKSZ * RELSEG_SIZE);
  #else
!     seekpos = (off_t)BLCKSZ * (blocknum);
  #endif

      if (FileSeek(v->mdfd_vfd, seekpos, SEEK_SET) != seekpos)
***************
*** 571,577 ****
  void
  mdwrite(SMgrRelation reln, BlockNumber blocknum, char *buffer, bool isTemp)
  {
!     long        seekpos;
      int            nbytes;
      MdfdVec    *v;

--- 571,577 ----
  void
  mdwrite(SMgrRelation reln, BlockNumber blocknum, char *buffer, bool isTemp)
  {
!     off_t        seekpos;
      int            nbytes;
      MdfdVec    *v;

***************
*** 583,592 ****
      v = _mdfd_getseg(reln, blocknum, isTemp, EXTENSION_FAIL);

  #ifndef LET_OS_MANAGE_FILESIZE
!     seekpos = (long) (BLCKSZ * (blocknum % ((BlockNumber) RELSEG_SIZE)));
!     Assert(seekpos < BLCKSZ * RELSEG_SIZE);
  #else
!     seekpos = (long) (BLCKSZ * (blocknum));
  #endif

      if (FileSeek(v->mdfd_vfd, seekpos, SEEK_SET) != seekpos)
--- 583,592 ----
      v = _mdfd_getseg(reln, blocknum, isTemp, EXTENSION_FAIL);

  #ifndef LET_OS_MANAGE_FILESIZE
!     seekpos = (off_t)BLCKSZ * (blocknum % ((BlockNumber) RELSEG_SIZE));
!     Assert(seekpos < (off_t)BLCKSZ * RELSEG_SIZE);
  #else
!     seekpos = (off_t)BLCKSZ * (blocknum);
  #endif

      if (FileSeek(v->mdfd_vfd, seekpos, SEEK_SET) != seekpos)
***************
*** 1297,1303 ****
  static BlockNumber
  _mdnblocks(SMgrRelation reln, MdfdVec *seg)
  {
!     long        len;

      len = FileSeek(seg->mdfd_vfd, 0L, SEEK_END);
      if (len < 0)
--- 1297,1303 ----
  static BlockNumber
  _mdnblocks(SMgrRelation reln, MdfdVec *seg)
  {
!     off_t        len;

      len = FileSeek(seg->mdfd_vfd, 0L, SEEK_END);
      if (len < 0)
Index: src/backend/utils/sort/tuplestore.c
===================================================================
RCS file: /projects/cvsroot/pgsql/src/backend/utils/sort/tuplestore.c,v
retrieving revision 1.30
diff -c -r1.30 tuplestore.c
*** src/backend/utils/sort/tuplestore.c    5 Jan 2007 22:19:47 -0000    1.30
--- src/backend/utils/sort/tuplestore.c    6 Apr 2007 12:08:49 -0000
***************
*** 130,143 ****
      bool        eof_reached;    /* read reached EOF (always valid) */
      int            current;        /* next array index (valid if INMEM) */
      int            readpos_file;    /* file# (valid if WRITEFILE and not eof) */
!     long        readpos_offset; /* offset (valid if WRITEFILE and not eof) */
      int            writepos_file;    /* file# (valid if READFILE) */
!     long        writepos_offset;    /* offset (valid if READFILE) */

      /* markpos_xxx holds marked position for mark and restore */
      int            markpos_current;    /* saved "current" */
      int            markpos_file;    /* saved "readpos_file" */
!     long        markpos_offset; /* saved "readpos_offset" */
  };

  #define COPYTUP(state,tup)    ((*(state)->copytup) (state, tup))
--- 130,143 ----
      bool        eof_reached;    /* read reached EOF (always valid) */
      int            current;        /* next array index (valid if INMEM) */
      int            readpos_file;    /* file# (valid if WRITEFILE and not eof) */
!     off_t        readpos_offset; /* offset (valid if WRITEFILE and not eof) */
      int            writepos_file;    /* file# (valid if READFILE) */
!     off_t        writepos_offset;    /* offset (valid if READFILE) */

      /* markpos_xxx holds marked position for mark and restore */
      int            markpos_current;    /* saved "current" */
      int            markpos_file;    /* saved "readpos_file" */
!     off_t        markpos_offset; /* saved "readpos_offset" */
  };

  #define COPYTUP(state,tup)    ((*(state)->copytup) (state, tup))
Index: src/include/storage/buffile.h
===================================================================
RCS file: /projects/cvsroot/pgsql/src/include/storage/buffile.h,v
retrieving revision 1.20
diff -c -r1.20 buffile.h
*** src/include/storage/buffile.h    5 Jan 2007 22:19:57 -0000    1.20
--- src/include/storage/buffile.h    6 Apr 2007 12:08:49 -0000
***************
*** 26,31 ****
--- 26,34 ----
  #ifndef BUFFILE_H
  #define BUFFILE_H

+ #include <sys/types.h>
+ #include "block.h"
+
  /* BufFile is an opaque type whose details are not known outside buffile.c. */

  typedef struct BufFile BufFile;
***************
*** 38,45 ****
  extern void BufFileClose(BufFile *file);
  extern size_t BufFileRead(BufFile *file, void *ptr, size_t size);
  extern size_t BufFileWrite(BufFile *file, void *ptr, size_t size);
! extern int    BufFileSeek(BufFile *file, int fileno, long offset, int whence);
! extern void BufFileTell(BufFile *file, int *fileno, long *offset);
! extern int    BufFileSeekBlock(BufFile *file, long blknum);

  #endif   /* BUFFILE_H */
--- 41,48 ----
  extern void BufFileClose(BufFile *file);
  extern size_t BufFileRead(BufFile *file, void *ptr, size_t size);
  extern size_t BufFileWrite(BufFile *file, void *ptr, size_t size);
! extern int    BufFileSeek(BufFile *file, int fileno, off_t offset, int whence);
! extern void BufFileTell(BufFile *file, int *fileno, off_t *offset);
! extern int    BufFileSeekBlock(BufFile *file, BlockNumber blknum);

  #endif   /* BUFFILE_H */
Index: src/include/storage/fd.h
===================================================================
RCS file: /projects/cvsroot/pgsql/src/include/storage/fd.h,v
retrieving revision 1.57
diff -c -r1.57 fd.h
*** src/include/storage/fd.h    5 Jan 2007 22:19:57 -0000    1.57
--- src/include/storage/fd.h    6 Apr 2007 12:08:50 -0000
***************
*** 67,74 ****
  extern int    FileRead(File file, char *buffer, int amount);
  extern int    FileWrite(File file, char *buffer, int amount);
  extern int    FileSync(File file);
! extern long FileSeek(File file, long offset, int whence);
! extern int    FileTruncate(File file, long offset);

  /* Operations that allow use of regular stdio --- USE WITH CAUTION */
  extern FILE *AllocateFile(const char *name, const char *mode);
--- 67,74 ----
  extern int    FileRead(File file, char *buffer, int amount);
  extern int    FileWrite(File file, char *buffer, int amount);
  extern int    FileSync(File file);
! extern off_t FileSeek(File file, off_t offset, int whence);
! extern int    FileTruncate(File file, off_t offset);

  /* Operations that allow use of regular stdio --- USE WITH CAUTION */
  extern FILE *AllocateFile(const char *name, const char *mode);

Re: Fix for large file support

From
Andrew Dunstan
Date:

Zdenek Kotala wrote:
> Current version of postgres support only 1GB chunks. This limit is
> defined in the pg_config_manual.h header file. However this setting
> allows to have maximal 2GB chunks. Main problem is that md storage
> manager and buffile use "long" data type (32bits) for offset instead
> "off_t" defined in <sys/types.h>.
>
> off_t is 32bits long on 32bits OS and 64bits long on 64bits OS or when
> application is compiled with large file support.
>
> Attached patch allow to setup bigger chunks than 4GB on OS with large
> file support.
>
> I tested it on 7GB table and it works.
>
>
>

What does it actually buy us, though? Does it mean the maximum field
size will grow beyond 1Gb? Or give better performance?

cheers

andrew

Re: Fix for large file support

From
Zdenek Kotala
Date:
Andrew Dunstan wrote:
>
>
> Does it mean the maximum field size will grow beyond 1Gb?

No. Because it is limited by varlena size. See
http://www.postgresql.org/docs/8.2/interactive/storage-toast.html

> Or give better performance?

Yes. List of chunks is stored as linked list and for some operation
(e.g. expand) are all chunks opened and their size is checked. On big
tables it takes some time. For example if you have 1TB big table and you
want to add new block you must go and open all 1024 files.

By the way ./configure script performs check for __LARGE_FILE_ support,
but it looks that it is nowhere used.

There could be small time penalty in 64bit arithmetics. However it
happens only if large file support is enabled on 32bit OS.

    Zdenek


Re: Fix for large file support

From
Tom Lane
Date:
Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
> Andrew Dunstan wrote:
>> Or give better performance?

> Yes. List of chunks is stored as linked list and for some operation
> (e.g. expand) are all chunks opened and their size is checked. On big
> tables it takes some time. For example if you have 1TB big table and you
> want to add new block you must go and open all 1024 files.

Indeed, but that would be far more effectively addressed by fixing the
*other* code path that doesn't segment at all (the
LET_OS_MANAGE_FILESIZE option, which is most likely broken these days
for lack of testing).  I don't see the point of a halfway measure like
increasing RELSEG_SIZE.

            regards, tom lane

Re: Fix for large file support

From
Zdenek Kotala
Date:
Tom Lane wrote:
> Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
>> Andrew Dunstan wrote:

> Indeed, but that would be far more effectively addressed by fixing the
> *other* code path that doesn't segment at all (the
> LET_OS_MANAGE_FILESIZE option, which is most likely broken these days
> for lack of testing).  I don't see the point of a halfway measure like
> increasing RELSEG_SIZE.

LET_OS_MANAGE_FILESIZE is good way. I think one problem of this option I
fixed. It is size of offset. I went thru the code and did not see any
other problem there. However, how you mentioned it need more testing. I
going to take server with large disk array and I will test it.

I would like to add --enable-largefile switch to configure file to
enable access to wide group of users. What you think about it?

        Zdenek

Re: Fix for large file support (nonsegment mode support)

From
Zdenek Kotala
Date:
Tom Lane wrote:
> [ redirecting to -hackers for wider comment ]
>
> Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
>> Tom Lane wrote:
>> LET_OS_MANAGE_FILESIZE is good way. I think one problem of this option I
>> fixed. It is size of offset. I went thru the code and did not see any
>> other problem there. However, how you mentioned it need more testing. I
>> going to take server with large disk array and I will test it.
>
>> I would like to add --enable-largefile switch to configure file to
>> enable access to wide group of users. What you think about it?
>
> Yeah, I was going to suggest the same thing --- but not with that switch
> name.  We already use enable/disable-largefile to control whether 64-bit
> file access is built at all (this mostly affects pg_dump at the moment).
>
> I think the clearest way might be to flip the sense of the variable.
> I never found "LET_OS_MANAGE_FILESIZE" to be a good name anyway.  I'd
> suggest "USE_SEGMENTED_FILES", which defaults to "on", and you can
> turn it off via --disable-segmented-files if configure confirms your
> OS has largefile support (thus you could not specify both this and
> --disable-largefile).
>

There is latest version of nonsegment support patch. I changed
LET_OS_MANAGE_FILESIZE to USE_SEGMENTED_FILES and I added
-disable-segmented-files switch to configure. I kept tuplestore behavior
and it still split file in both mode.

I also little bit cleanup some other datatypes (e.g int->mode_t).
Autoconf and autoheader must be run after patch application.

I tested it with 9GB table and both mode works fine.

Please, let me know your comments.

        Zdenek

Attachment

Re: Fix for large file support (nonsegment mode support)

From
Bruce Momjian
Date:
This has been saved for the 8.4 release:

    http://momjian.postgresql.org/cgi-bin/pgpatches_hold

---------------------------------------------------------------------------

 Zdenek Kotala wrote:
> Tom Lane wrote:
> > [ redirecting to -hackers for wider comment ]
> >
> > Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
> >> Tom Lane wrote:
> >> LET_OS_MANAGE_FILESIZE is good way. I think one problem of this option I
> >> fixed. It is size of offset. I went thru the code and did not see any
> >> other problem there. However, how you mentioned it need more testing. I
> >> going to take server with large disk array and I will test it.
> >
> >> I would like to add --enable-largefile switch to configure file to
> >> enable access to wide group of users. What you think about it?
> >
> > Yeah, I was going to suggest the same thing --- but not with that switch
> > name.  We already use enable/disable-largefile to control whether 64-bit
> > file access is built at all (this mostly affects pg_dump at the moment).
> >
> > I think the clearest way might be to flip the sense of the variable.
> > I never found "LET_OS_MANAGE_FILESIZE" to be a good name anyway.  I'd
> > suggest "USE_SEGMENTED_FILES", which defaults to "on", and you can
> > turn it off via --disable-segmented-files if configure confirms your
> > OS has largefile support (thus you could not specify both this and
> > --disable-largefile).
> >
>
> There is latest version of nonsegment support patch. I changed
> LET_OS_MANAGE_FILESIZE to USE_SEGMENTED_FILES and I added
> -disable-segmented-files switch to configure. I kept tuplestore behavior
> and it still split file in both mode.
>
> I also little bit cleanup some other datatypes (e.g int->mode_t).
> Autoconf and autoheader must be run after patch application.
>
> I tested it with 9GB table and both mode works fine.
>
> Please, let me know your comments.
>
>         Zdenek

[ application/x-gzip is not supported, skipping... ]

>
> ---------------------------(end of broadcast)---------------------------
> TIP 3: Have you checked our extensive FAQ?
>
>                http://www.postgresql.org/docs/faq

--
  Bruce Momjian  <bruce@momjian.us>          http://momjian.us
  EnterpriseDB                               http://www.enterprisedb.com

  + If your life is a hard drive, Christ can be your backup. +

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Tom Lane
Date:
Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
> There is latest version of nonsegment support patch. I changed
> LET_OS_MANAGE_FILESIZE to USE_SEGMENTED_FILES and I added
> -disable-segmented-files switch to configure. I kept tuplestore behavior
> and it still split file in both mode.

Applied with minor corrections.

            regards, tom lane

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Peter Eisentraut
Date:
Tom Lane wrote:
> Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
> > There is latest version of nonsegment support patch. I changed
> > LET_OS_MANAGE_FILESIZE to USE_SEGMENTED_FILES and I added
> > -disable-segmented-files switch to configure. I kept tuplestore behavior
> > and it still split file in both mode.
>
> Applied with minor corrections.

Why is this not the default when supported?  I am wondering both from the
point of view of the user, and in terms of development direction.

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Tom Lane
Date:
Peter Eisentraut <peter_e@gmx.net> writes:
> Tom Lane wrote:
>> Applied with minor corrections.

> Why is this not the default when supported?

Fear.

Maybe eventually, but right now I think it's too risky.

One point that I already found out the hard way is that sizeof(off_t) = 8
does not guarantee the availability of largefile support; there can also
be filesystem-level constraints, and perhaps other things we know not of
at this point.

I think this needs to be treated as experimental until it's got a few
more than zero miles under its belt.  I wouldn't be too surprised to
find that we have to implement it as a run-time switch instead of
compile-time, in order to not fail miserably when somebody sticks a
tablespace on an archaic filesystem.

            regards, tom lane

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Alvaro Herrera
Date:
Peter Eisentraut wrote:
> Tom Lane wrote:
> > Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
> > > There is latest version of nonsegment support patch. I changed
> > > LET_OS_MANAGE_FILESIZE to USE_SEGMENTED_FILES and I added
> > > -disable-segmented-files switch to configure. I kept tuplestore behavior
> > > and it still split file in both mode.
> >
> > Applied with minor corrections.
>
> Why is this not the default when supported?  I am wondering both from the
> point of view of the user, and in terms of development direction.

Also it would get more buildfarm coverage if it were default.  If it
breaks something we'll notice earlier.

--
Alvaro Herrera                                http://www.CommandPrompt.com/
PostgreSQL Replication, Consulting, Custom Development, 24x7 support

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Tom Lane
Date:
Alvaro Herrera <alvherre@commandprompt.com> writes:
> Also it would get more buildfarm coverage if it were default.  If it
> breaks something we'll notice earlier.

Since nothing the regression tests do even approach 1GB, the odds that
the buildfarm will notice problems are approximately zero.

            regards, tom lane

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Peter Eisentraut
Date:
Tom Lane wrote:
> I think this needs to be treated as experimental until it's got a few
> more than zero miles under its belt.

OK, then maybe we should document that.

> I wouldn't be too surprised to
> find that we have to implement it as a run-time switch instead of
> compile-time, in order to not fail miserably when somebody sticks a
> tablespace on an archaic filesystem.

Yes, that sounds quite useful.  Let's wait and see what happens.

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
"Zeugswetter Andreas OSB SD"
Date:
> > Why is this not the default when supported?
>
> Fear.
>
> Maybe eventually, but right now I think it's too risky.
>
> One point that I already found out the hard way is that sizeof(off_t)
= 8
> does not guarantee the availability of largefile support; there can
also
> be filesystem-level constraints, and perhaps other things we know not
of
> at this point.

Exactly, e.g. AIX is one of those. jfs (not the newer jfs2) has an
option
to enable large files, which is not the default and cannot be changed
post crfs.
And even if it is enabled, jfs has a 64 Gb filesize limit !
Anybody know others that support large but not huge files ?

Andreas

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
"Zeugswetter Andreas OSB SD"
Date:
> > Why is this not the default when supported?  I am wondering both
from the
> > point of view of the user, and in terms of development direction.
>
> Also it would get more buildfarm coverage if it were default.  If it
> breaks something we'll notice earlier.

No we don't, because the buildfarm does not test huge files.

Andreas

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Larry Rosenman
Date:
On Mon, 10 Mar 2008, Tom Lane wrote:

> Peter Eisentraut <peter_e@gmx.net> writes:
>> Tom Lane wrote:
>>> Applied with minor corrections.
>
>> Why is this not the default when supported?
>
> Fear.
>
> Maybe eventually, but right now I think it's too risky.
>
> One point that I already found out the hard way is that sizeof(off_t) = 8
> does not guarantee the availability of largefile support; there can also
> be filesystem-level constraints, and perhaps other things we know not of
> at this point.
>
Just to note an additional filesystem that will need special action...
The VxFS filesystem has a largefiles option, per filesystem.  At least that
was the case on SCO UnixWare (No, I no longer run it).

LER


>
>             regards, tom lane
>
>

--
Larry Rosenman                     http://www.lerctr.org/~ler
Phone: +1 512-248-2683                 E-Mail: ler@lerctr.org
US Mail: 430 Valona Loop, Round Rock, TX 78681-3893

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Tom Lane
Date:
Peter Eisentraut <peter_e@gmx.net> writes:
> Tom Lane wrote:
>> I think this needs to be treated as experimental until it's got a few
>> more than zero miles under its belt.

> OK, then maybe we should document that.

Agreed, but at this point we don't even know what hazards we need to
document.

            regards, tom lane

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Tom Lane
Date:
"Zeugswetter Andreas OSB SD" <Andreas.Zeugswetter@s-itsolutions.at> writes:
> Exactly, e.g. AIX is one of those. jfs (not the newer jfs2) has an
> option
> to enable large files, which is not the default and cannot be changed
> post crfs.
> And even if it is enabled, jfs has a 64 Gb filesize limit !
> Anybody know others that support large but not huge files ?

Yeah, HPUX 10 is similar --- 128GB hard maximum.  It does say you
can convert an existing filesystem to largefile support, but it has
to be unmounted.

These examples suggest that maybe what we want is not so much a "no
segments ever" mode as a segment size larger than 1GB.

            regards, tom lane

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Zdenek Kotala
Date:
Tom Lane napsal(a):
> "Zeugswetter Andreas OSB SD" <Andreas.Zeugswetter@s-itsolutions.at> writes:
>> Exactly, e.g. AIX is one of those. jfs (not the newer jfs2) has an
>> option
>> to enable large files, which is not the default and cannot be changed
>> post crfs.
>> And even if it is enabled, jfs has a 64 Gb filesize limit !
>> Anybody know others that support large but not huge files ?
>
> Yeah, HPUX 10 is similar --- 128GB hard maximum.  It does say you
> can convert an existing filesystem to largefile support, but it has
> to be unmounted.
>
> These examples suggest that maybe what we want is not so much a "no
> segments ever" mode as a segment size larger than 1GB.

Patch allows to use bigger than 2/4GB segment files and it is possible
changed it in source file. But how it was mentioned in this thread maybe
somethink like this "CREATE TABLESPACE name LOCATION '/my/location'
SEGMENTS 10GB" should good solution. If segments is not mentioned then
default value is used.


            Zdenek

PS: ZFS is happy with 2^64bit size and UFS has 1TB file size limit
(depends on solaris version)

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Tom Lane
Date:
Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
> Tom Lane napsal(a):
>> These examples suggest that maybe what we want is not so much a "no
>> segments ever" mode as a segment size larger than 1GB.

> PS: ZFS is happy with 2^64bit size and UFS has 1TB file size limit
> (depends on solaris version)

So even on Solaris, "no segments ever" is actually a pretty awful idea.
As it stands, the code would fail on tables > 1TB.

I'm thinking we need to reconsider this patch.  Rather than disabling
segmentation altogether, we should see it as allowing use of segments
larger than 1GB.  I suggest that we ought to just flat rip out the "non
segmenting" code paths in md.c, and instead look into what segment sizes
are appropriate on different platforms.

            regards, tom lane

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Bruce Momjian
Date:
Tom Lane wrote:
> Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
> > Tom Lane napsal(a):
> >> These examples suggest that maybe what we want is not so much a "no
> >> segments ever" mode as a segment size larger than 1GB.
>
> > PS: ZFS is happy with 2^64bit size and UFS has 1TB file size limit
> > (depends on solaris version)
>
> So even on Solaris, "no segments ever" is actually a pretty awful idea.
> As it stands, the code would fail on tables > 1TB.
>
> I'm thinking we need to reconsider this patch.  Rather than disabling
> segmentation altogether, we should see it as allowing use of segments
> larger than 1GB.  I suggest that we ought to just flat rip out the "non
> segmenting" code paths in md.c, and instead look into what segment sizes
> are appropriate on different platforms.

Agreed.

--
  Bruce Momjian  <bruce@momjian.us>        http://momjian.us
  EnterpriseDB                             http://postgres.enterprisedb.com

  + If your life is a hard drive, Christ can be your backup. +

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Zdenek Kotala
Date:
Tom Lane napsal(a):
> Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
>> Tom Lane napsal(a):
>>> These examples suggest that maybe what we want is not so much a "no
>>> segments ever" mode as a segment size larger than 1GB.
>
>> PS: ZFS is happy with 2^64bit size and UFS has 1TB file size limit
>> (depends on solaris version)
>
> So even on Solaris, "no segments ever" is actually a pretty awful idea.
> As it stands, the code would fail on tables > 1TB.
>
> I'm thinking we need to reconsider this patch.  Rather than disabling
> segmentation altogether, we should see it as allowing use of segments
> larger than 1GB.  I suggest that we ought to just flat rip out the "non
> segmenting" code paths in md.c, and instead look into what segment sizes
> are appropriate on different platforms.

Yes, agree. It seems only ZFS is OK at this moment and if somebody sets
32TB he gets nonsegment mode anyway. I looked into posix standard and
there is useful function which can be used. See

http://www.opengroup.org/onlinepubs/009695399/functions/pathconf.html

Maybe we can put additional test into configure and collect appropriate
data from buildfarm.

I think current patch could stay in CVS and I will rip out non segment
code path in a new patch.

        Zdenek

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Tom Lane
Date:
Zdenek Kotala <Zdenek.Kotala@Sun.COM> writes:
> I think current patch could stay in CVS and I will rip out non segment
> code path in a new patch.

Sure, I feel no need to revert what's applied.  Have at it.

            regards, tom lane

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Peter Eisentraut
Date:
Zdenek Kotala wrote:
> Yes, agree. It seems only ZFS is OK at this moment and if somebody sets
> 32TB he gets nonsegment mode anyway.

Surely if you set the segment size to INT64_MAX, you will get nonsegmented
behavior anyway, so two code paths might not be necessary at all.

> I looked into posix standard and
> there is useful function which can be used. See
>
> http://www.opengroup.org/onlinepubs/009695399/functions/pathconf.html
>
> Maybe we can put additional test into configure and collect appropriate
> data from buildfarm.

It might be good to just check first if it returns realistic values for the
example cases that have been mentioned.

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Tom Lane
Date:
Peter Eisentraut <peter_e@gmx.net> writes:
> Zdenek Kotala wrote:
>> Maybe we can put additional test into configure and collect appropriate
>> data from buildfarm.

> It might be good to just check first if it returns realistic values for the
> example cases that have been mentioned.

Yeah, please just make up a ten-line C program that prints the numbers
you want, and post it on -hackers for people to try.  If manual testing
says that it's printing useful numbers, then it would be time enough to
think about how to get it into the buildfarm.

            regards, tom lane

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Peter Eisentraut
Date:
Zdenek Kotala wrote:
> But how it was mentioned in this thread maybe
> somethink like this "CREATE TABLESPACE name LOCATION '/my/location'
> SEGMENTS 10GB" should good solution. If segments is not mentioned then
> default value is used.

I think you would need a tool to resegmentize a table or tablespace offline,
usable for example when recovering a backup.

Also, tablespace configuration information is of course also stored in a
table.  pg_tablespace probably won't become large, but it would probably
still need to be special-cased, along with other system catalogs perhaps.

An then, how to coordindate offline resegmenting and online tablespace
operations in a crash-safe way?

Another factor I just thought of is that tar, commonly used as part of a
backup procedure, can on some systems only handle files up to 8 GB in size.
There are supposed to be newer formats that can avoid that restriction, but
it's not clear how widely available these are and what the incantation is to
get at them.  Of course we don't use tar directly, but if we ever make large
segments the default, we ought to provide some clear advice for the user on
how to make their backups.

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Martijn van Oosterhout
Date:
On Wed, Mar 19, 2008 at 09:38:12AM +0100, Peter Eisentraut wrote:
> Another factor I just thought of is that tar, commonly used as part of a
> backup procedure, can on some systems only handle files up to 8 GB in size.
> There are supposed to be newer formats that can avoid that restriction, but
> it's not clear how widely available these are and what the incantation is to
> get at them.  Of course we don't use tar directly, but if we ever make large
> segments the default, we ought to provide some clear advice for the user on
> how to make their backups.

By my reading, GNU tar handles larger files and no-one else (not even a
POSIX standard tar) can...

Have a nice day,
--
Martijn van Oosterhout   <kleptog@svana.org>   http://svana.org/kleptog/
> Please line up in a tree and maintain the heap invariant while
> boarding. Thank you for flying nlogn airlines.

Attachment

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Kenneth Marshall
Date:
On Wed, Mar 19, 2008 at 10:51:12AM +0100, Martijn van Oosterhout wrote:
> On Wed, Mar 19, 2008 at 09:38:12AM +0100, Peter Eisentraut wrote:
> > Another factor I just thought of is that tar, commonly used as part of a
> > backup procedure, can on some systems only handle files up to 8 GB in size.
> > There are supposed to be newer formats that can avoid that restriction, but
> > it's not clear how widely available these are and what the incantation is to
> > get at them.  Of course we don't use tar directly, but if we ever make large
> > segments the default, we ought to provide some clear advice for the user on
> > how to make their backups.
>
> By my reading, GNU tar handles larger files and no-one else (not even a
> POSIX standard tar) can...
>
> Have a nice day,
> --
> Martijn van Oosterhout   <kleptog@svana.org>   http://svana.org/kleptog/
> > Please line up in a tree and maintain the heap invariant while
> > boarding. Thank you for flying nlogn airlines.

The star program written by Joerg Schilling is a very well written
POSIX compatible tar program that can easily handle files larger than
8GB. It is another backup option.

Cheers,
Ken

Re: [HACKERS] Fix for large file support (nonsegment mode support)

From
Zdeněk Kotala
Date:
Peter Eisentraut napsal(a):
> Zdenek Kotala wrote:
>> But how it was mentioned in this thread maybe
>> somethink like this "CREATE TABLESPACE name LOCATION '/my/location'
>> SEGMENTS 10GB" should good solution. If segments is not mentioned then
>> default value is used.
>
> I think you would need a tool to resegmentize a table or tablespace offline,
> usable for example when recovering a backup.

Do you mean something like strip(1) command? I don't see any usecase for
terrabytes data. You usually have a problem to find place where you can backup.

> Also, tablespace configuration information is of course also stored in a
> table.  pg_tablespace probably won't become large, but it would probably
> still need to be special-cased, along with other system catalogs perhaps.

It is true and unfortunately singularity. Same as database list which is in a
table as well, but it is stored also as a text file for startup purpose. I more
incline to use non table configuration file for tablespaces, because I don't see
any advantage to have it under MVCC control and it allow also to define storage
for pg_global and pg_default.

> An then, how to coordindate offline resegmenting and online tablespace
> operations in a crash-safe way?
>
> Another factor I just thought of is that tar, commonly used as part of a
> backup procedure, can on some systems only handle files up to 8 GB in size.
> There are supposed to be newer formats that can avoid that restriction, but
> it's not clear how widely available these are and what the incantation is to
> get at them.  Of course we don't use tar directly, but if we ever make large
> segments the default, we ought to provide some clear advice for the user on
> how to make their backups.

I think tar is OK - minimal on Solaris. See man largefile.

Default segment size still should be 1GB. If DBA makes a decision to increase
this to higher value, then it is his responsibility to find way how to process
this big files.

        Zdenek