|
IDS Forum
Re: Performance problem during log backup
Posted By: Art Kagel Date: Thursday, 28 August 2014, at 9:03 a.m.
In Response To: Re: Performance problem during log backup (Marcus Haarmann)
Marcus: All of the BLOB/CLOB records in a table do not have to go to the
same sbspace! You could create multiple sbspaces and have your
applications place the individual documents into multiple sbspaces using
some schema of your own just like fragmentation. That will eliminate the
problem you ran into originally!
Art
Art S. Kagel, Principal Consultant
ASK Database Management
Blog: http://informix-myview.blogspot.com/
Disclaimer: Please keep in mind that my own opinions are my own opinions
and do not reflect on the IIUG, nor any other organization with which I am
associated either explicitly, implicitly, or by inference. Neither do
those opinions reflect those of other individuals affiliated with any
entity with which I am affiliated nor those of the entities themselves.
On Thu, Aug 28, 2014 at 8:43 AM, Marcus Haarmann <marcus.haarmann@midoco.de>
wrote:
> Fernando,
>
> Search for " informix.LO_hdr_partn" in IIUG and you will find my post
> there.
> We opened a Bug with IBM through our distributor, sent tons of data
> (oncheck
> took a day to produce 200GB output),
> but no answer.
> The initial setup did run for two years and suddenly we got SQL errors -136
> (no more extents).
> The only place to see extent problems was in meta data of SBlob, which
> reached
> the maximum number of pages
> for some chunks.
> We reproduced the error situation and generated an af file with the
> stacktrace, initiated by detection of the -136 SQL error.
> Instance was alive all the time, not crashing, but inserts were not
> possible
> any more.
> We suspect that the error was occuring because some single documents had
> been
> deleted before,
> which might have been located in a full chunk. The resulting free space was
> about to be re-allocated,
> but the meta data could not be extended since maximum number of pages was
> reached, which resulted in the error.
> We have also seen, that deleting data will not free pages in the meta data
> partition (at least onstat -d does not show that).
> But nobody could give us a clue what we had done wrong (to make sure we
> would
> not run in the same situation
> when setting up a new instance). So we decided to go "back to the roots"
> and
> use bytes instead of blob.
> Now we are stuck again ...
>
> Marcus Haarmann
> Geschäftsführer
>
> Midoco GmbH
> Otto-Hahn-Str. 12
> 40721 Hilden
>
> Tel. +49 (2103) 28 74 0
> Fax. +49 (2103) 28 74 28
> www.midoco.de
>
> Member of Pisano Holding GmbH
> Better Travel Technology
> www.pisano-holding.com
>
> Amtsgericht Düsseldorf - HRB 51420 - USt.ID DE814319276 -
> Geschäftsführer: Steffen Faradi, Marcus Haarmann, Jörg Hauschild -
>
> ----- Ursprüngliche Mail -----
>
> Von: "Fernando Nunes" <domusonline@gmail.com>
> An: ids@iiug.org
> Gesendet: Donnerstag, 28. August 2014 14:19:50
> Betreff: Re: Performance problem during log backup [33629]
>
> I'd go back to the beginning.... An error inserting data... from what I
> understand it caused an AF.
> What did IBM have to say about that? You mean the instance was crashing
> frequently during INSERTs?
> I find it hard to believe that you go through so much trouble to solve a
> bug (an AF is almost certainly a bug unless there are hardware errors or
> severe administration errors (like messing with storage etc.)
>
> Regards.
>
> On Thu, Aug 28, 2014 at 2:08 PM, Art Kagel <art.kagel@gmail.com> wrote:
>
> > Hmm, using JSON/BSON in Informix MIGHT be a work around. If you put the
> > document data into a BSON column in a relational table (so v12.10
> > preferably .xC4 or later for ease of use) the engine will shift the BSON
> > column to Smartblob space automagically. Because the engine is the only
> > one dealing with the smartblob space maybe it will work without getting
> the
> > errors you were getting with BLOB columns. JSON/BSON in Informix can be
> > accessed via SQL and the Informix client protocols or via MongoDB's wire
> > listener protocol (so any development tool that can access MongoDB). Just
> > a thought.
> >
> > Note that in MongoDB JSON documents are limited to 16MB, JSON columns in
> > Informix can be up to 4GB each.
> >
> > Art
> >
> > Art S. Kagel, Principal Consultant
> > ASK Database Management
> >
> > Blog: http://informix-myview.blogspot.com/
> >
> > Disclaimer: Please keep in mind that my own opinions are my own opinions
> > and do not reflect on the IIUG, nor any other organization with which I
> am
> > associated either explicitly, implicitly, or by inference. Neither do
> > those opinions reflect those of other individuals affiliated with any
> > entity with which I am affiliated nor those of the entities themselves.
> >
> > On Thu, Aug 28, 2014 at 7:46 AM, Marcus Haarmann <
> > marcus.haarmann@midoco.de>
> > wrote:
> >
> > > Hi Art,
> > >
> > > Sounds good, but we run into other problems:
> > > 1) Maximum number of pages will be reached too soon depending on page
> > size
> > > (2k
> > > ~ 2mio rows)
> > > We want to store > 200mio rows
> > > 2) Fragmentation cannot be used since this is growth edition. We would
> > > have to
> > > buy an Ultimate just
> > > for the document stuff.
> > >
> > > So, SBlob not possible, Blob not possible, Blob in TBLSpace not
> possible.
> > > Other ideas to store such an amount of binary data within the DB ?
> > >
> > > A more complex solution could be a cluster filesystem and just
> > remembering
> > > the
> > > file path
> > > in the db, which would obviously not create any problem ...
> > >
> > > I think we have to evaluate other products because of the limitations
> in
> > > blob
> > > storage.
> > > Maybe we will investigate in document oriented databases like Mongo or
> > > similar.
> > > But the conversion to a non-SQL DB is complex and application will
> need a
> > > significant
> > > change.
> > >
> > > Marcus Haarmann
> > >
> > > ----- Ursprüngliche Mail -----
> > >
> > > Von: "Art Kagel" <art.kagel@gmail.com>
> > > An: ids@iiug.org
> > > Gesendet: Mittwoch, 27. August 2014 23:10:25
> > > Betreff: Re: Performance problem during log backup [33606]
> > >
> > > Marcus:
> > >
> > > The fix is to move the BYTE and TEXT columns from a blobspace to IN
> TABLE
> > > and place the tables into a dbspace with a pagesize that is appropriate
> > to
> > > the average blob sizes so little space is wasted. Then the logs will
> > > contain the blob pages, you will be able to use hdr, and the logs will
> > back
> > > up faster because the backup will not have to go find the blob pages.
> You
> > > will have to increase the logical log file sizes and the number of logs
> > > because the blobs will now be logged, but that's a small price.
> > >
> > > Art
> > >
> > > Art S. Kagel, Principal Consultant
> > > ASK Database Management
> > >
> > > Blog: http://informix-myview.blogspot.com/
> > >
> > > Disclaimer: Please keep in mind that my own opinions are my own
> opinions
> > > and do not reflect on the IIUG, nor any other organization with which I
> > am
> > > associated either explicitly, implicitly, or by inference. Neither do
> > > those opinions reflect those of other individuals affiliated with any
> > > entity with which I am affiliated nor those of the entities themselves.
> > >
> > > On Wed, Aug 27, 2014 at 4:41 PM, Marcus Haarmann <
> > > marcus.haarmann@midoco.de>
> > > wrote:
> > >
> > > > Hi experts,
> > > >
> > > > we were forced to create a new instance for our document storage
> > > > (originally
> > > > set up as table with blob column,
> > > > using a smart blob storage). The new instance is using byte columns
> > (yes
> > > I
> > > > know we cannot use HDR ...).
> > > >
> > > > The old instance was stuck because of an undefined error on inserting
> > new
> > > > data
> > > > and nobody was able to solve
> > > > the situation.
> > > > The error was related to the meta data, looking at the af file.
> > > > Now, we converted everything to byte content, cause we do not really
> > > need a
> > > > smartblob functionality
> > > > (other than HDR, we are using mirroring now to cope with hardware
> > > errors).
> > > > Took a while
> > > > to copy the data from blob to byte.
> > > > When the instance was copied, we had logging turned off, to copy the
> > data
> > > > with
> > > > higher throughput.
> > > > The data was copied with a JAVA program, which inserted the records
> > with
> > > 10
> > > > parallel threads on the
> > > > target machine (since no internal cast from blob to byte is
> available).
> > > > Now, the machine has taken over the load and we have switched on
> > logging.
> > > >
> > > > The new situation is that whenever a logfile gets full (like every 15
> > > > minutes), the backup (ontape) starts
> > > > writing the header of the logfile and then stops. About 20 minutes
> > later
> > > > the
> > > > logfile will be written.
> > > > Then the next logfile is already full and backup starts again.
> > > > Normally, on our other instances, writing logfiles is a process
> running
> > > for
> > > > some seconds, not several minutes.
> > > > Inserting new data in this state is veeery slow, can take minutes for
> > one
> > > > row.
> > > >
> > > > We have now turned off log files for the moment, and added a separate
> > > > instance
> > > > to collect the traffic
> > > > because the instance is basically unusable in this state.
> > > >
> > > > Also, the smaller instance with comparable setup (same table, also
> byte
> > > col
> > > > for document content),
> > > > tends to have the same behaviour (though we can write a logfile there
> > in
> > > > about
> > > > 3-4 minutes at the moment).
> > > >
> > > > The hardware is not very fast, cause normally the documents (mostly
> > > > compressed
> > > > pdf) should be written once,
> > > > never changed and are retrieved again by primary key on demand
> (number
> > of
> > > > written documents 100 compared to 10
> > > > read documents).
> > > > Initially we had the instance running with smart blobs, and memory
> > usage
> > > > was 500mb. Storage size with smart blobs was around 2.5TB.
> > > > We ran into several problems:
> > > > - Backup Level 1 blocked instance for 1.5h at night (fixed bug in
> > > FC8W...).
> > > > - finally SQL Error on insert of new rows, an unclear situation which
> > IBM
> > > > was
> > > > not able to reproduce by now.
> > > > We changed the application to be able to address two instances.
> > > > Since there was no hint from IBM how to setup the database to not run
> > > into
> > > > the
> > > > situation again,
> > > > we decided to use "old" byte columns instead of blob columns.
> > > >
> > > > The setup now is as follows:
> > > > IDS 11.70FC7GE on Linux, Double processor Opteron, FC 8GBit to SAN,
> Raw
> > > > Devices on lvm partitions
> > > > Three tables in the database, biggest one around 200mio rows. In
> total
> > > > around
> > > > 350mio records.
> > > > Chunk layout: 1x rootdbs 100GB (for table), 6x blobdbs 500GB (for
> > bytes).
> > > > Allocated memory 5.5GB (SHM 256mb, add 64mb, buffers 80000x2k)
> > > >
> > > > What we found out so far:
> > > > Written logfiles are significantly bigger than logsize (probably
> > because
> > > > the
> > > > byte content is not in the logs
> > > > and is read separately).
> > > > While logfiles are backed up, we encounter a extreme read activity on
> > the
> > > > blob
> > > > chunks (onstat -g iof).
> > > > Ok, this is not a very common setup for a relational database, we are
> > > > currently evaluating other databases
> > > > which might do a better job on this amount of documents.
> > > > I do not like that personally, since our administration is used to do
> > > > basically everything with Informix.
> > > > But the question is:
> > > > What is generally wrong with the setup ?
> > > > Why is logfile writing taking that long ?
> > > > Why does the big instance block without any error message when
> > inserting
> > > > new
> > > > rows ?
> > > > What can we tune to make it work ?
> > > >
> > > > Maybe somebody can give us a hint. The intended setup for us would
> be a
> > > HDR
> > > > instance pair,
> > > > using Smart Blobs, Level 0 Backup on sundays, Level 1 backup +
> > Logbackup
> > > > during the week.
> > > > Since there are only a couple of tables involved, maybe an ER setup
> > with
> > > > bytes
> > > > could be a solution,
> > > > but we are sort of desperate because each try to store such an amount
> > of
> > > > documents in an Informix
> > > > instance was not very successful so far.
> > > >
> > > > Thanks in advance for your input.
> > > >
> > > > Marcus Haarmann
> > > >
> > > >
> > > >
> > > >
> > >
> > >
> > >
> >
> >
>
>
> *******************************************************************************
> > > > Forum Note: Use "Reply" to post a response in the discussion forum.
> > > >
> > > >
> > >
> > > --001a11c34b18f4e3870501a2d8d6
> > >
> > >
> > >
> > >
> >
> >
>
>
> *******************************************************************************
> > > Forum Note: Use "Reply" to post a response in the discussion forum.
> > >
> > >
> > >
> > >
> >
> >
>
>
> *******************************************************************************
> > > Forum Note: Use "Reply" to post a response in the discussion forum.
> > >
> > >
> >
> > --001a11c34b18e6a7de0501af6229
> >
> >
> >
> >
>
>
> *******************************************************************************
> > Forum Note: Use "Reply" to post a response in the discussion forum.
> >
> >
>
> --
> Fernando Nunes
> Portugal
>
> http://informix-technology.blogspot.com
> My email works... but I don't check it frequently...
>
> --001a11c2d676c17dc30501af8ca0
>
>
>
> *******************************************************************************
> Forum Note: Use "Reply" to post a response in the discussion forum.
>
>
>
> *******************************************************************************
> Forum Note: Use "Reply" to post a response in the discussion forum.
>
>
--089e0160a638dc9f620501b025ad
Messages In This Thread
IDS Forum is maintained by Administrator with WebBBS 5.12.
|
|