db-derby-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Varma R (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (DERBY-5234) Unable to insert data into table. Failed due be "ERROR XSDG0: Page Page(51919,Container(0, 1104)) could not be read from disk."
Date Tue, 17 May 2011 10:19:47 GMT

    [ https://issues.apache.org/jira/browse/DERBY-5234?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13034679#comment-13034679
] 

Varma R commented on DERBY-5234:
--------------------------------

Hi,

Tried the above options mentioned.
a) Given the below debug option which starting the derby database.
-Dderby.debug.true=CacheTrace,DaemonTrace,DeadLockTrace,LockStackTrace,LockTrace,LogTrace,memoryLeakTrace,ScanTrace,SerializedTrace,SpaceTrace,TEST_LOG_FULL

But this didn't print any additional information in the logs.

b) Below are the command output's on test environment. 
-bash-3.00$ ulimit -a | grep file
core file size        (blocks, -c) unlimited
file size             (blocks, -f) unlimited
open files                    (-n) 256
-bash-3.00$ plimit 5036
5036:   java -Dderby.system.home=/export/home/cmd -Dderby.locks.monitor=true -
   resource              current         maximum
  time(seconds)         unlimited       unlimited
  file(blocks)          unlimited       unlimited
  data(kbytes)          unlimited       unlimited
  stack(kbytes)         10240           unlimited
  coredump(blocks)      unlimited       unlimited
  nofiles(descriptors)  65536           65536
  vmemory(kbytes)       unlimited       unlimited

you can see that there are no limits on the file system and 65536 is the hard limit for the
file descriptors.

c) After running SYSCS_UTIL.SYSCS_COMPRESS_TABLE procedure on KPI.KPI_MERGEIN table. INSERT
command is working on KPI.KPI_MERGEIN table.

d) How to find out the list of .dat files corresponding to a table in seg0 directory?
 
This same application was load tested for higher transcation volume and no such error were
reported during development. However the problem in production seems to be pointing to cumulative
data/disk usage. Also would it be recommeneded to run SYSCS_UTIL.SYSCS_COMPRESS_TABLE once
in a while?

> Unable to insert data into table. Failed due be "ERROR XSDG0: Page Page(51919,Container(0,
1104)) could not be read from disk."
> -------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: DERBY-5234
>                 URL: https://issues.apache.org/jira/browse/DERBY-5234
>             Project: Derby
>          Issue Type: Bug
>          Components: JDBC, Network Server, SQL
>    Affects Versions: 10.5.3.0
>         Environment: HP-UX 11iv2 in production environment with JDK1.6; Solaris 5/10
in test environment with JDK 1.6
>            Reporter: Varma R
>              Labels: ERROR, XSDG0, apache, corruption, data, derby
>
> One of the derby database table "gets corrupted"/"indicates connection not available"
during processing inserts from java client application as shown in the trace and the only
way to recover from this error is to rebuild the DB - by deleting the data and creating the
tables again. This happens once in a while (thrice in a span of two months) and the java application
(run in multiple servers), which updates the database, processes around 100 million transactions
per hour (in total and each transation results in 4-5 updates to the DB) 
> There are eight tables in the derby database.
>    TABLE NAME                           ROWS COUNT (at time of corruption)
> ---------------------------------------------------------------------------------
>    KPI.KPI_MERGEIN;                     362917
>    KPI.KPI_IN;                                 422508
>    KPI.KPI_DROPPED;                    53667
>    KPI.KPI_ERROR1;                       0
>    KPI.KPI_ERROR2;                       2686
>    KPI.KPI_ERRORMERGE;            0
>    KPI.KPI_MERGEOUT;                 362669
>    KPI.KPI_OUT;                             125873
> The derby database has been started with the following parameters 
> CMD="java -Dderby.system.home=$DERBY_OPTS -Dderby.locks.monitor=true -Dderby.locks.deadlockTrace=true
-Dderby.locks.escalationThreshold=50000 -Dderby.locks.waitTimeout=
> -1 -Dderby.storage.pageCacheSize=100000 -Xms512M -Xmx3072M -XX:NewSize=256M -classpath
$DERBY_CLASSPATH org.apache.derby.drda.NetworkServerControl start -h $KPIDERBYHOST -p $DERBY_KPI_PORT"
> The corrupted database tar (filesystem) in live environment was moved to a test system
(Solaris system) and few checks were run on the corrupted DB as part of analysis (DB does
start fine)
> While trying to insert a row in any table expect KPI.KPI_MERGEIN, it is successful. But
when a new row is inserted into KPI.KPI_MERGEIN table using command line tool it's throwing
below error message (the same message that appeared in live 
> ij> INSERT INTO KPI.KPI_MERGEIN (A0_TXN_ID, A1_NE_ID, A2_CHU_IP_ADDR, A3_BATCH_DATE,A5_CODE)
VALUES (-1, 'BMTDE', '192.2.1.3', 231456879, 'KSD');
> ERROR 08006: A network protocol error was encountered and the connection has been terminated:
the requested command encountered an unarchitected and implementation-specific condition for
which there was no architected message
> and in derby.log file it shows below error stacktrace.
> ERROR XSDG0: Page Page(51919,Container(0, 1104)) could not be read from disk.
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.CachedPage.readPage(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.CachedPage.setIdentity(Unknown Source)
>         at org.apache.derby.impl.services.cache.ConcurrentCache.find(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.FileContainer.initPage(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.FileContainer.newPage(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.BaseContainer.addPage(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.BaseContainerHandle.addPage(Unknown Source)
>         at org.apache.derby.impl.store.access.heap.HeapController.doInsert(Unknown Source)
>         at org.apache.derby.impl.store.access.heap.HeapController.insertAndFetchLocation(Unknown
Source)
>         at org.apache.derby.impl.sql.execute.RowChangerImpl.insertRow(Unknown Source)
>         at org.apache.derby.impl.sql.execute.InsertResultSet.normalInsertCore(Unknown
Source)
>         at org.apache.derby.impl.sql.execute.InsertResultSet.open(Unknown Source)
>         at org.apache.derby.impl.sql.GenericPreparedStatement.executeStmt(Unknown Source)
>         at org.apache.derby.impl.sql.GenericPreparedStatement.execute(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedStatement.executeStatement(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedStatement.execute(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedStatement.executeUpdate(Unknown Source)
>         at org.apache.derby.impl.drda.DRDAConnThread.parseEXCSQLIMM(Unknown Source)
>         at org.apache.derby.impl.drda.DRDAConnThread.processCommands(Unknown Source)
>         at org.apache.derby.impl.drda.DRDAConnThread.run(Unknown Source)
> Caused by: java.io.EOFException: Reached end of file while attempting to read a whole
page.
>         at org.apache.derby.impl.store.raw.data.RAFContainer4.readFull(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.RAFContainer4.readPage0(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.RAFContainer4.readPage(Unknown Source)
>         ... 20 more
> ============= begin nested exception, level (1) ===========
> java.io.EOFException: Reached end of file while attempting to read a whole page.
>         at org.apache.derby.impl.store.raw.data.RAFContainer4.readFull(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.RAFContainer4.readPage0(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.RAFContainer4.readPage(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.CachedPage.readPage(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.CachedPage.setIdentity(Unknown Source)
>         at org.apache.derby.impl.services.cache.ConcurrentCache.find(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.FileContainer.initPage(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.FileContainer.newPage(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.BaseContainer.addPage(Unknown Source)
>         at org.apache.derby.impl.store.raw.data.BaseContainerHandle.addPage(Unknown Source)
>         at org.apache.derby.impl.store.access.heap.HeapController.doInsert(Unknown Source)
>         at org.apache.derby.impl.store.access.heap.HeapController.insertAndFetchLocation(Unknown
Source)
>         at org.apache.derby.impl.sql.execute.RowChangerImpl.insertRow(Unknown Source)
>         at org.apache.derby.impl.sql.execute.InsertResultSet.normalInsertCore(Unknown
Source)
>         at org.apache.derby.impl.sql.execute.InsertResultSet.open(Unknown Source)
>         at org.apache.derby.impl.sql.GenericPreparedStatement.executeStmt(Unknown Source)
>         at org.apache.derby.impl.sql.GenericPreparedStatement.execute(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedStatement.executeStatement(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedStatement.execute(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedStatement.executeUpdate(Unknown Source)
>         at org.apache.derby.impl.drda.DRDAConnThread.parseEXCSQLIMM(Unknown Source)
>         at org.apache.derby.impl.drda.DRDAConnThread.processCommands(Unknown Source)
>         at org.apache.derby.impl.drda.DRDAConnThread.run(Unknown Source)
> ============= end nested exception, level (1) ===========
> 2011-05-16 10:37:21.392 GMT:
> Shutting down instance a816c00e-012f-f85f-7892-ffff874c3ff6
> ----------------------------------------------------------------
> Cleanup action completed
> The problem is only with INSERT statement. When i try SELECT statement on KPI.KPI_MERGEIN
table it is working well.The database file system size (in seg0) is 1.3 GB
> Can anyone help me out in identifying the problem that why for one table alone its throwing
the above error message ? Would upgrade to a new version help ? 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Mime
View raw message