hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Stack <st...@duboce.net>
Subject Re: Support for MultiGet / SQL In clause
Date Sun, 24 Jan 2010 19:37:51 GMT
Are they recent?  Seems like you have a mismatch in versions between
what wrote the data and what is reading it?  For sure you installed
the same hadoop everywhere?  Do the following exceptions come up into
hbase?  Is hbase running ok?

St.Ack

On Sat, Jan 23, 2010 at 11:35 PM, Sriram Muthuswamy Chittathoor
<sriramc@ivycomptech.com> wrote:
> What do errors like this in my  datanode mean (comes in all the boxes) -
> - something is corrupted and I clean everything and restart
>
> 2010-01-24 02:27:46,349 ERROR
> org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException:
> Incompatible namespaceIDs in /home/ppoker/test/hadoop-0.20.1/data/data:
> namenode namespaceID = 1033461714; datanode namespaceID = 1483221888
>        at
> org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStor
> age.java:233)
>        at
> org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead
> (DataStorage.java:148)
>        at
> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.j
> ava:298)
>        at
> org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:216
> )
>        at
> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.ja
> va:1283)
>        at
> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(Data
> Node.java:1238)
>        at
> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.
> java:1246)
>        at
> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1368)
>
> 2010-01-24 02:27:46,349 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
>
>
>
>
>
> -----Original Message-----
> From: Daniel Washusen [mailto:dan@reactive.org]
> Sent: Sunday, January 24, 2010 12:41 PM
> To: hbase-user@hadoop.apache.org
> Subject: Re: Support for MultiGet / SQL In clause
>
> Also, I would suggest that you allocate one of the machines to be
> hmaster & namenode. The other three machines run rserver, datanode &
> zookeeper.
>
> On 24/01/2010, at 5:36 PM, "Sriram Muthuswamy Chittathoor"
> <sriramc@ivycomptech.com
>  > wrote:
>
>> You were right. This is what I have in the regionserver.out file
>>
>> java.lang.OutOfMemoryError: Java heap space
>> Dumping heap to java_pid28602.hprof ...
>> Heap dump file created [1081182427 bytes in 53.046 secs]
>>
>> Do I increase the jvm heapsize for all the machines.  This is my
>> config
>>
>> 1.  4 boxes running the cluster
>> 2.  Habse Regionserver and HDFS data node on all the boxes
>> 3.  One of the boxes where this error occurred also has the NameNode
>> and
>> the HBasemaster running
>>
>> How do I selectively increase the heapsize just for regionserver.  I
>> changed  in bin/hbase.
>>
>> -----Original Message-----
>> From: Daniel Washusen [mailto:dan@reactive.org]
>> Sent: Sunday, January 24, 2010 3:44 AM
>> To: hbase-user@hadoop.apache.org
>> Subject: Re: Support for MultiGet / SQL In clause
>>
>> What does the regionserver.out log file say?  Maybe you are out of
>> memory...  The master web ui will report how much heap space the
>> region server is using...
>>
>> On 24/01/2010, at 8:43 AM, "Sriram Muthuswamy Chittathoor"
>> <sriramc@ivycomptech.com
>>> wrote:
>>
>>> I created a new table with indexes.  Initially created 100000 rows
>>> and then did a scan.  At that time it was okay.  Then I started
>>> creating a million rows in a loop and then after some time I get
>>> this exception and the table disappeared (even from the hbase
>>> shell).  One other table also disappeared.
>>>
>>> This is very consistent.  I tried a few times and every time it is
>>> the same on creating a lot of rows.  Rows are not too big (Just some
>>> 6 columns in one family) each of say type long or string.  Created 3
>>> indexes -- 2 byte array and 1 long.
>>>
>>>
>>>
>>> Cur Rows : 349999
>>> Cur : 359999
>>> Cur : 369999
>>> Cur : 379999
>>> Cur Rows : 389999   <--  Crashed after this
>>> Exception in thread "main"
>>> org.apache.hadoop.hbase.client.RetriesExhaustedException: Trying to
>>> contact region server 10.1.162.25:60020 for region
>> POKERHH6,,1264279773695
>>> , row '0000392413', but failed after 10 attempts.
>>> Exceptions:
>>> java.io.IOException: Call to /10.1.162.25:60020 failed on local
>>> exception: java.io.EOFException
>>> org.apache.hadoop.hbase.TableNotFoundException: POKERHH6
>>> org.apache.hadoop.hbase.TableNotFoundException: POKERHH6
>>> org.apache.hadoop.hbase.TableNotFoundException: POKERHH6
>>> org.apache.hadoop.hbase.TableNotFoundException: POKERHH6
>>> org.apache.hadoop.hbase.TableNotFoundException: POKERHH6
>>> org.apache.hadoop.hbase.TableNotFoundException: POKERHH6
>>> org.apache.hadoop.hbase.TableNotFoundException: POKERHH6
>>> org.apache.hadoop.hbase.TableNotFoundException: POKERHH6
>>> org.apache.hadoop.hbase.TableNotFoundException: POKERHH6
>>>
>>>       at org.apache.hadoop.hbase.client.HConnectionManager
>>> $TableServers.getRegionServerWithRetries(HConnectionManager.java:
>>> 1048)
>>>       at org.apache.hadoop.hbase.client.HConnectionManager
>>> $TableServers$3.doCall(HConnectionManager.java:1239)
>>>       at org.apache.hadoop.hbase.client.HConnectionManager
>>> $TableServers$Batch.process(HConnectionManager.java:1161)
>>>       at org.apache.hadoop.hbase.client.HConnectionManager
>>> $TableServers.processBatchOfRows(HConnectionManager.java:1247)
>>>       at org.apache.hadoop.hbase.client.HTable.flushCommits
>>> (HTable.java:609)
>>>       at org.apache.hadoop.hbase.client.HTable.put(HTable.java:474)
>>>       at test.TestExtIndexedTable.main(TestExtIndexedTable.java:110)
>>> 10/01/23 16:21:28 INFO zookeeper.ZooKeeper: Closing session:
>>> 0x265caacc7a001b
>>> 10/01/23 16:21:28 INFO zookeeper.ClientCnxn: Closing ClientCnxn for
>>> session: 0x265caacc7a001b
>>> 10/01/23 16:21:28 INFO zookeeper.ClientCnxn: Exception while closing
>>> send thread for session 0x265caacc7a001b : Read error rc = -1
>>> java.nio.DirectByteBuffer[pos=0 lim=4 cap=4]
>>>
>>>
>>>
>>> -----Original Message-----
>>> From: saint.ack@gmail.com [mailto:saint.ack@gmail.com] On Behalf Of
>>> Stack
>>> Sent: Sunday, January 24, 2010 1:33 AM
>>> To: hbase-user@hadoop.apache.org
>>> Subject: Re: Support for MultiGet / SQL In clause -- error in patch
>>> HBASE-1845
>>>
>>> On Sat, Jan 23, 2010 at 2:52 AM, Sriram Muthuswamy Chittathoor
>>> <sriramc@ivycomptech.com> wrote:
>>>> Thanks all.  I messed it up when I was trying to upgrade to
>>>> 0.20.3.  I deleted the data directory and formatted it thinking it
>>>> will reset the whole cluster.
>>>>
>>>> I started fresh by deleting the data directory on all the nodes and
>>>> then everything worked.  I was also able to create the indexed
>>>> table using the 0.20.3 patch.  Let me run some tests on a few
>>>> million rows and see how it holds up.
>>>>
>>>> BTW --  what would be the right way when I moved versions.  Do I
>>>> run migrate scripts to migrate the data to newer versions ?
>>>>
>>> Just install the new binaries every and restart or perform a rolling
>>> restart -- see http://wiki.apache.org/hadoop/Hbase/RollingRestart --
>>> if you would avoid taking down your cluster during the upgrade.
>>>
>>> You'll be flagged on start if you need to run a migration but general
>>> rule is that there (should) never be need of a migration between
>>> patch
>>> releases: e.g. between 0.20.2 to 0.20.3.  There may be need of
>>> migrations moving between minor numbers; e.g. from 0.19 to 0.20.
>>>
>>> Let us know how IHBase works out for you (indexed hbase).  Its a RAM
>>> hog but the speed improvement finding matching cells can be
>>> startling.
>>>
>>> St.Ack
>>>
>>>> -----Original Message-----
>>>> From: saint.ack@gmail.com [mailto:saint.ack@gmail.com] On Behalf Of
>>>> Stack
>>>> Sent: Saturday, January 23, 2010 5:00 AM
>>>> To: hbase-user@hadoop.apache.org
>>>> Subject: Re: Support for MultiGet / SQL In clause -- error in patch
>>>> HBASE-1845
>>>>
>>>> Check your master log.  Something is seriously off if you do not
>>>> have
>>>> a reachable .META. table.
>>>> St.Ack
>>>>
>>>> On Fri, Jan 22, 2010 at 1:09 PM, Sriram Muthuswamy Chittathoor
>>>> <sriramc@ivycomptech.com> wrote:
>>>>> I applied the hbase-0.20.3 version / hadoop 0.20.1.  But after
>>>>> starting
>>>>> hbase I keep getting the error below when I go to the hbase shell
>>>>>
>>>>> [ppoker@karisimbivir1 hbase-0.20.3]$ ./bin/hbase shell
>>>>> HBase Shell; enter 'help<RETURN>' for list of supported commands.
>>>>> Version: 0.20.3, r900041, Sat Jan 16 17:20:21 PST 2010
>>>>> hbase(main):001:0> list
>>>>> NativeException:
>>>>> org.apache.hadoop.hbase.client.RetriesExhaustedException: Trying to
>>>>> contact region server null for region , row '', but failed after 7
>>>>> attempts.
>>>>> Exceptions:
>>>>> org.apache.hadoop.hbase.TableNotFoundException: .META.
>>>>> org.apache.hadoop.hbase.TableNotFoundException: .META.
>>>>> org.apache.hadoop.hbase.TableNotFoundException: .META.
>>>>> org.apache.hadoop.hbase.TableNotFoundException: .META.
>>>>> org.apache.hadoop.hbase.TableNotFoundException: .META.
>>>>> org.apache.hadoop.hbase.TableNotFoundException: .META.
>>>>> org.apache.hadoop.hbase.TableNotFoundException: .META.
>>>>>
>>>>>
>>>>>
>>>>> Also when I try to create a table programatically I get this --
>>>>> 10/01/22 15:48:23 INFO zookeeper.ClientCnxn: Attempting connection
>>>>> to
>>>>> server localhost/127.0.0.1:2181
>>>>> 10/01/22 15:48:23 INFO zookeeper.ClientCnxn: Priming connection to
>>>>> java.nio.channels.SocketChannel[connected local=/127.0.0.1:43775
>>>>> remote=localhost/127.0.0.1:2181]
>>>>> 10/01/22 15:48:23 INFO zookeeper.ClientCnxn: Server connection
>>>>> successful
>>>>> Exception in thread "main"
>>>>> org.apache.hadoop.hbase.TableNotFoundException: .META.
>>>>>       at
>>>>> org.apache.hadoop.hbase.client.HConnectionManager
>>>>> $TableServers.locateReg
>>>>> ionInMeta(HConnectionManager.java:684)
>>>>>       at
>>>>> org.apache.hadoop.hbase.client.HConnectionManager
>>>>> $TableServers.locateReg
>>>>> ion(HConnectionManager.java:634)
>>>>>       at
>>>>> org.apache.hadoop.hbase.client.HConnectionManager
>>>>> $TableServers.locateReg
>>>>> ion(HConnectionManager.java:601)
>>>>>       at
>>>>> org.apache.hadoop.hbase.client.HConnectionManager
>>>>> $TableServers.locateReg
>>>>> ionInMeta(HConnectionManager.java:675)
>>>>>       at
>>>>> org.apache.hadoop.hbase.client.HConnectionManager
>>>>> $TableServers.locateReg
>>>>> ion(HConnectionManager.java:638)
>>>>>       at
>>>>> org.apache.hadoop.hbase.client.HConnectionManager
>>>>> $TableServers.locateReg
>>>>> ion(HConnectionManager.java:601)
>>>>>       at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:
>>>>> 128)
>>>>>       at org.apache.hadoop.hbase.client.HTable.<init>(HTable.java:
>>>>> 106)
>>>>>       at test.CreateTable.main(CreateTable.java:36)
>>>>>
>>>>>
>>>>>
>>>>> Any clues ?
>>>>>
>>>>>
>>>>>
>>>>> -----Original Message-----
>>>>> From: Dan Washusen [mailto:dan@reactive.org]
>>>>> Sent: Friday, January 22, 2010 4:53 AM
>>>>> To: hbase-user@hadoop.apache.org
>>>>> Subject: Re: Support for MultiGet / SQL In clause -- error in patch
>>>>> HBASE-1845
>>>>>
>>>>> If you want to give the "indexed" contrib package a try you'll
>>>>> need to
>>>>> do
>>>>> the following:
>>>>>
>>>>>  1. Include the contrib jars (export HBASE_CLASSPATH=(`find
>>>>>  /path/to/hbase/hbase-0.20.3/contrib/indexed -name '*jar' | tr -s
>>>>> "\n"
>>>>> ":"`)
>>>>>  2. Set the 'hbase.hregion.impl' property to
>>>>>  'org.apache.hadoop.hbase.regionserver.IdxRegion' in your
>>>>> hbase-site.xml
>>>>>
>>>>> Once you've done that you can create a table with an index using:
>>>>>
>>>>>>    // define which qualifiers need an index (choosing the correct
>>>>> type)
>>>>>>    IdxColumnDescriptor columnDescriptor = new
>>>>>> IdxColumnDescriptor("columnFamily");
>>>>>>    columnDescriptor.addIndexDescriptor(
>>>>>>      new IdxIndexDescriptor("qualifier",
>>>>>> IdxQualifierType.BYTE_ARRAY)
>>>>>>    );
>>>>>>
>>>>>>    HTableDescriptor tableDescriptor = new HTableDescriptor
>>>>>> ("table");
>>>>>>    tableDescriptor.addFamily(columnDescriptor);
>>>>>>
>>>>>
>>>>> Then when you want to perform a scan with an index hint:
>>>>>
>>>>>>    Scan scan = new IdxScan(
>>>>>>          new Comparison("columnFamily", "qualifier",
>>>>>>              Comparison.Operator.EQ, Bytes.toBytes("foo"))
>>>>>>      );
>>>>>>
>>>>>
>>>>> You have to keep in mind that the index hint is only a hint.  It
>>>>> guarantees
>>>>> that your scan will get all rows that match the hint but you'll
>>>>> more
>>>>> than
>>>>> likely receive rows that don't.  For this reason I'd suggest that
>>>>> you
>>>>> also
>>>>> include a filter along with the scan:
>>>>>
>>>>>>      Scan scan = new IdxScan(
>>>>>>          new Comparison("columnFamily", "qualifier",
>>>>>>              Comparison.Operator.EQ, Bytes.toBytes("foo"))
>>>>>>      );
>>>>>>      scan.setFilter(
>>>>>>          new SingleColumnValueFilter(
>>>>>>              "columnFamily", "qualifer",
>>>>> CompareFilter.CompareOp.EQUAL,
>>>>>>              new BinaryComparator("foo")
>>>>>>          )
>>>>>>      );
>>>>>>
>>>>>
>>>>> Cheers,
>>>>> Dan
>>>>>
>>>>>
>>>>> 2010/1/22 stack <stack@duboce.net>
>>>>>
>>>>>>
>>>>>
>>
> http://people.apache.org/~jdcryans/hbase-0.20.3-candidate-2/<http://peop
>>>>> le.apache.org/%7Ejdcryans/hbase-0.20.3-candidate-2/>
>>>>>>
>>>>>> There is a bit of documentation if you look at javadoc for the
>>>>>> 'indexed' contrib (This is what hbase-2073 is called on commit).
>>>>>>
>>>>>> St.Ack
>>>>>>
>>>>>> P.S. We had a thread going named "HBase bulk load".  You got all
>>>>>> the
>>>>>> answers you need on that one?
>>>>>>
>>>>>> On Thu, Jan 21, 2010 at 11:19 AM, Sriram Muthuswamy Chittathoor
>>>>>> <sriramc@ivycomptech.com> wrote:
>>>>>>>
>>>>>>> Great.  Can I migrate to 0.20.3RC2 easily.  I am on 0.20.2.
Can u
>>>>> pass
>>>>>>> me the link
>>>>>>>
>>>>>>> -----Original Message-----
>>>>>>> From: saint.ack@gmail.com [mailto:saint.ack@gmail.com] On Behalf
>>>>>>> Of
>>>>>>> stack
>>>>>>> Sent: Friday, January 22, 2010 12:42 AM
>>>>>>> To: hbase-user@hadoop.apache.org
>>>>>>> Subject: Re: Support for MultiGet / SQL In clause -- error in
>>>>>>> patch
>>>>>>> HBASE-1845
>>>>>>>
>>>>>>> IIRC, hbase-1845 was a sketch only and not yet complete.  Its
>>>>> probably
>>>>>>> rotted since any ways.
>>>>>>>
>>>>>>> Have you looked at hbase-2037 since committed and available in
>>>>>>> 0.20.3RC2.
>>>>>>> Would this help you with your original problem?
>>>>>>>
>>>>>>> St.Ack
>>>>>>>
>>>>>>> On Thu, Jan 21, 2010 at 9:10 AM, Sriram Muthuswamy Chittathoor
<
>>>>>>> sriramc@ivycomptech.com> wrote:
>>>>>>>
>>>>>>>> I tried applying the patch to the hbase source code  hbase
>>>>>>>> 0.20.2
>>>>> and
>>>>>>> I
>>>>>>>> get the errors below.  Do you know if this needs to be applied
>>>>>>>> to
>>>>> a
>>>>>>>> specific hbase version. Is there a version which works with
>>>>>>>> 0.20.2
>>>>> or
>>>>>>>> later ??
>>>>>>>> Basically HRegionServer  and HTable patching fails.
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks for the help
>>>>>>>>
>>>>>>>> patch -p0 -i batch.patch
>>>>>>>>
>>>>>>>> patching file src/java/org/apache/hadoop/hbase/client/Get.java
>>>>>>>> Hunk #1 succeeded at 61 (offset 2 lines).
>>>>>>>> Hunk #2 succeeded at 347 (offset 31 lines).
>>>>>>>> patching file
>>>>> src/java/org/apache/hadoop/hbase/client/HConnection.java
>>>>>>>> patching file
>>>>>>>> src/java/org/apache/hadoop/hbase/client/HConnectionManager.java
>>>>>>>> Hunk #3 succeeded at 1244 (offset 6 lines).
>>>>>>>> patching file src/java/org/apache/hadoop/hbase/client/
>>>>>>>> HTable.java
>>>>>>>> Hunk #2 succeeded at 73 (offset 8 lines).
>>>>>>>> Hunk #4 FAILED at 405.
>>>>>>>> Hunk #5 succeeded at 671 with fuzz 2 (offset 26 lines).
>>>>>>>> 1 out of 5 hunks FAILED -- saving rejects to file
>>>>>>>> src/java/org/apache/hadoop/hbase/client/HTable.java.rej
>>>>>>>> patching file src/java/org/apache/hadoop/hbase/client/Multi.java
>>>>>>>> patching file
>>>>>>> src/java/org/apache/hadoop/hbase/client/MultiCallable.java
>>>>>>>> patching file
>>>>> src/java/org/apache/hadoop/hbase/client/MultiResult.java
>>>>>>>> patching file src/java/org/apache/hadoop/hbase/client/Row.java
>>>>>>>> patching file
>>>>>>>> src/java/org/apache/hadoop/hbase/io/HbaseObjectWritable.java
>>>>>>>> Hunk #2 succeeded at 156 with fuzz 1 (offset 3 lines).
>>>>>>>> patching file
>>>>>>> src/java/org/apache/hadoop/hbase/ipc/HRegionInterface.java
>>>>>>>> Hunk #2 succeeded at 247 (offset 2 lines).
>>>>>>>> patching file
>>>>>>>> src/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
>>>>>>>> Hunk #1 succeeded at 78 (offset -1 lines).
>>>>>>>> Hunk #2 FAILED at 2515.
>>>>>>>> 1 out of 2 hunks FAILED -- saving rejects to file
>>>>>>>>
>>>>> src/java/org/apache/hadoop/hbase/regionserver/
>>>>> HRegionServer.java.rej
>>>>>>>> patching file
>>>>> src/test/org/apache/hadoop/hbase/client/TestHTable.java
>>>>>>>> Hunk #2 FAILED at 333.
>>>>>>>> 1 out of 2 hunks FAILED -- saving rejects to file
>>>>>>>> src/test/org/apache/hadoop/hbase/client/TestHTable.java.rej
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> -----Original Message-----
>>>>>>>> From: Marc Limotte [mailto:mslimotte@gmail.com]
>>>>>>>> Sent: Tuesday, January 19, 2010 10:26 PM
>>>>>>>> To: hbase-user@hadoop.apache.org
>>>>>>>> Subject: Re: Support for MultiGet / SQL In clause
>>>>>>>>
>>>>>>>> Sriram,
>>>>>>>>
>>>>>>>> Would a secondary index help you:
>>>>>>>>
>>>>>>>
>>>>>
>>
> http://hadoop.apache.org/hbase/docs/r0.20.2/api/org/apache/hadoop/hbase/
>>>>>>>> client/tableindexed/package-summary.html#package_description
>>>>>>>> .
>>>>>>>>
>>>>>>>> The index is stored in a separate table, but the index is
>>>>>>>> managed
>>>>> for
>>>>>>>> you.
>>>>>>>>
>>>>>>>> I don't think you can do an arbitrary "in" query, though.
 If
>>>>>>>> the
>>>>> keys
>>>>>>>> that
>>>>>>>> you want to include in the "in" are reasonably close neighbors,
>>>>> you
>>>>>>>> could do
>>>>>>>> a scan and skip ones that are uninteresting.  You could
also
>>>>>>>> try a
>>>>>>> batch
>>>>>>>> Get
>>>>>>>> by applying a separate patch, see
>>>>>>>> http://issues.apache.org/jira/browse/HBASE-1845.
>>>>>>>>
>>>>>>>> Marc Limotte
>>>>>>>>
>>>>>>>> On Tue, Jan 19, 2010 at 8:45 AM, Sriram Muthuswamy Chittathoor
<
>>>>>>>> sriramc@ivycomptech.com> wrote:
>>>>>>>>
>>>>>>>>> Is there any support for this.  I want to do this
>>>>>>>>>
>>>>>>>>> 1.  Create a second table to maintain mapping between
secondary
>>>>>>> column
>>>>>>>>> and the rowid's of the primary table
>>>>>>>>>
>>>>>>>>> 2.  Use this second table to get the rowid's to lookup
from the
>>>>>>>> primary
>>>>>>>>> table using a SQL In like clause ---
>>>>>>>>>
>>>>>>>>> Basically I am doing this to speed up querying by  Non-row
key
>>>>>>>> columns.
>>>>>>>>>
>>>>>>>>> Thanks
>>>>>>>>>
>>>>>>>>> Sriram C
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> This email is sent for and on behalf of Ivy Comptech
Private
>>>>>>> Limited.
>>>>>>>> Ivy
>>>>>>>>> Comptech Private Limited is a limited liability company.
>>>>>>>>>
>>>>>>>>> This email and any attachments are confidential, and
may be
>>>>> legally
>>>>>>>>> privileged and protected by copyright. If you are not
the
>>>>> intended
>>>>>>>> recipient
>>>>>>>>> dissemination or copying of this email is prohibited.
If you
>>>>> have
>>>>>>>> received
>>>>>>>>> this in error, please notify the sender by replying by
email
>>>>>>>>> and
>>>>>>> then
>>>>>>>> delete
>>>>>>>>> the email completely from your system.
>>>>>>>>> Any views or opinions are solely those of the sender.
 This
>>>>>>>> communication
>>>>>>>>> is not intended to form a binding contract on behalf
of Ivy
>>>>> Comptech
>>>>>>>> Private
>>>>>>>>> Limited unless expressly indicated to the contrary and
properly
>>>>>>>> authorised.
>>>>>>>>> Any actions taken on the basis of this email are at the
>>>>> recipient's
>>>>>>>> own
>>>>>>>>> risk.
>>>>>>>>>
>>>>>>>>> Registered office:
>>>>>>>>> Ivy Comptech Private Limited, Cyber Spazio, Road No.
2, Banjara
>>>>>>> Hills,
>>>>>>>>> Hyderabad 500 033, Andhra Pradesh, India. Registered
number:
>>>>> 37994.
>>>>>>>>> Registered in India. A list of members' names is available
for
>>>>>>>> inspection at
>>>>>>>>> the registered office.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>
>>>>>
>>>>
>>>> This email is sent for and on behalf of Ivy Comptech Private
>>>> Limited. Ivy Comptech Private Limited is a limited liability
>>>> company.
>>>>
>>>> This email and any attachments are confidential, and may be legally
>>>> privileged and protected by copyright. If you are not the intended
>>>> recipient dissemination or copying of this email is prohibited. If
>>>> you have received this in error, please notify the sender by
>>>> replying by email and then delete the email completely from your
>>>> system.
>>>> Any views or opinions are solely those of the sender.  This
>>>> communication is not intended to form a binding contract on behalf
>>>> of Ivy Comptech Private Limited unless expressly indicated to the
>>>> contrary and properly authorised. Any actions taken on the basis of
>>>> this email are at the recipient's own risk.
>>>>
>>>> Registered office:
>>>> Ivy Comptech Private Limited, Cyber Spazio, Road No. 2, Banjara
>>>> Hills, Hyderabad 500 033, Andhra Pradesh, India. Registered number:
>>>> 37994. Registered in India. A list of members' names is available
>>>> for inspection at the registered office.
>>>>
>>>>
>>>
>>> This email is sent for and on behalf of Ivy Comptech Private
>>> Limited. Ivy Comptech Private Limited is a limited liability company.
>>>
>>> This email and any attachments are confidential, and may be legally
>>> privileged and protected by copyright. If you are not the intended
>>> recipient dissemination or copying of this email is prohibited. If
>>> you have received this in error, please notify the sender by
>>> replying by email and then delete the email completely from your
>>> system.
>>> Any views or opinions are solely those of the sender.  This
>>> communication is not intended to form a binding contract on behalf
>>> of Ivy Comptech Private Limited unless expressly indicated to the
>>> contrary and properly authorised. Any actions taken on the basis of
>>> this email are at the recipient's own risk.
>>>
>>> Registered office:
>>> Ivy Comptech Private Limited, Cyber Spazio, Road No. 2, Banjara
>>> Hills, Hyderabad 500 033, Andhra Pradesh, India. Registered number:
>>> 37994. Registered in India. A list of members' names is available
>>> for inspection at the registered office.
>>>
>>
>> This email is sent for and on behalf of Ivy Comptech Private
>> Limited. Ivy Comptech Private Limited is a limited liability company.
>>
>> This email and any attachments are confidential, and may be legally
>> privileged and protected by copyright. If you are not the intended
>> recipient dissemination or copying of this email is prohibited. If
>> you have received this in error, please notify the sender by
>> replying by email and then delete the email completely from your
>> system.
>> Any views or opinions are solely those of the sender.  This
>> communication is not intended to form a binding contract on behalf
>> of Ivy Comptech Private Limited unless expressly indicated to the
>> contrary and properly authorised. Any actions taken on the basis of
>> this email are at the recipient's own risk.
>>
>> Registered office:
>> Ivy Comptech Private Limited, Cyber Spazio, Road No. 2, Banjara
>> Hills, Hyderabad 500 033, Andhra Pradesh, India. Registered number:
>> 37994. Registered in India. A list of members' names is available
>> for inspection at the registered office.
>>
>
> This email is sent for and on behalf of Ivy Comptech Private Limited. Ivy Comptech Private
Limited is a limited liability company.
>
> This email and any attachments are confidential, and may be legally privileged and protected
by copyright. If you are not the intended recipient dissemination or copying of this email
is prohibited. If you have received this in error, please notify the sender by replying by
email and then delete the email completely from your system.
> Any views or opinions are solely those of the sender.  This communication is not intended
to form a binding contract on behalf of Ivy Comptech Private Limited unless expressly indicated
to the contrary and properly authorised. Any actions taken on the basis of this email are
at the recipient's own risk.
>
> Registered office:
> Ivy Comptech Private Limited, Cyber Spazio, Road No. 2, Banjara Hills, Hyderabad 500
033, Andhra Pradesh, India. Registered number: 37994. Registered in India. A list of members'
names is available for inspection at the registered office.
>
>

Mime
View raw message