hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Ted Yu <yuzhih...@gmail.com>
Subject Re: hbase 0.96 stop master receive ERROR ipc.RPC: RPC.stopProxy called on non proxy.
Date Mon, 25 Nov 2013 18:54:53 GMT
Henry:
See HBASE-7635 Proxy created by HFileSystem#createReorderingProxy() should
implement Closeable


On Tue, Nov 26, 2013 at 1:56 AM, Ted Yu <yuzhihong@gmail.com> wrote:

> Here is the caller to createReorderingProxy():
>
>       ClientProtocol cp1 = createReorderingProxy(namenode, lrb, conf);
>
> where namenode
> is org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB :
>
> public class ClientNamenodeProtocolTranslatorPB implements
>
>     ProtocolMetaInterface, ClientProtocol, Closeable, ProtocolTranslator {
>
> In createReorderingProxy() :
>
>             new Class[]{ClientProtocol.class, Closeable.class},
>
> We ask for Closeable interface.
>
>
> Did the error persist after you replaced with the hadoop-hdfs-2.2.0.jar ?
> Meaning, did you start HBase using the new hadoop jars ?
>
> Cheers
>
>
> On Mon, Nov 25, 2013 at 1:04 PM, Henry Hung <YTHung1@winbond.com> wrote:
>
>> I looked into the source code of
>> org/apache/hadoop/hbase/fs/HFileSystem.java
>> and whenever I execute hbase-daemon.sh stop master (or regionserver), the
>> method.getName() is "close",
>> but the org/apache/hadoop/hdfs/protocol/ClientProtocol.java do not have
>> method named "close",
>> thus it result in error "object is not an instance of declaring class"
>>
>> Could someone that familiar with hbase-0.96.0 hadoop2 can tell me if this
>> problem need to be fixed? And how to fix it?
>>
>>   private static ClientProtocol createReorderingProxy(final
>> ClientProtocol cp, final ReorderBlocks lrb, final Configuration conf) {
>>     return (ClientProtocol) Proxy.newProxyInstance
>>         (cp.getClass().getClassLoader(),
>>             new Class[]{ClientProtocol.class, Closeable.class},
>>             new InvocationHandler() {
>>               public Object invoke(Object proxy, Method method,
>>                                    Object[] args) throws Throwable {
>>                 try {
>>                   // method.invoke will failed if
>> method.getName().equals("close")
>>                   // because ClientProtocol do not have method "close"
>>                   Object res = method.invoke(cp, args);
>>                   if (res != null && args != null && args.length
== 3
>>                       && "getBlockLocations".equals(method.getName())
>>                       && res instanceof LocatedBlocks
>>                       && args[0] instanceof String
>>                       && args[0] != null) {
>>                     lrb.reorderBlocks(conf, (LocatedBlocks) res, (String)
>> args[0]);
>>                   }
>>                   return res;
>>                 } catch  (InvocationTargetException ite) {
>>                   // We will have this for all the exception, checked on
>> not, sent
>>                   //  by any layer, including the functional exception
>>                   Throwable cause = ite.getCause();
>>                   if (cause == null){
>>                     throw new RuntimeException(
>>                       "Proxy invocation failed and getCause is null",
>> ite);
>>                   }
>>                   if (cause instanceof UndeclaredThrowableException) {
>>                     Throwable causeCause = cause.getCause();
>>                     if (causeCause == null) {
>>                       throw new
>> RuntimeException("UndeclaredThrowableException had null cause!");
>>                     }
>>                     cause = cause.getCause();
>>                   }
>>                   throw cause;
>>                 }
>>               }
>>             });
>>   }
>>
>>
>>
>> -----Original Message-----
>> From: MA11 YTHung1
>> Sent: Thursday, November 21, 2013 9:57 AM
>> To: user@hbase.apache.org
>> Subject: RE: hbase 0.96 stop master receive ERROR ipc.RPC: RPC.stopProxy
>> called on non proxy.
>>
>> Additional information:
>>
>> I replace all files with prefix hadoop in hbase-0.96.0-hadoop2/lib with
>> hadoop-2.2.0 libraries.
>>
>> the ls -l of hbase-0.96.0-hadoop2/lib as below:
>>
>> -rw-r--r-- 1 hadoop users    62983 Sep 17 16:13 activation-1.1.jar
>> -rw-r--r-- 1 hadoop users     4467 Sep 17 23:29 aopalliance-1.0.jar
>> -rw-r--r-- 1 hadoop users    43033 Sep 17 16:13 asm-3.1.jar
>> -rw-r--r-- 1 hadoop users   263268 Sep 17 16:27 avro-1.5.3.jar
>> -rw-r--r-- 1 hadoop users   188671 Sep 17 16:12
>> commons-beanutils-1.7.0.jar
>> -rw-r--r-- 1 hadoop users   206035 Sep 17 16:13
>> commons-beanutils-core-1.8.0.jar
>> -rw-r--r-- 1 hadoop users    41123 Sep 17 16:12 commons-cli-1.2.jar
>> -rw-r--r-- 1 hadoop users   259600 Sep 17 16:13 commons-codec-1.7.jar
>> -rw-r--r-- 1 hadoop users   575389 Sep 17 16:12
>> commons-collections-3.2.1.jar
>> -rw-r--r-- 1 hadoop users   238681 Sep 17 16:27 commons-compress-1.4.jar
>> -rw-r--r-- 1 hadoop users   298829 Sep 17 16:13
>> commons-configuration-1.6.jar
>> -rw-r--r-- 1 hadoop users    24239 Sep 17 23:28 commons-daemon-1.0.13.jar
>> -rw-r--r-- 1 hadoop users   143602 Sep 17 16:12 commons-digester-1.8.jar
>> -rw-r--r-- 1 hadoop users   112341 Sep 17 16:13 commons-el-1.0.jar
>> -rw-r--r-- 1 hadoop users   305001 Sep 17 16:12 commons-httpclient-3.1.jar
>> -rw-r--r-- 1 hadoop users   185140 Sep 17 16:13 commons-io-2.4.jar
>> -rw-r--r-- 1 hadoop users   284220 Sep 17 16:12 commons-lang-2.6.jar
>> -rw-r--r-- 1 hadoop users    60686 Sep 17 16:12 commons-logging-1.1.1.jar
>> -rw-r--r-- 1 hadoop users   988514 Sep 17 16:13 commons-math-2.2.jar
>> -rw-r--r-- 1 hadoop users   273370 Sep 17 16:27 commons-net-3.1.jar
>> -rw-r--r-- 1 hadoop users  3566844 Sep 17 16:15 core-3.1.1.jar
>> -rw-r--r-- 1 hadoop users    15322 Sep 17 16:12
>> findbugs-annotations-1.3.9-1.jar
>> -rw-r--r-- 1 hadoop users    21817 Sep 17 23:29
>> gmbal-api-only-3.0.0-b023.jar
>> -rw-r--r-- 1 hadoop users   684337 Sep 17 23:29
>> grizzly-framework-2.1.1.jar
>> -rw-r--r-- 1 hadoop users   210846 Sep 17 23:29
>> grizzly-framework-2.1.1-tests.jar
>> -rw-r--r-- 1 hadoop users   248346 Sep 17 23:29 grizzly-http-2.1.1.jar
>> -rw-r--r-- 1 hadoop users   193583 Sep 17 23:29
>> grizzly-http-server-2.1.1.jar
>> -rw-r--r-- 1 hadoop users   336878 Sep 17 23:29
>> grizzly-http-servlet-2.1.1.jar
>> -rw-r--r-- 1 hadoop users     8072 Sep 17 23:29 grizzly-rcm-2.1.1.jar
>> -rw-r--r-- 1 hadoop users  1795932 Sep 17 16:13 guava-12.0.1.jar
>> -rw-r--r-- 1 hadoop users   710492 Sep 17 23:29 guice-3.0.jar
>> -rw-r--r-- 1 hadoop users    65012 Sep 17 23:29 guice-servlet-3.0.jar
>> -rw-r--r-- 1 hadoop users    16778 Nov 20 17:39
>> hadoop-annotations-2.2.0.jar
>> -rw-r--r-- 1 hadoop users    49750 Nov 20 17:40 hadoop-auth-2.2.0.jar
>> -rw-r--r-- 1 hadoop users     2576 Oct 12 06:20
>> hadoop-client-2.1.0-beta.jar
>> -rw-r--r-- 1 hadoop users  2735584 Nov 20 17:50 hadoop-common-2.2.0.jar
>> -rw-r--r-- 1 hadoop users  5242252 Nov 21 08:48 hadoop-hdfs-2.2.0.jar
>> -rw-r--r-- 1 hadoop users  1988460 Nov 21 08:48
>> hadoop-hdfs-2.2.0-tests.jar
>> -rw-r--r-- 1 hadoop users   482042 Nov 21 08:49
>> hadoop-mapreduce-client-app-2.2.0.jar
>> -rw-r--r-- 1 hadoop users   656365 Nov 21 08:49
>> hadoop-mapreduce-client-common-2.2.0.jar
>> -rw-r--r-- 1 hadoop users  1455001 Nov 21 08:50
>> hadoop-mapreduce-client-core-2.2.0.jar
>> -rw-r--r-- 1 hadoop users    35216 Nov 21 08:50
>> hadoop-mapreduce-client-jobclient-2.2.0.jar
>> -rw-r--r-- 1 hadoop users  1434852 Nov 21 08:50
>> hadoop-mapreduce-client-jobclient-2.2.0-tests.jar
>> -rw-r--r-- 1 hadoop users    21537 Nov 21 08:51
>> hadoop-mapreduce-client-shuffle-2.2.0.jar
>> -rw-r--r-- 1 hadoop users  1158936 Nov 21 08:51 hadoop-yarn-api-2.2.0.jar
>> -rw-r--r-- 1 hadoop users    94728 Nov 21 08:51
>> hadoop-yarn-client-2.2.0.jar
>> -rw-r--r-- 1 hadoop users  1301627 Nov 21 08:51
>> hadoop-yarn-common-2.2.0.jar
>> -rw-r--r-- 1 hadoop users   175554 Nov 21 08:52
>> hadoop-yarn-server-common-2.2.0.jar
>> -rw-r--r-- 1 hadoop users   467638 Nov 21 08:52
>> hadoop-yarn-server-nodemanager-2.2.0.jar
>> -rw-r--r-- 1 hadoop users   825853 Oct 12 06:28
>> hbase-client-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users   354845 Oct 12 06:28
>> hbase-common-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users   132690 Oct 12 06:28
>> hbase-common-0.96.0-hadoop2-tests.jar
>> -rw-r--r-- 1 hadoop users    97428 Oct 12 06:28
>> hbase-examples-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users    72765 Oct 12 06:28
>> hbase-hadoop2-compat-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users    32096 Oct 12 06:28
>> hbase-hadoop-compat-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users    12174 Oct 12 06:28
>> hbase-it-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users   288784 Oct 12 06:28
>> hbase-it-0.96.0-hadoop2-tests.jar
>> -rw-r--r-- 1 hadoop users    94784 Oct 12 06:28
>> hbase-prefix-tree-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users  3134214 Oct 12 06:28
>> hbase-protocol-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users  3058804 Oct 12 06:28
>> hbase-server-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users  3150292 Oct 12 06:28
>> hbase-server-0.96.0-hadoop2-tests.jar
>> -rw-r--r-- 1 hadoop users    12554 Oct 12 06:28
>> hbase-shell-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users    10941 Oct 12 06:28
>> hbase-testing-util-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users  2276333 Oct 12 06:28
>> hbase-thrift-0.96.0-hadoop2.jar
>> -rw-r--r-- 1 hadoop users    95975 Sep 17 16:15 high-scale-lib-1.1.1.jar
>> -rw-r--r-- 1 hadoop users    31020 Sep 17 16:14 htrace-core-2.01.jar
>> -rw-r--r-- 1 hadoop users   352585 Sep 17 16:15 httpclient-4.1.3.jar
>> -rw-r--r-- 1 hadoop users   181201 Sep 17 16:15 httpcore-4.1.3.jar
>> -rw-r--r-- 1 hadoop users   227517 Sep 17 16:13 jackson-core-asl-1.8.8.jar
>> -rw-r--r-- 1 hadoop users    17884 Sep 17 16:13 jackson-jaxrs-1.8.8.jar
>> -rw-r--r-- 1 hadoop users   669065 Sep 17 16:13
>> jackson-mapper-asl-1.8.8.jar
>> -rw-r--r-- 1 hadoop users    32353 Sep 17 16:13 jackson-xc-1.8.8.jar
>> -rw-r--r-- 1 hadoop users    20642 Sep 17 16:15 jamon-runtime-2.3.1.jar
>> -rw-r--r-- 1 hadoop users   408133 Sep 17 16:13 jasper-compiler-5.5.23.jar
>> -rw-r--r-- 1 hadoop users    76844 Sep 17 16:13 jasper-runtime-5.5.23.jar
>> -rw-r--r-- 1 hadoop users     2497 Sep 17 23:29 javax.inject-1.jar
>> -rw-r--r-- 1 hadoop users    83586 Sep 17 23:29 javax.servlet-3.0.jar
>> -rw-r--r-- 1 hadoop users   105134 Sep 17 16:27 jaxb-api-2.2.2.jar
>> -rw-r--r-- 1 hadoop users   890168 Sep 17 16:13 jaxb-impl-2.2.3-1.jar
>> -rw-r--r-- 1 hadoop users   129217 Sep 17 23:29 jersey-client-1.8.jar
>> -rw-r--r-- 1 hadoop users   458233 Sep 17 16:13 jersey-core-1.8.jar
>> -rw-r--r-- 1 hadoop users    17585 Sep 17 23:29 jersey-grizzly2-1.8.jar
>> -rw-r--r-- 1 hadoop users    14712 Sep 17 23:29 jersey-guice-1.8.jar
>> -rw-r--r-- 1 hadoop users   147933 Sep 17 16:13 jersey-json-1.8.jar
>> -rw-r--r-- 1 hadoop users   694352 Sep 17 16:13 jersey-server-1.8.jar
>> -rw-r--r-- 1 hadoop users    28034 Sep 17 23:29
>> jersey-test-framework-core-1.8.jar
>> -rw-r--r-- 1 hadoop users    12907 Sep 17 23:29
>> jersey-test-framework-grizzly2-1.8.jar
>> -rw-r--r-- 1 hadoop users   321806 Sep 17 16:27 jets3t-0.6.1.jar
>> -rw-r--r-- 1 hadoop users    75963 Sep 17 16:13 jettison-1.3.1.jar
>> -rw-r--r-- 1 hadoop users   539912 Sep 17 16:13 jetty-6.1.26.jar
>> -rw-r--r-- 1 hadoop users    18891 Sep 17 16:15 jetty-sslengine-6.1.26.jar
>> -rw-r--r-- 1 hadoop users   177131 Sep 17 16:13 jetty-util-6.1.26.jar
>> -rw-r--r-- 1 hadoop users 13832273 Sep 17 16:15 jruby-complete-1.6.8.jar
>> -rw-r--r-- 1 hadoop users   185746 Sep 17 16:27 jsch-0.1.42.jar
>> -rw-r--r-- 1 hadoop users  1024680 Sep 17 16:13 jsp-2.1-6.1.14.jar
>> -rw-r--r-- 1 hadoop users   134910 Sep 17 16:13 jsp-api-2.1-6.1.14.jar
>> -rw-r--r-- 1 hadoop users   100636 Sep 17 16:27 jsp-api-2.1.jar
>> -rw-r--r-- 1 hadoop users    33015 Sep 17 16:13 jsr305-1.3.9.jar
>> -rw-r--r-- 1 hadoop users   245039 Sep 17 16:12 junit-4.11.jar
>> -rw-r--r-- 1 hadoop users   347531 Sep 17 16:15 libthrift-0.9.0.jar
>> -rw-r--r-- 1 hadoop users   489884 Sep 17 16:12 log4j-1.2.17.jar
>> -rw-r--r-- 1 hadoop users    42212 Sep 17 23:29
>> management-api-3.0.0-b012.jar
>> -rw-r--r-- 1 hadoop users    82445 Sep 17 16:14 metrics-core-2.1.2.jar
>> drwxr-xr-x 3 hadoop users     4096 Nov 21 09:10 native
>> -rw-r--r-- 1 hadoop users  1206119 Sep 18 04:00 netty-3.6.6.Final.jar
>> -rw-r--r-- 1 hadoop users    29555 Sep 17 16:27 paranamer-2.3.jar
>> -rw-r--r-- 1 hadoop users   533455 Sep 17 16:13 protobuf-java-2.5.0.jar
>> drwxr-xr-x 5 hadoop users     4096 Sep 28 10:37 ruby
>> -rw-r--r-- 1 hadoop users   132368 Sep 17 16:13 servlet-api-2.5-6.1.14.jar
>> -rw-r--r-- 1 hadoop users   105112 Sep 17 16:12 servlet-api-2.5.jar
>> -rw-r--r-- 1 hadoop users    25962 Sep 17 16:14 slf4j-api-1.6.4.jar
>> -rw-r--r-- 1 hadoop users     9748 Oct  3 07:15 slf4j-log4j12-1.6.4.jar
>> -rw-r--r-- 1 hadoop users   995720 Sep 17 16:27 snappy-java-1.0.3.2.jar
>> -rw-r--r-- 1 hadoop users    26514 Sep 17 16:13 stax-api-1.0.1.jar
>> -rw-r--r-- 1 hadoop users    15010 Sep 17 16:13 xmlenc-0.52.jar
>> -rw-r--r-- 1 hadoop users    94672 Sep 17 16:27 xz-1.0.jar
>> -rw-r--r-- 1 hadoop users   779974 Sep 17 16:14 zookeeper-3.4.5.jar
>>
>> Best regards,
>> Henry
>>
>> -----Original Message-----
>> From: MA11 YTHung1
>> Sent: Thursday, November 21, 2013 9:51 AM
>> To: user@hbase.apache.org
>> Subject: RE: hbase 0.96 stop master receive ERROR ipc.RPC: RPC.stopProxy
>> called on non proxy.
>>
>> I'm using hadoop-2.2.0 stable
>>
>> -----Original Message-----
>> From: Jimmy Xiang [mailto:jxiang@cloudera.com]
>> Sent: Thursday, November 21, 2013 9:49 AM
>> To: user
>> Subject: Re: hbase 0.96 stop master receive ERROR ipc.RPC: RPC.stopProxy
>> called on non proxy.
>>
>> Which version of Hadoop do you use?
>>
>>
>> On Wed, Nov 20, 2013 at 5:43 PM, Henry Hung <YTHung1@winbond.com> wrote:
>>
>> > Hi All,
>> >
>> > When stopping master or regionserver, I found some ERROR and WARN in
>> > the log files, are these errors can cause problem in hbase:
>> >
>> > 13/11/21 09:31:16 INFO zookeeper.ClientCnxn: EventThread shut down
>> > 13/11/21 09:35:36 ERROR ipc.RPC: RPC.stopProxy called on non proxy.
>> > java.lang.IllegalArgumentException: object is not an instance of
>> > declaring class
>> >         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> >         at
>> >
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> >         at
>> >
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> >         at java.lang.reflect.Method.invoke(Method.java:597)
>> >         at
>> > org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:266)
>> >         at $Proxy18.close(Unknown Source)
>> >         at org.apache.hadoop.ipc.RPC.stopProxy(RPC.java:621)
>> >         at
>> >
>> org.apache.hadoop.hdfs.DFSClient.closeConnectionToNamenode(DFSClient.java:738)
>> >         at org.apache.hadoop.hdfs.DFSClient.close(DFSClient.java:794)
>> >         at
>> >
>> org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSystem.java:847)
>> >         at
>> > org.apache.hadoop.fs.FileSystem$Cache.closeAll(FileSystem.java:2524)
>> >         at
>> >
>> org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer.run(FileSystem.java:2541)
>> >         at
>> > org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.j
>> > ava:54)
>> > 13/11/21 09:35:36 WARN util.ShutdownHookManager: ShutdownHook
>> > 'ClientFinalizer' failed,
>> org.apache.hadoop.HadoopIllegalArgumentException:
>> > Cannot close proxy - is not Closeable or does not provide closeable
>> > invocation handler class $Proxy18
>> > org.apache.hadoop.HadoopIllegalArgumentException: Cannot close proxy -
>> > is not Closeable or does not provide closeable invocation handler
>> > class
>> > $Proxy18
>> >         at org.apache.hadoop.ipc.RPC.stopProxy(RPC.java:639)
>> >         at
>> >
>> org.apache.hadoop.hdfs.DFSClient.closeConnectionToNamenode(DFSClient.java:738)
>> >         at org.apache.hadoop.hdfs.DFSClient.close(DFSClient.java:794)
>> >         at
>> >
>> org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSystem.java:847)
>> >         at
>> > org.apache.hadoop.fs.FileSystem$Cache.closeAll(FileSystem.java:2524)
>> >         at
>> >
>> org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer.run(FileSystem.java:2541)
>> >         at
>> > org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.j
>> > ava:54)
>> >
>> > Best regards,
>> > Henry
>> >
>> > ________________________________
>> > The privileged confidential information contained in this email is
>> > intended for use only by the addressees as indicated by the original
>> > sender of this email. If you are not the addressee indicated in this
>> > email or are not responsible for delivery of the email to such a
>> > person, please kindly reply to the sender indicating this fact and
>> > delete all copies of it from your computer and network server
>> > immediately. Your cooperation is highly appreciated. It is advised
>> > that any unauthorized use of confidential information of Winbond is
>> > strictly prohibited; and any information in this email irrelevant to
>> > the official business of Winbond shall be deemed as neither given nor
>> endorsed by Winbond.
>> >
>>
>> The privileged confidential information contained in this email is
>> intended for use only by the addressees as indicated by the original sender
>> of this email. If you are not the addressee indicated in this email or are
>> not responsible for delivery of the email to such a person, please kindly
>> reply to the sender indicating this fact and delete all copies of it from
>> your computer and network server immediately. Your cooperation is highly
>> appreciated. It is advised that any unauthorized use of confidential
>> information of Winbond is strictly prohibited; and any information in this
>> email irrelevant to the official business of Winbond shall be deemed as
>> neither given nor endorsed by Winbond.
>>
>> The privileged confidential information contained in this email is
>> intended for use only by the addressees as indicated by the original sender
>> of this email. If you are not the addressee indicated in this email or are
>> not responsible for delivery of the email to such a person, please kindly
>> reply to the sender indicating this fact and delete all copies of it from
>> your computer and network server immediately. Your cooperation is highly
>> appreciated. It is advised that any unauthorized use of confidential
>> information of Winbond is strictly prohibited; and any information in this
>> email irrelevant to the official business of Winbond shall be deemed as
>> neither given nor endorsed by Winbond.
>>
>> The privileged confidential information contained in this email is
>> intended for use only by the addressees as indicated by the original sender
>> of this email. If you are not the addressee indicated in this email or are
>> not responsible for delivery of the email to such a person, please kindly
>> reply to the sender indicating this fact and delete all copies of it from
>> your computer and network server immediately. Your cooperation is highly
>> appreciated. It is advised that any unauthorized use of confidential
>> information of Winbond is strictly prohibited; and any information in this
>> email irrelevant to the official business of Winbond shall be deemed as
>> neither given nor endorsed by Winbond.
>>
>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message