Skip to main content

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [List Home]
Re: [geomesa-dev] Problems running GeoMesa GDELT tutorial

Bob, Chris E, et al,

 

I grabbed the link that Bob used (http://data.gdeltproject.org/events/GDELT.MASTERREDUCEDV2.1979-2013.zip) and it is a reduced data set that only has 17 columns.  The main historical back files available at http://data.gdeltproject.org/events/index.html have 57 columns.

 

Maybe the GeoMesa  GDELT ingestion project can add a –reduced=[true | false] which then operated on the reduced file vice a full historical backfile?

 

On a side note, the geomesa-gdelt project readme on github indicates that “…ingestFile be a gdelt format TSV …” meaning that they should be Tab separated values.  However, all the downloads I have found in a quick perusal of the gdelt project download page are all CSV formatted.  Is the geomesa-gdelt readme in error and the files really are CSV, or should the CSV formats be mapped to TSV before running through the geomesa ingestion?

 

Thanks,

 

Chris Snider

Senior Software Engineer

Intelligent Software Solutions, Inc.

Direct (719) 452-7257

Description: Description: Description: cid:image001.png@01CA1F1F.CBC93990

 

From: Barnhart, Bob M. [mailto:ROBERT.M.BARNHART@xxxxxxxxxx]
Sent: Thursday, May 08, 2014 4:45 PM
To: Chris Snider; Discussions between GeoMesa committers
Subject: RE: [geomesa-dev] Problems running GeoMesa GDELT tutorial

 

Thanks Chris. Good luck!

 

From: Chris Snider [mailto:chris.snider@xxxxxxxxxx]
Sent: Thursday, May 08, 2014 3:21 PM
To: Barnhart, Bob M.; Discussions between GeoMesa committers
Subject: RE: [geomesa-dev] Problems running GeoMesa GDELT tutorial

 

Bob,

 

I haven’t run the gdelt dataset myself yet.  That is something I plan on doing tomorrow or early next week as I learn more about Hadoop.  However, when I see errors like that, it is usually my data that I have messed up somehow.  I agree that it would be a good thing to try a smaller dataset and see if the job completes.

 

I’ll post back to the list if I am successful or run into the same/other issues.

 

Chris Snider

Senior Software Engineer

Intelligent Software Solutions, Inc.

Direct (719) 452-7257

Description: Description: Description: cid:image001.png@01CA1F1F.CBC93990

 

From: Barnhart, Bob M. [mailto:ROBERT.M.BARNHART@xxxxxxxxxx]
Sent: Thursday, May 08, 2014 3:26 PM
To: Chris Snider; Discussions between GeoMesa committers
Subject: RE: [geomesa-dev] Problems running GeoMesa GDELT tutorial

 

Chris,

 

Thanks for the fix…it worked fine using an ad-hoc authorization (“GDELT” vs. “COMMA”).

 

I don’t know whether I can collaborate with you directly (as opposed to mailing to geomesa-dev@xxxxxxxxxxxxxxxx) but I thought I’d see if you were willing to take a look at another issue.

 

Still trying to ingest the data loaded into Hadoop from http://data.gdeltproject.org/events/GDELT.MASTERREDUCEDV2.1979-2013.zip, I’m now getting an “ArrayIndexOutOfBoundsException” error (see below).

 

I’m not sure if (1) the data might be at issue, or (2) my Accumulo/Hadoop installation might be at issue. I’m tempted to try loading a smaller dataset, perhaps only (some of) the 2014 files from http://data.gdeltproject.org/events/index.html.

 

Any ideas what might be going on here?

 

Thanks,

Bob Barnhart

 

(ingest log…)

 

-----------------------------------------------------

Running: hadoop jar ./target/geomesa-gdelt-1.0-SNAPSHOT.jar                  geomesa.gdelt.GDELTIngest                                       -instanceId ntc-irad                             -zookeepers localhost:2181                         -user root -password (r00t)                         -auths GDELT                      -tableName gdelt -featureName event                -ingestFile hdfs:///gdelt/uncompressed/gdelt.tsv

-----------------------------------------------------

14/05/08 13:39:09 INFO HSQLDB45DD8FA39A.ENGINE: dataFileCache open start

14/05/08 13:39:09 INFO HSQLDB45DD8FA39A.ENGINE: Checkpoint start

14/05/08 13:39:09 INFO HSQLDB45DD8FA39A.ENGINE: Checkpoint end

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.5-1392090, built on 09/30/2012 17:52 GMT

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:host.name=localhost

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:java.version=1.7.0_55

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:java.home=/usr/lib/jvm/java-1.7.0-openjdk-1.7.0.55.x86_64/jre

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:java.class.path=/usr/local/hadoop-2.4.0/etc/hadoop:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/guava-11.0.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsch-0.1.42.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/activation-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jasper-compiler-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-digester-1.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-el-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/stax-api-1.0-2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/java-xmlbuilder-0.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/junit-4.8.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/avro-1.7.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jettison-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jersey-json-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/xmlenc-0.52.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/zookeeper-3.4.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-beanutils-1.7.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-xc-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/httpclient-4.2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-collections-3.2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-net-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/paranamer-2.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-lang-2.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-beanutils-core-1.8.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/hadoop-auth-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-httpclient-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsp-api-2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-configuration-1.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-math3-3.1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/mockito-all-1.8.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/httpcore-4.2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jets3t-0.9.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/slf4j-api-1.7.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/hadoop-nfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/hadoop-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/hadoop-common-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/guava-11.0.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-el-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/xmlenc-0.52.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jsp-api-2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-nfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/guava-11.0.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/activation-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/leveldbjni-all-1.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/stax-api-1.0-2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/aopalliance-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/guice-servlet-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jettison-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-json-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/zookeeper-3.4.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-xc-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-collections-3.2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/guice-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jline-0.9.94.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-lang-2.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-httpclient-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-client-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/javax.inject-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-guice-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-client-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-tests-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-web-proxy-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-api-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/guice-servlet-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/hamcrest-core-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/guice-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/junit-4.10.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/javax.inject-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-app-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-2.4.0.jar:/usr/local/hadoop-2.4.0/contrib/capacity-scheduler/*.jar

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:java.library.path=/usr/local/hadoop-2.4.0/lib/native

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:java.compiler=<NA>

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:os.name=Linux

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:os.arch=amd64

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:os.version=2.6.32-431.11.2.el6.x86_64

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:user.name=barnhartr

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:user.home=/home/barnhartr

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client environment:user.dir=/usr/local/geomesa-gdelt-master

14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:2181 sessionTimeout=30000 watcher=org.apache.accumulo.fate.zookeeper.ZooSession$ZooWatcher@325c4c8

14/05/08 13:39:11 INFO zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:2181. Will not attempt to authenticate using SASL (unknown error)

14/05/08 13:39:11 INFO zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:2181, initiating session

14/05/08 13:39:11 INFO zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:2181, sessionid = 0x145dd8bc67d0007, negotiated timeout = 30000

OpenJDK 64-Bit Server VM warning: You have loaded library /usr/local/hadoop-2.4.0/lib/native/libhadoop.so.1.0.0 which might have disabled stack guard. The VM will try to fix the stack guard now.

It's highly recommended that you fix the library with 'execstack -c <libfile>', or link it with '-z noexecstack'.

14/05/08 13:39:15 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable

14/05/08 13:39:18 INFO Configuration.deprecation: session.id is deprecated. Instead, use dfs.metrics.session-id

14/05/08 13:39:18 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=JobTracker, sessionId=

14/05/08 13:39:19 WARN mapreduce.JobSubmitter: Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this.

14/05/08 13:39:19 WARN mapreduce.JobSubmitter: No job jar file set.  User classes may not be found. See Job or Job#setJar(String).

14/05/08 13:39:19 INFO input.FileInputFormat: Total input paths to process : 1

14/05/08 13:39:19 INFO mapreduce.JobSubmitter: number of splits:49

14/05/08 13:39:19 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_local1091701005_0001

14/05/08 13:39:19 WARN conf.Configuration: file:/hadoop/tmp/mapred/staging/barnhartr1091701005/.staging/job_local1091701005_0001/job.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval;  Ignoring.

14/05/08 13:39:19 WARN conf.Configuration: file:/hadoop/tmp/mapred/staging/barnhartr1091701005/.staging/job_local1091701005_0001/job.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts;  Ignoring.

14/05/08 13:39:58 INFO mapred.LocalDistributedCacheManager: Creating symlink: /hadoop/tmp/mapred/local/1399581560010/geomesa-gdelt-1.0-SNAPSHOT.jar <- /usr/local/geomesa-gdelt-master/geomesa-gdelt-1.0-SNAPSHOT.jar

14/05/08 13:39:58 INFO mapred.LocalDistributedCacheManager: Localized hdfs://localhost:8020/tmp/geomesa-gdelt-1.0-SNAPSHOT.jar as file:/hadoop/tmp/mapred/local/1399581560010/geomesa-gdelt-1.0-SNAPSHOT.jar

14/05/08 13:39:58 WARN conf.Configuration: file:/hadoop/tmp/mapred/local/localRunner/barnhartr/job_local1091701005_0001/job_local1091701005_0001.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval;  Ignoring.

14/05/08 13:39:58 WARN conf.Configuration: file:/hadoop/tmp/mapred/local/localRunner/barnhartr/job_local1091701005_0001/job_local1091701005_0001.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts;  Ignoring.

14/05/08 13:39:58 INFO mapred.LocalDistributedCacheManager: file:/hadoop/tmp/mapred/local/1399581560010/geomesa-gdelt-1.0-SNAPSHOT.jar/

14/05/08 13:39:58 INFO mapreduce.Job: The url to track the job: http://localhost:8080/

14/05/08 13:39:58 INFO mapreduce.Job: Running job: job_local1091701005_0001

14/05/08 13:39:58 INFO mapred.LocalJobRunner: OutputCommitter set in config null

14/05/08 13:39:58 INFO mapred.LocalJobRunner: OutputCommitter is org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter

14/05/08 13:39:59 INFO mapred.LocalJobRunner: Waiting for map tasks

14/05/08 13:39:59 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000000_0

14/05/08 13:39:59 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:39:59 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6442450944+138958463

14/05/08 13:39:59 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:39:59 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:39:59 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:39:59 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:39:59 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:39:59 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:39:59 INFO mapreduce.Job: Job job_local1091701005_0001 running in uber mode : false

14/05/08 13:39:59 INFO mapreduce.Job:  map 0% reduce 0%

14/05/08 13:40:01 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:01 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000001_0

14/05/08 13:40:01 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:01 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:0+134217728

14/05/08 13:40:01 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:01 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:01 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:01 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:01 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:01 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:01 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:01 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000002_0

14/05/08 13:40:01 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:01 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:134217728+134217728

14/05/08 13:40:01 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:01 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:01 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:01 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:01 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:01 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:02 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:02 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000003_0

14/05/08 13:40:02 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:02 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:268435456+134217728

14/05/08 13:40:02 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:02 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:02 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:02 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:02 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:02 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:02 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:02 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000004_0

14/05/08 13:40:02 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:02 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:402653184+134217728

14/05/08 13:40:02 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:02 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:02 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:02 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:02 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:02 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:02 WARN impl.ThriftTransportPool: Server 127.0.0.1:9997:9997 (120000) had 20 failures in a short time period, will not complain anymore

14/05/08 13:40:02 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:02 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000005_0

14/05/08 13:40:02 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:02 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:536870912+134217728

14/05/08 13:40:02 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:02 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:02 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:02 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:02 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:02 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:03 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:03 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000006_0

14/05/08 13:40:03 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:03 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:671088640+134217728

14/05/08 13:40:03 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:03 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:03 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:03 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:03 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:03 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:03 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:03 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000007_0

14/05/08 13:40:03 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:03 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:805306368+134217728

14/05/08 13:40:03 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:03 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:03 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:03 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:03 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:03 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:03 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:03 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000008_0

14/05/08 13:40:03 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:03 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:939524096+134217728

14/05/08 13:40:03 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:03 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:03 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:03 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:03 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:03 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:03 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:03 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000009_0

14/05/08 13:40:03 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:03 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1073741824+134217728

14/05/08 13:40:03 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:03 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:03 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:03 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:03 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:03 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:04 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:04 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000010_0

14/05/08 13:40:04 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:04 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1207959552+134217728

14/05/08 13:40:04 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:04 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:04 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:04 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:04 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:04 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:04 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:04 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000011_0

14/05/08 13:40:04 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:04 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1342177280+134217728

14/05/08 13:40:04 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:04 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:04 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:04 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:04 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:04 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:04 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:04 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000012_0

14/05/08 13:40:04 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:04 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1476395008+134217728

14/05/08 13:40:04 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:04 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:04 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:04 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:04 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:04 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:04 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:04 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000013_0

14/05/08 13:40:04 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:04 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1610612736+134217728

14/05/08 13:40:04 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:04 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:04 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:04 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:04 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:04 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:05 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:05 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000014_0

14/05/08 13:40:05 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:05 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1744830464+134217728

14/05/08 13:40:05 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:05 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:05 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:05 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:05 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:05 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:05 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:05 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:05 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000015_0

14/05/08 13:40:05 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:05 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1879048192+134217728

14/05/08 13:40:05 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:05 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:05 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:05 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:05 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:05 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:05 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:05 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000016_0

14/05/08 13:40:05 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:05 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2013265920+134217728

14/05/08 13:40:05 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:05 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:05 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:05 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:05 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:05 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:05 INFO mapreduce.Job:  map 2% reduce 0%

14/05/08 13:40:05 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:05 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000017_0

14/05/08 13:40:05 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:05 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2147483648+134217728

14/05/08 13:40:05 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:06 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:06 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:06 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:06 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:06 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:06 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:06 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000018_0

14/05/08 13:40:06 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:06 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2281701376+134217728

14/05/08 13:40:06 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:06 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:06 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:06 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:06 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:06 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:06 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:06 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000019_0

14/05/08 13:40:06 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:06 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2415919104+134217728

14/05/08 13:40:06 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:06 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:06 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:06 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:06 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:06 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:06 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:06 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000020_0

14/05/08 13:40:06 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:06 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2550136832+134217728

14/05/08 13:40:06 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:06 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:06 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:06 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:06 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:06 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:06 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:06 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000021_0

14/05/08 13:40:06 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:06 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2684354560+134217728

14/05/08 13:40:06 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:07 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:07 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:07 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:07 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:07 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:07 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:07 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:07 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000022_0

14/05/08 13:40:07 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:07 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2818572288+134217728

14/05/08 13:40:07 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:07 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:07 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:07 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:07 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:07 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:07 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:07 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000023_0

14/05/08 13:40:07 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:07 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2952790016+134217728

14/05/08 13:40:07 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:07 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:07 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:07 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:07 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:07 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:07 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:07 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000024_0

14/05/08 13:40:07 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:07 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3087007744+134217728

14/05/08 13:40:07 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:07 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:07 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:07 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:07 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:07 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:07 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:07 INFO mapreduce.Job:  map 6% reduce 0%

14/05/08 13:40:07 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:07 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000025_0

14/05/08 13:40:07 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:07 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3221225472+134217728

14/05/08 13:40:07 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:08 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:08 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:08 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:08 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:08 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:08 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:08 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:08 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000026_0

14/05/08 13:40:08 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:08 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3355443200+134217728

14/05/08 13:40:08 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:08 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:08 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:08 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:08 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:08 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:08 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:08 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:08 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000027_0

14/05/08 13:40:08 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:08 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3489660928+134217728

14/05/08 13:40:08 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:08 INFO mapreduce.Job:  map 10% reduce 0%

14/05/08 13:40:08 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:08 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:08 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:08 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:08 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:08 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:09 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:09 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000028_0

14/05/08 13:40:09 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:09 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3623878656+134217728

14/05/08 13:40:09 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:09 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:09 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:09 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:09 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:09 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:09 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:09 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:09 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:09 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000029_0

14/05/08 13:40:09 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:09 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3758096384+134217728

14/05/08 13:40:09 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:09 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:09 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:09 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:09 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:09 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:09 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:09 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:09 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000030_0

14/05/08 13:40:09 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:09 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3892314112+134217728

14/05/08 13:40:09 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:09 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:09 INFO mapreduce.Job:  map 20% reduce 0%

14/05/08 13:40:10 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:10 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:10 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:10 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:10 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:10 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:10 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:10 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000031_0

14/05/08 13:40:10 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:10 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4026531840+134217728

14/05/08 13:40:10 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:10 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:10 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:10 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:10 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:10 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:10 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:10 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:10 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000032_0

14/05/08 13:40:10 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:10 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4160749568+134217728

14/05/08 13:40:10 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:10 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:10 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:10 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:10 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:10 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:10 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:10 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:10 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000033_0

14/05/08 13:40:10 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:10 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4294967296+134217728

14/05/08 13:40:10 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:10 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:10 INFO mapreduce.Job:  map 29% reduce 0%

14/05/08 13:40:10 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:10 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:10 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:10 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:10 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:11 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:11 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000034_0

14/05/08 13:40:11 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:11 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4429185024+134217728

14/05/08 13:40:11 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:11 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:11 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:11 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:11 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:11 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:11 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:11 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:11 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:11 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000035_0

14/05/08 13:40:11 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:11 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4563402752+134217728

14/05/08 13:40:11 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:11 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:11 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:11 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:11 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:11 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:11 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:11 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:11 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000036_0

14/05/08 13:40:11 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:11 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4697620480+134217728

14/05/08 13:40:11 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:11 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:11 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:11 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:11 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:11 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:11 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:11 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000037_0

14/05/08 13:40:11 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:11 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4831838208+134217728

14/05/08 13:40:11 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:12 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:12 INFO mapreduce.Job:  map 37% reduce 0%

14/05/08 13:40:12 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:12 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:12 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:12 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:12 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:12 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:12 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000038_0

14/05/08 13:40:12 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:12 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4966055936+134217728

14/05/08 13:40:12 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:12 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:12 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:12 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:12 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:12 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:12 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:12 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:12 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000039_0

14/05/08 13:40:12 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:12 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5100273664+134217728

14/05/08 13:40:12 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:12 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:12 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:12 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:12 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:12 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:12 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:12 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:12 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000040_0

14/05/08 13:40:12 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:12 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5234491392+134217728

14/05/08 13:40:12 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:12 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:12 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:12 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:12 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:12 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:12 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:12 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:12 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000041_0

14/05/08 13:40:12 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:12 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5368709120+134217728

14/05/08 13:40:12 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:12 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:12 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:12 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:12 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:12 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:13 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:13 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000042_0

14/05/08 13:40:13 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:13 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5502926848+134217728

14/05/08 13:40:13 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:13 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:13 INFO mapreduce.Job:  map 45% reduce 0%

14/05/08 13:40:13 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:13 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:13 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:13 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:13 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:13 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:13 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000043_0

14/05/08 13:40:13 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:13 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5637144576+134217728

14/05/08 13:40:13 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:13 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:13 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:13 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:13 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:13 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:13 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:13 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:13 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000044_0

14/05/08 13:40:13 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:13 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5771362304+134217728

14/05/08 13:40:13 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:13 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:13 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:13 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:13 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:13 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:13 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:13 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:13 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000045_0

14/05/08 13:40:13 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:13 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5905580032+134217728

14/05/08 13:40:13 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:13 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:13 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:13 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:13 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:13 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:13 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:13 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:13 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000046_0

14/05/08 13:40:13 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:13 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6039797760+134217728

14/05/08 13:40:13 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:14 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:14 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:14 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:14 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:14 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:14 INFO mapreduce.Job:  map 51% reduce 0%

14/05/08 13:40:14 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:14 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000047_0

14/05/08 13:40:14 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:14 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6174015488+134217728

14/05/08 13:40:14 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:14 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:14 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:14 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:14 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:14 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:14 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:14 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:14 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:14 INFO mapred.LocalJobRunner: Starting task: attempt_local1091701005_0001_m_000048_0

14/05/08 13:40:14 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 13:40:14 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6308233216+134217728

14/05/08 13:40:14 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 13:40:14 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 13:40:14 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 13:40:14 INFO mapred.MapTask: soft limit at 83886080

14/05/08 13:40:14 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 13:40:14 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 13:40:14 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:14 INFO mapred.MapTask: Starting flush of map output

14/05/08 13:40:14 INFO mapred.LocalJobRunner: map task executor complete.

14/05/08 13:40:14 WARN mapred.LocalJobRunner: job_local1091701005_0001

java.lang.Exception: java.lang.ArrayIndexOutOfBoundsException: 39

                at org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462)

                at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:522)

Caused by: java.lang.ArrayIndexOutOfBoundsException: 39

                at geomesa.gdelt.GDELTIngestMapper.map(GDELTIngestMapper.java:60)

                at geomesa.gdelt.GDELTIngestMapper.map(GDELTIngestMapper.java:27)

                at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145)

                at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)

                at org.apache.hadoop.mapred.MapTask.run(MapTask.java:340)

                at org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:243)

                at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)

                at java.util.concurrent.FutureTask.run(FutureTask.java:262)

                at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)

                at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)

                at java.lang.Thread.run(Thread.java:744)

14/05/08 13:40:15 INFO mapreduce.Job:  map 57% reduce 0%

14/05/08 13:40:15 INFO mapreduce.Job: Job job_local1091701005_0001 failed with state FAILED due to: NA

14/05/08 13:40:15 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:15 INFO mapreduce.Job: Counters: 25

                File System Counters

                                FILE: Number of bytes read=1100796112

                                FILE: Number of bytes written=1112405102

                                FILE: Number of read operations=0

                                FILE: Number of large read operations=0

                                FILE: Number of write operations=0

                                HDFS: Number of bytes read=1101293020

                                HDFS: Number of bytes written=1097385436

                                HDFS: Number of read operations=1465

                                HDFS: Number of large read operations=0

                                HDFS: Number of write operations=112

                Map-Reduce Framework

                                Map input records=28

                                Map output records=0

                                Map output bytes=0

                                Map output materialized bytes=168

                                Input split bytes=3220

                                Combine input records=0

                                Spilled Records=0

                                Failed Shuffles=0

                                Merged Map outputs=0

                                GC time elapsed (ms)=70902

                                CPU time spent (ms)=0

                                Physical memory (bytes) snapshot=0

                                Virtual memory (bytes) snapshot=0

                                Total committed heap usage (bytes)=6046191616

                File Input Format Counters

                                Bytes Read=114688

Exception in thread "main" java.lang.Exception: Job failed

                at geomesa.gdelt.GDELTIngest.runMapReduceJob(GDELTIngest.java:152)

                at geomesa.gdelt.GDELTIngest.main(GDELTIngest.java:110)

                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

                at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

                at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

                at java.lang.reflect.Method.invoke(Method.java:606)

                at org.apache.hadoop.util.RunJar.main(RunJar.java:212)

14/05/08 13:40:15 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:15 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:16 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:16 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:16 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:17 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:17 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:17 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:18 INFO mapred.LocalJobRunner: map > sort

14/05/08 13:40:18 INFO mapred.LocalJobRunner: map > sort

 

From: Chris Snider [mailto:chris.snider@xxxxxxxxxx]
Sent: Thursday, May 08, 2014 1:01 PM
To: Discussions between GeoMesa committers; Barnhart, Bob M.
Subject: RE: [geomesa-dev] Problems running GeoMesa GDELT tutorial

 

Bob,

 

I think there may be a misunderstanding between the DB permissions and the user authorizations inherent in the Accumulo Schema.

 

In an accumulo shell, run a

getauths –u root

 

This is the Authorizations that the connector is expecting. 

 

You can set authorizations for the root user with the following accumulo shell

setauths –s “MY,COMMA,DELIMITED,AUTHS” –u root

 

Use the new auths in the connection block for -auths COMMA (for example)

 

Chris Snider

Senior Software Engineer

Intelligent Software Solutions, Inc.

Description: Description: Description: cid:image001.png@01CA1F1F.CBC93990

 

From: geomesa-dev-bounces@xxxxxxxxxxxxxxxx [mailto:geomesa-dev-bounces@xxxxxxxxxxxxxxxx] On Behalf Of Andrew Ross
Sent: Thursday, May 08, 2014 1:47 PM
To: Barnhart, Bob M.; geomesa-dev@xxxxxxxxxxxxxxxx
Subject: Re: [geomesa-dev] Problems running GeoMesa GDELT tutorial

 

Hi Bob,

Thank you so much for your interest in Geomesa.

I'll check into what's up with geomesa-user. In the meantime, I've added geomesa-dev, which has the developer team on it.

Cheers!

Andrew

On 08/05/14 21:39, Barnhart, Bob M. wrote:

I’ve run into problems trying to work through the GeoMesa GDELT Analysis tutorial at http://geomesa.github.io/2014/04/17/geomesa-gdelt-analysis/ . I sent the following email to ‘geomesa-user@xxxxxxxxxxxxxxxx’ but it bounced with a non-existent user error. I don’t know if there are any GeoMesa communities of interest to whom I could send a description of my problem, so I’m directing my question to you in hopes that you could provide a (simple) solution, or direct me to a person/site where I might find an answer.

 

I am running Accumulo 1.5.1, Hadoop 2.4.0 and Zookeeper 3.4.6.

 

I’ve been able to load the GDELT data file http://data.gdeltproject.org/events/GDELT.MASTERREDUCEDV2.1979-2013.zip into Hadoop and am trying to ingest this data into Accumulo using the Hadoop command in the tutorial.

 

As shown in the execution trace below, the ingest process fails with the error:

 

java.lang.Exception: java.lang.RuntimeException: org.apache.accumulo.core.client.AccumuloSecurityException: Error BAD_AUTHORIZATIONS for user root - The user does not have the specified authorizations assigned

 

I don’t know to what “specified authorizations” this error might be referring. As shown below, the Accumulo “root” user has all possible System.* and Table.* permissions, including the ‘gdelt’ table:

 

$ accumulo shell –u root

root@ntc-irad> userpermissions -u root

System permissions: System.GRANT, System.CREATE_TABLE, System.DROP_TABLE, System.ALTER_TABLE, System.CREATE_USER, System.DROP_USER, System.ALTER_USER, System.SYSTEM

 

Table permissions (!METADATA): Table.READ, Table.ALTER_TABLE

Table permissions (gdelt): Table.READ, Table.WRITE, Table.BULK_IMPORT, Table.ALTER_TABLE, Table.GRANT, Table.DROP_TABLE

Table permissions (trace): Table.READ, Table.WRITE, Table.BULK_IMPORT, Table.ALTER_TABLE, Table.GRANT, Table.DROP_TABLE

 

I would be grateful for any assistance in getting the GDELT data ingesting into Accumulo so that I could complete my assessment of GeoMesa via the GDELT tutorial.

 

Best regards,

Bob Barnhart

Chief Systems Engineer | 858 826 5596 (Office) | 619 972 9489 (Mobile) | barnhartr@xxxxxxxxxx

 

-----------------------------------------------------

Running:   hadoop jar ./target/geomesa-gdelt-1.0-SNAPSHOT.jar                

geomesa.gdelt.GDELTIngest                                          

-instanceId ntc-irad                            

-zookeepers 127.0.0.1                      

-user root -password (r00t)           

 -auths Table.READ,Table.WRITE,Table.BULK_IMPORT,Table.ALTER_TABLE,Table.GRANT,Table.DROP_TABLE       

 -tableName gdelt -featureName event                  

 -ingestFile hdfs:///gdelt/uncompressed/gdelt.tsv

-----------------------------------------------------

14/05/08 11:52:58 INFO HSQLDB45DD2E6EE0.ENGINE: dataFileCache open start

14/05/08 11:52:59 INFO HSQLDB45DD2E6EE0.ENGINE: Checkpoint start

14/05/08 11:52:59 INFO HSQLDB45DD2E6EE0.ENGINE: Checkpoint end

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.5-1392090, built on 09/30/2012 17:52 GMT

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:host.name=localhost

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:java.version=1.7.0_55

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:java.home=/usr/lib/jvm/java-1.7.0-openjdk-1.7.0.55.x86_64/jre

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:java.class.path=/usr/local/hadoop-2.4.0/etc/hadoop:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/guava-11.0.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsch-0.1.42.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/activation-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jasper-compiler-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-digester-1.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-el-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/stax-api-1.0-2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/java-xmlbuilder-0.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/junit-4.8.2.jar:/usr/local/hadoop-2.4.0/share/had oop/commo n/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/avro-1.7.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jettison-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jersey-json-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/xmlenc-0.52.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/zookeeper-3.4.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-beanutils-1.7.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-xc-1.8.8.jar:/usr/lo cal/hadoo p-2.4.0/share/hadoop/common/lib/httpclient-4.2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-collections-3.2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-net-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/paranamer-2.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-lang-2.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-beanutils-core-1.8.0.jar:/usr/local/hadoop-2.4.0 /share/ha doop/common/lib/hadoop-auth-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-httpclient-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsp-api-2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-configuration-1.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-math3-3.1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/mockito-all-1.8.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/httpcore-4.2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jets3t-0.9.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/slf4j-api-1.7.5.jar:/usr/local/hadoop-2.4.0/share/hadoo p/common/ hadoop-nfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/hadoop-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/hadoop-common-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/guava-11.0.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-el-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/xmlenc-0.52.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/log4 j-1.2.17. jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jsp-api-2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-nfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/g uava-11.0 .2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/activation-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/leveldbjni-all-1.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/stax-api-1.0-2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/aopalliance-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/guice-servlet-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jettison-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib /jersey-j son-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/zookeeper-3.4.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-xc-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-collections-3.2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/guice-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jline-0.9.94.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-lan g-2.6.jar :/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-httpclient-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-client-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/javax.inject-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-guice-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-client-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-tests-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-2.4.0.jar:/usr/local /hadoop-2 .4.0/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-web-proxy-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-api-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/usr/local/hadoop-2.4.0/share/h adoop/map reduce/lib/guice-servlet-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/hamcrest-core-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/guice-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/junit-4.10.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/javax.inject-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/usr/local/hadoop-2.4.0/share/h adoop/map reduce/hadoop-mapreduce-client-jobclient-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-app-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-2.4.0.jar:/usr/local/hadoop-2.4.0/contrib/capacity-scheduler/*.jar

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:java.library.path=/usr/local/hadoop-2.4.0/lib/native

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:java.compiler=<NA>

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:os.name=Linux

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:os.arch=amd64

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:os.version=2.6.32-431.11.2.el6.x86_64

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:user.name=barnhartr

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:user.home=/home/barnhartr

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client environment:user.dir=/usr/local/geomesa-gdelt-master

14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Initiating client connection, connectString=127.0.0.1 sessionTimeout=30000 watcher=org.apache.accumulo.fate.zookeeper.ZooSession$ZooWatcher@1ec896d2

14/05/08 11:53:01 INFO zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:2181. Will not attempt to authenticate using SASL (unknown error)

14/05/08 11:53:01 INFO zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:2181, initiating session

14/05/08 11:53:01 INFO zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:2181, sessionid = 0x145dc8e7394000e, negotiated timeout = 30000

OpenJDK 64-Bit Server VM warning: You have loaded library /usr/local/hadoop-2.4.0/lib/native/libhadoop.so.1.0.0 which might have disabled stack guard. The VM will try to fix the stack guard now.

It's highly recommended that you fix the library with 'execstack -c <libfile>', or link it with '-z noexecstack'.

14/05/08 11:53:05 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable

14/05/08 11:53:08 INFO Configuration.deprecation: session.id is deprecated. Instead, use dfs.metrics.session-id

14/05/08 11:53:08 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=JobTracker, sessionId=

14/05/08 11:53:08 WARN mapreduce.JobSubmitter: Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this.

14/05/08 11:53:08 WARN mapreduce.JobSubmitter: No job jar file set.  User classes may not be found. See Job or Job#setJar(String).

14/05/08 11:53:08 INFO input.FileInputFormat: Total input paths to process : 1

14/05/08 11:53:08 INFO mapreduce.JobSubmitter: number of splits:49

14/05/08 11:53:09 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_local422695915_0001

14/05/08 11:53:09 WARN conf.Configuration: file:/hadoop/tmp/mapred/staging/barnhartr422695915/.staging/job_local422695915_0001/job.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval;  Ignoring.

14/05/08 11:53:09 WARN conf.Configuration: file:/hadoop/tmp/mapred/staging/barnhartr422695915/.staging/job_local422695915_0001/job.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts;  Ignoring.

14/05/08 11:53:47 INFO mapred.LocalDistributedCacheManager: Creating symlink: /hadoop/tmp/mapred/local/1399575189575/geomesa-gdelt-1.0-SNAPSHOT.jar <- /usr/local/geomesa-gdelt-master/geomesa-gdelt-1.0-SNAPSHOT.jar

14/05/08 11:53:47 INFO mapred.LocalDistributedCacheManager: Localized hdfs://localhost:8020/tmp/geomesa-gdelt-1.0-SNAPSHOT.jar as file:/hadoop/tmp/mapred/local/1399575189575/geomesa-gdelt-1.0-SNAPSHOT.jar

14/05/08 11:53:47 WARN conf.Configuration: file:/hadoop/tmp/mapred/local/localRunner/barnhartr/job_local422695915_0001/job_local422695915_0001.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval;  Ignoring.

14/05/08 11:53:47 WARN conf.Configuration: file:/hadoop/tmp/mapred/local/localRunner/barnhartr/job_local422695915_0001/job_local422695915_0001.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts;  Ignoring.

14/05/08 11:53:47 INFO mapred.LocalDistributedCacheManager: file:/hadoop/tmp/mapred/local/1399575189575/geomesa-gdelt-1.0-SNAPSHOT.jar/

14/05/08 11:53:47 INFO mapreduce.Job: The url to track the job: http://localhost:8080/

14/05/08 11:53:47 INFO mapreduce.Job: Running job: job_local422695915_0001

14/05/08 11:53:47 INFO mapred.LocalJobRunner: OutputCommitter set in config null

14/05/08 11:53:47 INFO mapred.LocalJobRunner: OutputCommitter is org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter

14/05/08 11:53:47 INFO mapred.LocalJobRunner: Waiting for map tasks

14/05/08 11:53:47 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000000_0

14/05/08 11:53:47 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:47 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6442450944+138958463

14/05/08 11:53:47 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:48 INFO mapreduce.Job: Job job_local422695915_0001 running in uber mode : false

14/05/08 11:53:48 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:48 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:48 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:48 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:48 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:48 INFO mapreduce.Job:  map 0% reduce 0%

14/05/08 11:53:49 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:49 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000001_0

14/05/08 11:53:49 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:49 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:0+134217728

14/05/08 11:53:49 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:49 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:49 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:49 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:49 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:49 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:49 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:49 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:49 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000002_0

14/05/08 11:53:49 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:49 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:134217728+134217728

14/05/08 11:53:49 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:49 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:49 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:49 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:49 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:49 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:49 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:49 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:49 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000003_0

14/05/08 11:53:49 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:49 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:268435456+134217728

14/05/08 11:53:49 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:49 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:49 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:49 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:49 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:49 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:49 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:49 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:49 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000004_0

14/05/08 11:53:49 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:49 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:402653184+134217728

14/05/08 11:53:49 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:49 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:49 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:49 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:49 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:49 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:49 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:49 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:49 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000005_0

14/05/08 11:53:49 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:49 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:536870912+134217728

14/05/08 11:53:49 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:49 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:49 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:49 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:49 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:49 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:49 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000006_0

14/05/08 11:53:50 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:50 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:671088640+134217728

14/05/08 11:53:50 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:50 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:50 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000007_0

14/05/08 11:53:50 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:50 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:805306368+134217728

14/05/08 11:53:50 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:50 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:50 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000008_0

14/05/08 11:53:50 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:50 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:939524096+134217728

14/05/08 11:53:50 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:50 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:50 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000009_0

14/05/08 11:53:50 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:50 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1073741824+134217728

14/05/08 11:53:50 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:50 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:50 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000010_0

14/05/08 11:53:50 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:50 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1207959552+134217728

14/05/08 11:53:50 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:50 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:50 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000011_0

14/05/08 11:53:50 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:50 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1342177280+134217728

14/05/08 11:53:50 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:50 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:50 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000012_0

14/05/08 11:53:50 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:50 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1476395008+134217728

14/05/08 11:53:50 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:51 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:51 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000013_0

14/05/08 11:53:51 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:51 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1610612736+134217728

14/05/08 11:53:51 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:51 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:51 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000014_0

14/05/08 11:53:51 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:51 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1744830464+134217728

14/05/08 11:53:51 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:51 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:51 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000015_0

14/05/08 11:53:51 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:51 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1879048192+134217728

14/05/08 11:53:51 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:51 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:51 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000016_0

14/05/08 11:53:51 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:51 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2013265920+134217728

14/05/08 11:53:51 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:51 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:51 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000017_0

14/05/08 11:53:51 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:51 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2147483648+134217728

14/05/08 11:53:51 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:51 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:51 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000018_0

14/05/08 11:53:51 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:51 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2281701376+134217728

14/05/08 11:53:51 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:52 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:52 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000019_0

14/05/08 11:53:52 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:52 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2415919104+134217728

14/05/08 11:53:52 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:52 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:52 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000020_0

14/05/08 11:53:52 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:52 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2550136832+134217728

14/05/08 11:53:52 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:52 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:52 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000021_0

14/05/08 11:53:52 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:52 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2684354560+134217728

14/05/08 11:53:52 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:52 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:52 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000022_0

14/05/08 11:53:52 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:52 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2818572288+134217728

14/05/08 11:53:52 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:52 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:52 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000023_0

14/05/08 11:53:52 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:52 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2952790016+134217728

14/05/08 11:53:52 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:52 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:52 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000024_0

14/05/08 11:53:52 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:52 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3087007744+134217728

14/05/08 11:53:52 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:53 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:53 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000025_0

14/05/08 11:53:53 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:53 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3221225472+134217728

14/05/08 11:53:53 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:53 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:53 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000026_0

14/05/08 11:53:53 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:53 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3355443200+134217728

14/05/08 11:53:53 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:53 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:53 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000027_0

14/05/08 11:53:53 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:53 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3489660928+134217728

14/05/08 11:53:53 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:53 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:53 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000028_0

14/05/08 11:53:53 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:53 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3623878656+134217728

14/05/08 11:53:53 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:53 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:53 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000029_0

14/05/08 11:53:53 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:53 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3758096384+134217728

14/05/08 11:53:53 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:53 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:53 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000030_0

14/05/08 11:53:53 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:53 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3892314112+134217728

14/05/08 11:53:53 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:54 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:54 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000031_0

14/05/08 11:53:54 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:54 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4026531840+134217728

14/05/08 11:53:54 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:54 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:54 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000032_0

14/05/08 11:53:54 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:54 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4160749568+134217728

14/05/08 11:53:54 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:54 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:54 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000033_0

14/05/08 11:53:54 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:54 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4294967296+134217728

14/05/08 11:53:54 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:54 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:54 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000034_0

14/05/08 11:53:54 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:54 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4429185024+134217728

14/05/08 11:53:54 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:54 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:54 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000035_0

14/05/08 11:53:54 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:54 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4563402752+134217728

14/05/08 11:53:54 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:54 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:54 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000036_0

14/05/08 11:53:54 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:54 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4697620480+134217728

14/05/08 11:53:54 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:54 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:54 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000037_0

14/05/08 11:53:55 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:55 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4831838208+134217728

14/05/08 11:53:55 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:55 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:55 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000038_0

14/05/08 11:53:55 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:55 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4966055936+134217728

14/05/08 11:53:55 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:55 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:55 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000039_0

14/05/08 11:53:55 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:55 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5100273664+134217728

14/05/08 11:53:55 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:55 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:55 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000040_0

14/05/08 11:53:55 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:55 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5234491392+134217728

14/05/08 11:53:55 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:55 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:55 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000041_0

14/05/08 11:53:55 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:55 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5368709120+134217728

14/05/08 11:53:55 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:55 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:55 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000042_0

14/05/08 11:53:55 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:55 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5502926848+134217728

14/05/08 11:53:55 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:55 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:55 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000043_0

14/05/08 11:53:56 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:56 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5637144576+134217728

14/05/08 11:53:56 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:56 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:56 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:56 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:56 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:56 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:56 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000044_0

14/05/08 11:53:56 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:56 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5771362304+134217728

14/05/08 11:53:56 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:56 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:56 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:56 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:56 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:56 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:56 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000045_0

14/05/08 11:53:56 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:56 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5905580032+134217728

14/05/08 11:53:56 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:56 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:56 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:56 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:56 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:56 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:56 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000046_0

14/05/08 11:53:56 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:56 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6039797760+134217728

14/05/08 11:53:56 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:56 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:56 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:56 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:56 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:56 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:56 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000047_0

14/05/08 11:53:56 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:56 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6174015488+134217728

14/05/08 11:53:56 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:56 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:56 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:56 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:56 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:56 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:56 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task: attempt_local422695915_0001_m_000048_0

14/05/08 11:53:56 INFO mapred.Task:  Using ResourceCalculatorProcessTree : [ ]

14/05/08 11:53:56 INFO mapred.MapTask: Processing split: hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6308233216+134217728

14/05/08 11:53:56 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer

14/05/08 11:53:57 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:57 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)

14/05/08 11:53:57 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100

14/05/08 11:53:57 INFO mapred.MapTask: soft limit at 83886080

14/05/08 11:53:57 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600

14/05/08 11:53:57 INFO mapred.MapTask: kvstart = 26214396; length = 6553600

14/05/08 11:53:57 INFO mapred.MapTask: Starting flush of map output

14/05/08 11:53:57 INFO mapred.LocalJobRunner: map task executor complete.

14/05/08 11:53:57 WARN mapred.LocalJobRunner: job_local422695915_0001

java.lang.Exception: java.lang.RuntimeException: org.apache.accumulo.core.client.AccumuloSecurityException: Error BAD_AUTHORIZATIONS for user root - The user does not have the specified authorizations assigned

                at org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462)

                at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:522)

Caused by: java.lang.RuntimeException: org.apache.accumulo.core.client.AccumuloSecurityException: Error BAD_AUTHORIZATIONS for user root - The user does not have the specified authorizations assigned

                at org.apache.accumulo.core.client.impl.TabletServerBatchReaderIterator.hasNext(TabletServerBatchReaderIterator.java:210)

                at geomesa.core.data.AccumuloDataStore$$anonfun$readMetadataItem$1.apply(AccumuloDataStore.scala:169)

                at geomesa.core.data.AccumuloDataStore$$anonfun$readMetadataItem$1.apply(AccumuloDataStore.scala:157)

                at scala.collection.MapLike$class.getOrElse(MapLike.scala:128)

                at scala.collection.AbstractMap.getOrElse(Map.scala:58)

                at geomesa.core.data.AccumuloDataStore.readMetadataItem(AccumuloDataStore.scala:157)

                at geomesa.core.data.AccumuloDataStore.getAttributes(AccumuloDataStore.scala:220)

                at geomesa.core.data.AccumuloDataStore.getSchema(AccumuloDataStore.scala:267)

                at geomesa.gdelt.GDELTIngestMapper.setup(GDELTIngestMapper.java:53)

                at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142)

                at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)

                at org.apache.hadoop.mapred.MapTask.run(MapTask.java:340)

                at org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:243)

                at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)

                at java.util.concurrent.FutureTask.run(FutureTask.java:262)

                at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)

                at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)

                at java.lang.Thread.run(Thread.java:744)

Caused by: org.apache.accumulo.core.client.AccumuloSecurityException: Error BAD_AUTHORIZATIONS for user root - The user does not have the specified authorizations assigned

                at org.apache.accumulo.core.client.impl.TabletServerBatchReaderIterator.doLookup(TabletServerBatchReaderIterator.java:701)

                at org.apache.accumulo.core.client.impl.TabletServerBatchReaderIterator$QueryTask.run(TabletServerBatchReaderIterator.java:361)

                at org.apache.accumulo.trace.instrument.TraceRunnable.run(TraceRunnable.java:47)

                at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)

                at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)

                at org.apache.accumulo.trace.instrument.TraceRunnable.run(TraceRunnable.java:47)

                at org.apache.accumulo.core.util.LoggingRunnable.run(LoggingRunnable.java:34)

                ... 1 more

Caused by: ThriftSecurityException(user:root, code:BAD_AUTHORIZATIONS)

                at org.apache.accumulo.core.tabletserver.thrift.TabletClientService$startMultiScan_result$startMultiScan_resultStandardScheme.read(TabletClientService.java:8165)

                at org.apache.accumulo.core.tabletserver.thrift.TabletClientService$startMultiScan_result$startMultiScan_resultStandardScheme.read(TabletClientService.java:8142)

                at org.apache.accumulo.core.tabletserver.thrift.TabletClientService$startMultiScan_result.read(TabletClientService.java:8081)

                at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78)

                at org.apache.accumulo.core.tabletserver.thrift.TabletClientService$Client.recv_startMultiScan(TabletClientService.java:294)

                at org.apache.accumulo.core.tabletserver.thrift.TabletClientService$Client.startMultiScan(TabletClientService.java:274)

                at org.apache.accumulo.core.client.impl.TabletServerBatchReaderIterator.doLookup(TabletServerBatchReaderIterator.java:644)

                ... 7 more

14/05/08 11:53:57 WARN impl.TabletServerBatchReader: TabletServerBatchReader not shutdown; did you forget to call close()?

14/05/08 11:53:58 INFO mapreduce.Job: Job job_local422695915_0001 failed with state FAILED due to: NA

14/05/08 11:53:58 INFO mapreduce.Job: Counters: 0

Exception in thread "main" java.lang.Exception: Job failed

                at geomesa.gdelt.GDELTIngest.runMapReduceJob(GDELTIngest.java:152)

                at geomesa.gdelt.GDELTIngest.main(GDELTIngest.java:110)

                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

                at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

                at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

                at java.lang.reflect.Method.invoke(Method.java:606)

                at org.apache.hadoop.util.RunJar.main(RunJar.java:212)

 


Back to the top