Skip to main content

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [List Home]
Re: [geomesa-dev] Problems running GeoMesa GDELT tutorial

Chris Snider,

Well spotted.  We will take a look at this today, and post back when we
have a plan.

Thanks!

Sincerely,
  -- Chris Eichelberger


On Mon, 2014-05-12 at 13:15 +0000, Chris Snider wrote:
> Bob, Chris E, et al,
> 
>  
> 
> I grabbed the link that Bob used
> (http://data.gdeltproject.org/events/GDELT.MASTERREDUCEDV2.1979-2013.zip) and it is a reduced data set that only has 17 columns.  The main historical back files available at http://data.gdeltproject.org/events/index.html have 57 columns.
> 
>  
> 
> Maybe the GeoMesa  GDELT ingestion project can add a –reduced=[true |
> false] which then operated on the reduced file vice a full historical
> backfile?
> 
>  
> 
> On a side note, the geomesa-gdelt project readme on github indicates
> that “…ingestFile be a gdelt format TSV …” meaning that they should be
> Tab separated values.  However, all the downloads I have found in a
> quick perusal of the gdelt project download page are all CSV
> formatted.  Is the geomesa-gdelt readme in error and the files really
> are CSV, or should the CSV formats be mapped to TSV before running
> through the geomesa ingestion?
> 
>  
> 
> Thanks,
> 
>  
> 
> Chris Snider
> 
> Senior Software Engineer
> 
> Intelligent Software Solutions, Inc.
> 
> Direct (719) 452-7257
> 
> Description: Description: Description:
> cid:image001.png@01CA1F1F.CBC93990
> 
> 
>  
> 
> From: Barnhart, Bob M. [mailto:ROBERT.M.BARNHART@xxxxxxxxxx] 
> Sent: Thursday, May 08, 2014 4:45 PM
> To: Chris Snider; Discussions between GeoMesa committers
> Subject: RE: [geomesa-dev] Problems running GeoMesa GDELT tutorial
> 
> 
>  
> 
> Thanks Chris. Good luck!
> 
>  
> 
> From: Chris Snider [mailto:chris.snider@xxxxxxxxxx] 
> Sent: Thursday, May 08, 2014 3:21 PM
> To: Barnhart, Bob M.; Discussions between GeoMesa committers
> Subject: RE: [geomesa-dev] Problems running GeoMesa GDELT tutorial
> 
> 
>  
> 
> Bob,
> 
>  
> 
> I haven’t run the gdelt dataset myself yet.  That is something I plan
> on doing tomorrow or early next week as I learn more about Hadoop.
> However, when I see errors like that, it is usually my data that I
> have messed up somehow.  I agree that it would be a good thing to try
> a smaller dataset and see if the job completes.
> 
>  
> 
> I’ll post back to the list if I am successful or run into the
> same/other issues.
> 
>  
> 
> Chris Snider
> 
> Senior Software Engineer
> 
> Intelligent Software Solutions, Inc.
> 
> Direct (719) 452-7257
> 
> Description: Description: Description:
> cid:image001.png@01CA1F1F.CBC93990
> 
> 
>  
> 
> From: Barnhart, Bob M. [mailto:ROBERT.M.BARNHART@xxxxxxxxxx] 
> Sent: Thursday, May 08, 2014 3:26 PM
> To: Chris Snider; Discussions between GeoMesa committers
> Subject: RE: [geomesa-dev] Problems running GeoMesa GDELT tutorial
> 
> 
>  
> 
> Chris,
> 
>  
> 
> Thanks for the fix…it worked fine using an ad-hoc authorization
> (“GDELT” vs. “COMMA”).
> 
>  
> 
> I don’t know whether I can collaborate with you directly (as opposed
> to mailing to geomesa-dev@xxxxxxxxxxxxxxxx) but I thought I’d see if
> you were willing to take a look at another issue.
> 
>  
> 
> Still trying to ingest the data loaded into Hadoop
> fromhttp://data.gdeltproject.org/events/GDELT.MASTERREDUCEDV2.1979-2013.zip, I’m now getting an “ArrayIndexOutOfBoundsException” error (see below).
> 
>  
> 
> I’m not sure if (1) the data might be at issue, or (2) my
> Accumulo/Hadoop installation might be at issue. I’m tempted to try
> loading a smaller dataset, perhaps only (some of) the 2014 files from
> http://data.gdeltproject.org/events/index.html.
> 
>  
> 
> Any ideas what might be going on here?
> 
>  
> 
> Thanks,
> 
> Bob Barnhart
> 
>  
> 
> (ingest log…)
> 
>  
> 
> -----------------------------------------------------
> 
> Running: hadoop jar ./target/geomesa-gdelt-1.0-SNAPSHOT.jar
>                  geomesa.gdelt.GDELTIngest
>                 -instanceId ntc-irad
>    -zookeepers localhost:2181                         -user root
> -password (r00t)                         -auths GDELT
>                    -tableName gdelt -featureName event
>    -ingestFile hdfs:///gdelt/uncompressed/gdelt.tsv
> 
> -----------------------------------------------------
> 
> 14/05/08 13:39:09 INFO HSQLDB45DD8FA39A.ENGINE: dataFileCache open
> start
> 
> 14/05/08 13:39:09 INFO HSQLDB45DD8FA39A.ENGINE: Checkpoint start
> 
> 14/05/08 13:39:09 INFO HSQLDB45DD8FA39A.ENGINE: Checkpoint end
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:zookeeper.version=3.4.5-1392090, built on 09/30/2012 17:52
> GMT
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:host.name=localhost
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:java.version=1.7.0_55
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:java.vendor=Oracle Corporation
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:java.home=/usr/lib/jvm/java-1.7.0-openjdk-1.7.0.55.x86_64/jre
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:java.class.path=/usr/local/hadoop-2.4.0/etc/hadoop:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/guava-11.0.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsch-0.1.42.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/activation-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jasper-compiler-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-digester-1.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-el-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/stax-api-1.0-2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/java-xmlbuilder-0.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/junit-4.8.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/avro-1.7.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jettison-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jersey-json-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/xmlenc-0.52.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/zookeeper-3.4.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-beanutils-1.7.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-xc-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/httpclient-4.2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-collections-3.2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-net-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/paranamer-2.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-lang-2.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-beanutils-core-1.8.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/hadoop-auth-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-httpclient-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsp-api-2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-configuration-1.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-math3-3.1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/mockito-all-1.8.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/httpcore-4.2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jets3t-0.9.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/slf4j-api-1.7.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/hadoop-nfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/hadoop-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/hadoop-common-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/guava-11.0.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-el-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/xmlenc-0.52.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jsp-api-2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-nfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/guava-11.0.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/activation-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/leveldbjni-all-1.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/stax-api-1.0-2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/aopalliance-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/guice-servlet-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jettison-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-json-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/zookeeper-3.4.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-xc-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-collections-3.2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/guice-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jline-0.9.94.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-lang-2.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-httpclient-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-client-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/javax.inject-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-guice-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-client-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-tests-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-web-proxy-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-api-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/guice-servlet-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/hamcrest-core-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/guice-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/junit-4.10.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/javax.inject-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-app-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-2.4.0.jar:/usr/local/hadoop-2.4.0/contrib/capacity-scheduler/*.jar
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:java.library.path=/usr/local/hadoop-2.4.0/lib/native
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:java.io.tmpdir=/tmp
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:java.compiler=<NA>
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:os.name=Linux
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:os.arch=amd64
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:os.version=2.6.32-431.11.2.el6.x86_64
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:user.name=barnhartr
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:user.home=/home/barnhartr
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Client
> environment:user.dir=/usr/local/geomesa-gdelt-master
> 
> 14/05/08 13:39:11 INFO zookeeper.ZooKeeper: Initiating client
> connection, connectString=localhost:2181 sessionTimeout=30000
> watcher=org.apache.accumulo.fate.zookeeper.ZooSession
> $ZooWatcher@325c4c8
> 
> 14/05/08 13:39:11 INFO zookeeper.ClientCnxn: Opening socket connection
> to server localhost/127.0.0.1:2181. Will not attempt to authenticate
> using SASL (unknown error)
> 
> 14/05/08 13:39:11 INFO zookeeper.ClientCnxn: Socket connection
> established to localhost/127.0.0.1:2181, initiating session
> 
> 14/05/08 13:39:11 INFO zookeeper.ClientCnxn: Session establishment
> complete on server localhost/127.0.0.1:2181, sessionid =
> 0x145dd8bc67d0007, negotiated timeout = 30000
> 
> OpenJDK 64-Bit Server VM warning: You have loaded
> library /usr/local/hadoop-2.4.0/lib/native/libhadoop.so.1.0.0 which
> might have disabled stack guard. The VM will try to fix the stack
> guard now.
> 
> It's highly recommended that you fix the library with 'execstack -c
> <libfile>', or link it with '-z noexecstack'.
> 
> 14/05/08 13:39:15 WARN util.NativeCodeLoader: Unable to load
> native-hadoop library for your platform... using builtin-java classes
> where applicable
> 
> 14/05/08 13:39:18 INFO Configuration.deprecation: session.id is
> deprecated. Instead, use dfs.metrics.session-id
> 
> 14/05/08 13:39:18 INFO jvm.JvmMetrics: Initializing JVM Metrics with
> processName=JobTracker, sessionId=
> 
> 14/05/08 13:39:19 WARN mapreduce.JobSubmitter: Hadoop command-line
> option parsing not performed. Implement the Tool interface and execute
> your application with ToolRunner to remedy this.
> 
> 14/05/08 13:39:19 WARN mapreduce.JobSubmitter: No job jar file set.
> User classes may not be found. See Job or Job#setJar(String).
> 
> 14/05/08 13:39:19 INFO input.FileInputFormat: Total input paths to
> process : 1
> 
> 14/05/08 13:39:19 INFO mapreduce.JobSubmitter: number of splits:49
> 
> 14/05/08 13:39:19 INFO mapreduce.JobSubmitter: Submitting tokens for
> job: job_local1091701005_0001
> 
> 14/05/08 13:39:19 WARN conf.Configuration:
> file:/hadoop/tmp/mapred/staging/barnhartr1091701005/.staging/job_local1091701005_0001/job.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval;  Ignoring.
> 
> 14/05/08 13:39:19 WARN conf.Configuration:
> file:/hadoop/tmp/mapred/staging/barnhartr1091701005/.staging/job_local1091701005_0001/job.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts;  Ignoring.
> 
> 14/05/08 13:39:58 INFO mapred.LocalDistributedCacheManager: Creating
> symlink: /hadoop/tmp/mapred/local/1399581560010/geomesa-gdelt-1.0-SNAPSHOT.jar <- /usr/local/geomesa-gdelt-master/geomesa-gdelt-1.0-SNAPSHOT.jar
> 
> 14/05/08 13:39:58 INFO mapred.LocalDistributedCacheManager: Localized
> hdfs://localhost:8020/tmp/geomesa-gdelt-1.0-SNAPSHOT.jar as
> file:/hadoop/tmp/mapred/local/1399581560010/geomesa-gdelt-1.0-SNAPSHOT.jar
> 
> 14/05/08 13:39:58 WARN conf.Configuration:
> file:/hadoop/tmp/mapred/local/localRunner/barnhartr/job_local1091701005_0001/job_local1091701005_0001.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval;  Ignoring.
> 
> 14/05/08 13:39:58 WARN conf.Configuration:
> file:/hadoop/tmp/mapred/local/localRunner/barnhartr/job_local1091701005_0001/job_local1091701005_0001.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts;  Ignoring.
> 
> 14/05/08 13:39:58 INFO mapred.LocalDistributedCacheManager:
> file:/hadoop/tmp/mapred/local/1399581560010/geomesa-gdelt-1.0-SNAPSHOT.jar/
> 
> 14/05/08 13:39:58 INFO mapreduce.Job: The url to track the job:
> http://localhost:8080/
> 
> 14/05/08 13:39:58 INFO mapreduce.Job: Running job:
> job_local1091701005_0001
> 
> 14/05/08 13:39:58 INFO mapred.LocalJobRunner: OutputCommitter set in
> config null
> 
> 14/05/08 13:39:58 INFO mapred.LocalJobRunner: OutputCommitter is
> org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
> 
> 14/05/08 13:39:59 INFO mapred.LocalJobRunner: Waiting for map tasks
> 
> 14/05/08 13:39:59 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000000_0
> 
> 14/05/08 13:39:59 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:39:59 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6442450944
> +138958463
> 
> 14/05/08 13:39:59 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:39:59 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:39:59 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:39:59 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:39:59 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:39:59 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:39:59 INFO mapreduce.Job: Job job_local1091701005_0001
> running in uber mode : false
> 
> 14/05/08 13:39:59 INFO mapreduce.Job:  map 0% reduce 0%
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:01 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000001_0
> 
> 14/05/08 13:40:01 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:0+134217728
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:01 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000002_0
> 
> 14/05/08 13:40:01 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:134217728+134217728
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:01 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:02 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000003_0
> 
> 14/05/08 13:40:02 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:268435456+134217728
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:02 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000004_0
> 
> 14/05/08 13:40:02 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:402653184+134217728
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:02 WARN impl.ThriftTransportPool: Server
> 127.0.0.1:9997:9997 (120000) had 20 failures in a short time period,
> will not complain anymore 
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:02 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000005_0
> 
> 14/05/08 13:40:02 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:536870912+134217728
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:02 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:03 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000006_0
> 
> 14/05/08 13:40:03 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:671088640+134217728
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:03 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000007_0
> 
> 14/05/08 13:40:03 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:805306368+134217728
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:03 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000008_0
> 
> 14/05/08 13:40:03 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:939524096+134217728
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:03 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000009_0
> 
> 14/05/08 13:40:03 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1073741824
> +134217728
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:03 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:04 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000010_0
> 
> 14/05/08 13:40:04 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1207959552
> +134217728
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:04 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000011_0
> 
> 14/05/08 13:40:04 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1342177280
> +134217728
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:04 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000012_0
> 
> 14/05/08 13:40:04 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1476395008
> +134217728
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:04 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000013_0
> 
> 14/05/08 13:40:04 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1610612736
> +134217728
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:04 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:05 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000014_0
> 
> 14/05/08 13:40:05 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1744830464
> +134217728
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:05 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:05 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000015_0
> 
> 14/05/08 13:40:05 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1879048192
> +134217728
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:05 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000016_0
> 
> 14/05/08 13:40:05 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2013265920
> +134217728
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:05 INFO mapreduce.Job:  map 2% reduce 0%
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:05 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000017_0
> 
> 14/05/08 13:40:05 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2147483648
> +134217728
> 
> 14/05/08 13:40:05 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:06 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000018_0
> 
> 14/05/08 13:40:06 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2281701376
> +134217728
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:06 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000019_0
> 
> 14/05/08 13:40:06 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2415919104
> +134217728
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:06 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000020_0
> 
> 14/05/08 13:40:06 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2550136832
> +134217728
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:06 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000021_0
> 
> 14/05/08 13:40:06 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2684354560
> +134217728
> 
> 14/05/08 13:40:06 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:07 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:07 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000022_0
> 
> 14/05/08 13:40:07 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2818572288
> +134217728
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:07 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000023_0
> 
> 14/05/08 13:40:07 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2952790016
> +134217728
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:07 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000024_0
> 
> 14/05/08 13:40:07 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3087007744
> +134217728
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:07 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:07 INFO mapreduce.Job:  map 6% reduce 0%
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:07 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000025_0
> 
> 14/05/08 13:40:07 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3221225472
> +134217728
> 
> 14/05/08 13:40:07 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:08 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:08 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000026_0
> 
> 14/05/08 13:40:08 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3355443200
> +134217728
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:08 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:08 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000027_0
> 
> 14/05/08 13:40:08 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3489660928
> +134217728
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:08 INFO mapreduce.Job:  map 10% reduce 0%
> 
> 14/05/08 13:40:08 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:08 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:09 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000028_0
> 
> 14/05/08 13:40:09 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3623878656
> +134217728
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:09 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:09 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:09 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000029_0
> 
> 14/05/08 13:40:09 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3758096384
> +134217728
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:09 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:09 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000030_0
> 
> 14/05/08 13:40:09 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3892314112
> +134217728
> 
> 14/05/08 13:40:09 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:09 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:09 INFO mapreduce.Job:  map 20% reduce 0%
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:10 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:10 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000031_0
> 
> 14/05/08 13:40:10 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4026531840
> +134217728
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:10 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:10 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000032_0
> 
> 14/05/08 13:40:10 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4160749568
> +134217728
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:10 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:10 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000033_0
> 
> 14/05/08 13:40:10 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4294967296
> +134217728
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:10 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:10 INFO mapreduce.Job:  map 29% reduce 0%
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:10 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:11 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000034_0
> 
> 14/05/08 13:40:11 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4429185024
> +134217728
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:11 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:11 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:11 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000035_0
> 
> 14/05/08 13:40:11 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4563402752
> +134217728
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:11 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:11 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000036_0
> 
> 14/05/08 13:40:11 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4697620480
> +134217728
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:11 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000037_0
> 
> 14/05/08 13:40:11 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4831838208
> +134217728
> 
> 14/05/08 13:40:11 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:12 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:12 INFO mapreduce.Job:  map 37% reduce 0%
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:12 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000038_0
> 
> 14/05/08 13:40:12 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4966055936
> +134217728
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:12 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:12 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000039_0
> 
> 14/05/08 13:40:12 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5100273664
> +134217728
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:12 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:12 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000040_0
> 
> 14/05/08 13:40:12 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5234491392
> +134217728
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:12 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:12 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000041_0
> 
> 14/05/08 13:40:12 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5368709120
> +134217728
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:12 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:13 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000042_0
> 
> 14/05/08 13:40:13 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5502926848
> +134217728
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:13 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:13 INFO mapreduce.Job:  map 45% reduce 0%
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:13 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000043_0
> 
> 14/05/08 13:40:13 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5637144576
> +134217728
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:13 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:13 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000044_0
> 
> 14/05/08 13:40:13 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5771362304
> +134217728
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:13 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:13 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000045_0
> 
> 14/05/08 13:40:13 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5905580032
> +134217728
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:13 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:13 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000046_0
> 
> 14/05/08 13:40:13 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6039797760
> +134217728
> 
> 14/05/08 13:40:13 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:14 INFO mapreduce.Job:  map 51% reduce 0%
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:14 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000047_0
> 
> 14/05/08 13:40:14 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6174015488
> +134217728
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:14 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:14 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:14 INFO mapred.LocalJobRunner: Starting task:
> attempt_local1091701005_0001_m_000048_0
> 
> 14/05/08 13:40:14 INFO mapred.Task:  Using
> ResourceCalculatorProcessTree : [ ]
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: Processing split:
> hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6308233216
> +134217728
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: Map output collector class =
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: (EQUATOR) 0 kvi
> 26214396(104857584)
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: soft limit at 83886080
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: bufstart = 0; bufvoid =
> 104857600
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: kvstart = 26214396; length =
> 6553600
> 
> 14/05/08 13:40:14 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:14 INFO mapred.MapTask: Starting flush of map output
> 
> 14/05/08 13:40:14 INFO mapred.LocalJobRunner: map task executor
> complete.
> 
> 14/05/08 13:40:14 WARN mapred.LocalJobRunner: job_local1091701005_0001
> 
> java.lang.Exception: java.lang.ArrayIndexOutOfBoundsException: 39
> 
>                 at org.apache.hadoop.mapred.LocalJobRunner
> $Job.runTasks(LocalJobRunner.java:462)
> 
>                 at org.apache.hadoop.mapred.LocalJobRunner
> $Job.run(LocalJobRunner.java:522)
> 
> Caused by: java.lang.ArrayIndexOutOfBoundsException: 39
> 
>                 at
> geomesa.gdelt.GDELTIngestMapper.map(GDELTIngestMapper.java:60)
> 
>                 at
> geomesa.gdelt.GDELTIngestMapper.map(GDELTIngestMapper.java:27)
> 
>                 at
> org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145)
> 
>                 at
> org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
> 
>                 at
> org.apache.hadoop.mapred.MapTask.run(MapTask.java:340)
> 
>                 at org.apache.hadoop.mapred.LocalJobRunner$Job
> $MapTaskRunnable.run(LocalJobRunner.java:243)
> 
>                 at java.util.concurrent.Executors
> $RunnableAdapter.call(Executors.java:471)
> 
>                 at
> java.util.concurrent.FutureTask.run(FutureTask.java:262)
> 
>                 at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 
>                 at java.util.concurrent.ThreadPoolExecutor
> $Worker.run(ThreadPoolExecutor.java:615)
> 
>                 at java.lang.Thread.run(Thread.java:744)
> 
> 14/05/08 13:40:15 INFO mapreduce.Job:  map 57% reduce 0%
> 
> 14/05/08 13:40:15 INFO mapreduce.Job: Job job_local1091701005_0001
> failed with state FAILED due to: NA
> 
> 14/05/08 13:40:15 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:15 INFO mapreduce.Job: Counters: 25
> 
>                 File System Counters
> 
>                                 FILE: Number of bytes read=1100796112
> 
>                                 FILE: Number of bytes
> written=1112405102
> 
>                                 FILE: Number of read operations=0
> 
>                                 FILE: Number of large read
> operations=0
> 
>                                 FILE: Number of write operations=0
> 
>                                 HDFS: Number of bytes read=1101293020
> 
>                                 HDFS: Number of bytes
> written=1097385436
> 
>                                 HDFS: Number of read operations=1465
> 
>                                 HDFS: Number of large read
> operations=0
> 
>                                 HDFS: Number of write operations=112
> 
>                 Map-Reduce Framework
> 
>                                 Map input records=28
> 
>                                 Map output records=0
> 
>                                 Map output bytes=0
> 
>                                 Map output materialized bytes=168
> 
>                                 Input split bytes=3220
> 
>                                 Combine input records=0
> 
>                                 Spilled Records=0
> 
>                                 Failed Shuffles=0
> 
>                                 Merged Map outputs=0
> 
>                                 GC time elapsed (ms)=70902
> 
>                                 CPU time spent (ms)=0
> 
>                                 Physical memory (bytes) snapshot=0
> 
>                                 Virtual memory (bytes) snapshot=0
> 
>                                 Total committed heap usage
> (bytes)=6046191616
> 
>                 File Input Format Counters 
> 
>                                 Bytes Read=114688
> 
> Exception in thread "main" java.lang.Exception: Job failed
> 
>                 at
> geomesa.gdelt.GDELTIngest.runMapReduceJob(GDELTIngest.java:152)
> 
>                 at
> geomesa.gdelt.GDELTIngest.main(GDELTIngest.java:110)
> 
>                 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> Method)
> 
>                 at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> 
>                 at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 
>                 at java.lang.reflect.Method.invoke(Method.java:606)
> 
>                 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
> 
> 14/05/08 13:40:15 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:15 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:16 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:16 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:16 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:17 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:17 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:17 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:18 INFO mapred.LocalJobRunner: map > sort
> 
> 14/05/08 13:40:18 INFO mapred.LocalJobRunner: map > sort
> 
>  
> 
> From: Chris Snider [mailto:chris.snider@xxxxxxxxxx] 
> Sent: Thursday, May 08, 2014 1:01 PM
> To: Discussions between GeoMesa committers; Barnhart, Bob M.
> Subject: RE: [geomesa-dev] Problems running GeoMesa GDELT tutorial
> 
> 
>  
> 
> Bob,
> 
>  
> 
> I think there may be a misunderstanding between the DB permissions and
> the user authorizations inherent in the Accumulo Schema.
> 
>  
> 
> In an accumulo shell, run a 
> 
> getauths –u root
> 
>  
> 
> This is the Authorizations that the connector is expecting.  
> 
>  
> 
> You can set authorizations for the root user with the following
> accumulo shell
> 
> setauths –s “MY,COMMA,DELIMITED,AUTHS” –u root
> 
>  
> 
> Use the new auths in the connection block for-auths COMMA (for
> example)
> 
>  
> 
> Chris Snider
> 
> Senior Software Engineer
> 
> Intelligent Software Solutions, Inc.
> 
> Description: Description: Description:
> cid:image001.png@01CA1F1F.CBC93990
> 
> 
>  
> 
> From:geomesa-dev-bounces@xxxxxxxxxxxxxxxx
> [mailto:geomesa-dev-bounces@xxxxxxxxxxxxxxxx] On Behalf Of Andrew Ross
> Sent: Thursday, May 08, 2014 1:47 PM
> To: Barnhart, Bob M.; geomesa-dev@xxxxxxxxxxxxxxxx
> Subject: Re: [geomesa-dev] Problems running GeoMesa GDELT tutorial
> 
> 
>  
> 
> Hi Bob,
> 
> Thank you so much for your interest in Geomesa.
> 
> I'll check into what's up with geomesa-user. In the meantime, I've
> added geomesa-dev, which has the developer team on it.
> 
> Cheers!
> 
> Andrew
> 
> On 08/05/14 21:39, Barnhart, Bob M. wrote:
> 
> 
>         I’ve run into problems trying to work through the GeoMesa
>         GDELT Analysis tutorial at
>         http://geomesa.github.io/2014/04/17/geomesa-gdelt-analysis/ .
>         I sent the following email to ‘geomesa-user@xxxxxxxxxxxxxxxx’
>         but it bounced with a non-existent user error. I don’t know if
>         there are any GeoMesa communities of interest to whom I could
>         send a description of my problem, so I’m directing my question
>         to you in hopes that you could provide a (simple) solution, or
>         direct me to a person/site where I might find an answer.
>         
>          
>         
>         I am running Accumulo 1.5.1, Hadoop 2.4.0 and Zookeeper 3.4.6.
>         
>          
>         
>         I’ve been able to load the GDELT data file
>         http://data.gdeltproject.org/events/GDELT.MASTERREDUCEDV2.1979-2013.zip into Hadoop and am trying to ingest this data into Accumulo using the Hadoop command in the tutorial.
>         
>          
>         
>         As shown in the execution trace below, the ingest process
>         fails with the error:
>         
>          
>         
>         java.lang.Exception: java.lang.RuntimeException:
>         org.apache.accumulo.core.client.AccumuloSecurityException:
>         Error BAD_AUTHORIZATIONS for user root - The user does not
>         have the specified authorizations assigned
>         
>          
>         
>         I don’t know to what “specified authorizations” this error
>         might be referring. As shown below, the Accumulo “root” user
>         has all possible System.* and Table.* permissions, including
>         the ‘gdelt’ table:
>         
>          
>         
>         $ accumulo shell –u root
>         
>         root@ntc-irad> userpermissions -u root
>         
>         System permissions: System.GRANT, System.CREATE_TABLE,
>         System.DROP_TABLE, System.ALTER_TABLE, System.CREATE_USER,
>         System.DROP_USER, System.ALTER_USER, System.SYSTEM
>         
>          
>         
>         Table permissions (!METADATA): Table.READ, Table.ALTER_TABLE
>         
>         Table permissions (gdelt): Table.READ, Table.WRITE,
>         Table.BULK_IMPORT, Table.ALTER_TABLE, Table.GRANT,
>         Table.DROP_TABLE
>         
>         Table permissions (trace): Table.READ, Table.WRITE,
>         Table.BULK_IMPORT, Table.ALTER_TABLE, Table.GRANT,
>         Table.DROP_TABLE
>         
>          
>         
>         I would be grateful for any assistance in getting the GDELT
>         data ingesting into Accumulo so that I could complete my
>         assessment of GeoMesa via the GDELT tutorial.
>         
>          
>         
>         Best regards,
>         
>         Bob Barnhart
>         
>         Chief Systems Engineer | 858 826 5596 (Office) | 619 972 9489
>         (Mobile) | barnhartr@xxxxxxxxxx 
>         
>          
>         
>         -----------------------------------------------------
>         
>         Running:   hadoop jar ./target/geomesa-gdelt-1.0-SNAPSHOT.jar
>                        
>         
>         geomesa.gdelt.GDELTIngest
>                            
>         
>         -instanceId ntc-irad                            
>         
>         -zookeepers 127.0.0.1                      
>         
>         -user root -password (r00t)           
>         
>          -auths
>         Table.READ,Table.WRITE,Table.BULK_IMPORT,Table.ALTER_TABLE,Table.GRANT,Table.DROP_TABLE       
>         
>          -tableName gdelt -featureName event                  
>         
>          -ingestFile hdfs:///gdelt/uncompressed/gdelt.tsv
>         
>         -----------------------------------------------------
>         
>         14/05/08 11:52:58 INFO HSQLDB45DD2E6EE0.ENGINE: dataFileCache
>         open start
>         
>         14/05/08 11:52:59 INFO HSQLDB45DD2E6EE0.ENGINE: Checkpoint
>         start
>         
>         14/05/08 11:52:59 INFO HSQLDB45DD2E6EE0.ENGINE: Checkpoint end
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:zookeeper.version=3.4.5-1392090, built on
>         09/30/2012 17:52 GMT
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:host.name=localhost
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:java.version=1.7.0_55
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:java.vendor=Oracle Corporation
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:java.home=/usr/lib/jvm/java-1.7.0-openjdk-1.7.0.55.x86_64/jre
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:java.class.path=/usr/local/hadoop-2.4.0/etc/hadoop:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/guava-11.0.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsch-0.1.42.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/activation-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jasper-compiler-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-digester-1.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-el-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/stax-api-1.0-2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/java-xmlbuilder-0.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/junit-4.8.2.jar:/usr/local/hadoop-2.4.0/share/had oop/commo n/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/avro-1.7.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jettison-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jersey-json-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/xmlenc-0.52.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/zookeeper-3.4.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-beanutils-1.7.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-xc-1.8.8.jar:/usr/lo cal/hadoo p-2.4.0/share/hadoop/common/lib/httpclient-4.2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-collections-3.2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-net-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/paranamer-2.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-lang-2.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-beanutils-core-1.8.0.jar:/usr/local/hadoop-2.4.0 /share/ha doop/common/lib/hadoop-auth-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-httpclient-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jsp-api-2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-configuration-1.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-math3-3.1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/mockito-all-1.8.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/httpcore-4.2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/jets3t-0.9.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/lib/slf4j-api-1.7.5.jar:/usr/local/hadoop-2.4.0/share/hadoo p/common/ hadoop-nfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/hadoop-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/common/hadoop-common-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/guava-11.0.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-el-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/xmlenc-0.52.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/log4 j-1.2.17. jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jsp-api-2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/jasper-runtime-5.5.23.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/hdfs/hadoop-hdfs-nfs-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/g uava-11.0 .2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/activation-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jaxb-api-2.2.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/leveldbjni-all-1.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/stax-api-1.0-2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-jaxrs-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/aopalliance-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/guice-servlet-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jettison-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib /jersey-j son-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/zookeeper-3.4.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jetty-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-xc-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-collections-3.2.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-codec-1.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jsr305-1.3.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/guice-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-cli-1.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jline-0.9.94.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-lan g-2.6.jar :/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-httpclient-3.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jetty-util-6.1.26.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-client-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/commons-logging-1.1.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/javax.inject-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/servlet-api-2.5.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/lib/jersey-guice-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-client-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-tests-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-2.4.0.jar:/usr/local /hadoop-2 .4.0/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-server-web-proxy-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-api-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/commons-compress-1.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/netty-3.6.2.Final.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-core-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jackson-core-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/xz-1.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/usr/local/hadoop-2.4.0/share/h adoop/map reduce/lib/guice-servlet-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/hamcrest-core-1.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/commons-io-2.4.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/guice-3.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.8.8.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/asm-3.2.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/hadoop-annotations-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/junit-4.10.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/javax.inject-1.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/usr/local/hadoop-2.4.0/share/h adoop/map reduce/hadoop-mapreduce-client-jobclient-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-app-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.4.0-tests.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-common-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.4.0.jar:/usr/local/hadoop-2.4.0/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-2.4.0.jar:/usr/local/hadoop-2.4.0/contrib/capacity-scheduler/*.jar
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:java.library.path=/usr/local/hadoop-2.4.0/lib/native
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:java.io.tmpdir=/tmp
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:java.compiler=<NA>
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:os.name=Linux
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:os.arch=amd64
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:os.version=2.6.32-431.11.2.el6.x86_64
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:user.name=barnhartr
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:user.home=/home/barnhartr
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Client
>         environment:user.dir=/usr/local/geomesa-gdelt-master
>         
>         14/05/08 11:53:01 INFO zookeeper.ZooKeeper: Initiating client
>         connection, connectString=127.0.0.1 sessionTimeout=30000
>         watcher=org.apache.accumulo.fate.zookeeper.ZooSession
>         $ZooWatcher@1ec896d2
>         
>         14/05/08 11:53:01 INFO zookeeper.ClientCnxn: Opening socket
>         connection to server localhost/127.0.0.1:2181. Will not
>         attempt to authenticate using SASL (unknown error)
>         
>         14/05/08 11:53:01 INFO zookeeper.ClientCnxn: Socket connection
>         established to localhost/127.0.0.1:2181, initiating session
>         
>         14/05/08 11:53:01 INFO zookeeper.ClientCnxn: Session
>         establishment complete on server localhost/127.0.0.1:2181,
>         sessionid = 0x145dc8e7394000e, negotiated timeout = 30000
>         
>         OpenJDK 64-Bit Server VM warning: You have loaded
>         library /usr/local/hadoop-2.4.0/lib/native/libhadoop.so.1.0.0
>         which might have disabled stack guard. The VM will try to fix
>         the stack guard now.
>         
>         It's highly recommended that you fix the library with
>         'execstack -c <libfile>', or link it with '-z noexecstack'.
>         
>         14/05/08 11:53:05 WARN util.NativeCodeLoader: Unable to load
>         native-hadoop library for your platform... using builtin-java
>         classes where applicable
>         
>         14/05/08 11:53:08 INFO Configuration.deprecation: session.id
>         is deprecated. Instead, use dfs.metrics.session-id
>         
>         14/05/08 11:53:08 INFO jvm.JvmMetrics: Initializing JVM
>         Metrics with processName=JobTracker, sessionId=
>         
>         14/05/08 11:53:08 WARN mapreduce.JobSubmitter: Hadoop
>         command-line option parsing not performed. Implement the Tool
>         interface and execute your application with ToolRunner to
>         remedy this.
>         
>         14/05/08 11:53:08 WARN mapreduce.JobSubmitter: No job jar file
>         set.  User classes may not be found. See Job or
>         Job#setJar(String).
>         
>         14/05/08 11:53:08 INFO input.FileInputFormat: Total input
>         paths to process : 1
>         
>         14/05/08 11:53:08 INFO mapreduce.JobSubmitter: number of
>         splits:49
>         
>         14/05/08 11:53:09 INFO mapreduce.JobSubmitter: Submitting
>         tokens for job: job_local422695915_0001
>         
>         14/05/08 11:53:09 WARN conf.Configuration:
>         file:/hadoop/tmp/mapred/staging/barnhartr422695915/.staging/job_local422695915_0001/job.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval;  Ignoring.
>         
>         14/05/08 11:53:09 WARN conf.Configuration:
>         file:/hadoop/tmp/mapred/staging/barnhartr422695915/.staging/job_local422695915_0001/job.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts;  Ignoring.
>         
>         14/05/08 11:53:47 INFO mapred.LocalDistributedCacheManager:
>         Creating
>         symlink: /hadoop/tmp/mapred/local/1399575189575/geomesa-gdelt-1.0-SNAPSHOT.jar <- /usr/local/geomesa-gdelt-master/geomesa-gdelt-1.0-SNAPSHOT.jar
>         
>         14/05/08 11:53:47 INFO mapred.LocalDistributedCacheManager:
>         Localized
>         hdfs://localhost:8020/tmp/geomesa-gdelt-1.0-SNAPSHOT.jar as
>         file:/hadoop/tmp/mapred/local/1399575189575/geomesa-gdelt-1.0-SNAPSHOT.jar
>         
>         14/05/08 11:53:47 WARN conf.Configuration:
>         file:/hadoop/tmp/mapred/local/localRunner/barnhartr/job_local422695915_0001/job_local422695915_0001.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval;  Ignoring.
>         
>         14/05/08 11:53:47 WARN conf.Configuration:
>         file:/hadoop/tmp/mapred/local/localRunner/barnhartr/job_local422695915_0001/job_local422695915_0001.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts;  Ignoring.
>         
>         14/05/08 11:53:47 INFO mapred.LocalDistributedCacheManager:
>         file:/hadoop/tmp/mapred/local/1399575189575/geomesa-gdelt-1.0-SNAPSHOT.jar/
>         
>         14/05/08 11:53:47 INFO mapreduce.Job: The url to track the
>         job: http://localhost:8080/
>         
>         14/05/08 11:53:47 INFO mapreduce.Job: Running job:
>         job_local422695915_0001
>         
>         14/05/08 11:53:47 INFO mapred.LocalJobRunner: OutputCommitter
>         set in config null
>         
>         14/05/08 11:53:47 INFO mapred.LocalJobRunner: OutputCommitter
>         is org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
>         
>         14/05/08 11:53:47 INFO mapred.LocalJobRunner: Waiting for map
>         tasks
>         
>         14/05/08 11:53:47 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000000_0
>         
>         14/05/08 11:53:47 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:47 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6442450944
>         +138958463
>         
>         14/05/08 11:53:47 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:48 INFO mapreduce.Job: Job
>         job_local422695915_0001 running in uber mode : false
>         
>         14/05/08 11:53:48 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:48 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:48 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:48 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:48 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:48 INFO mapreduce.Job:  map 0% reduce 0%
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:49 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000001_0
>         
>         14/05/08 11:53:49 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:0+134217728
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:49 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:49 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:49 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000002_0
>         
>         14/05/08 11:53:49 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:134217728
>         +134217728
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:49 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:49 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:49 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000003_0
>         
>         14/05/08 11:53:49 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:268435456
>         +134217728
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:49 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:49 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:49 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000004_0
>         
>         14/05/08 11:53:49 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:402653184
>         +134217728
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:49 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:49 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:49 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000005_0
>         
>         14/05/08 11:53:49 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:536870912
>         +134217728
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:49 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:49 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:49 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000006_0
>         
>         14/05/08 11:53:50 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:671088640
>         +134217728
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:50 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:50 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000007_0
>         
>         14/05/08 11:53:50 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:805306368
>         +134217728
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:50 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:50 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000008_0
>         
>         14/05/08 11:53:50 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:939524096
>         +134217728
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:50 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:50 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000009_0
>         
>         14/05/08 11:53:50 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1073741824
>         +134217728
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:50 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:50 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000010_0
>         
>         14/05/08 11:53:50 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1207959552
>         +134217728
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:50 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:50 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000011_0
>         
>         14/05/08 11:53:50 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1342177280
>         +134217728
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:50 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:50 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:50 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000012_0
>         
>         14/05/08 11:53:50 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1476395008
>         +134217728
>         
>         14/05/08 11:53:50 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:51 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:51 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000013_0
>         
>         14/05/08 11:53:51 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1610612736
>         +134217728
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:51 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:51 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000014_0
>         
>         14/05/08 11:53:51 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1744830464
>         +134217728
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:51 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:51 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000015_0
>         
>         14/05/08 11:53:51 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:1879048192
>         +134217728
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:51 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:51 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000016_0
>         
>         14/05/08 11:53:51 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2013265920
>         +134217728
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:51 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:51 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000017_0
>         
>         14/05/08 11:53:51 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2147483648
>         +134217728
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:51 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:51 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:51 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000018_0
>         
>         14/05/08 11:53:51 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2281701376
>         +134217728
>         
>         14/05/08 11:53:51 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:52 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:52 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000019_0
>         
>         14/05/08 11:53:52 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2415919104
>         +134217728
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:52 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:52 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000020_0
>         
>         14/05/08 11:53:52 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2550136832
>         +134217728
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:52 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:52 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000021_0
>         
>         14/05/08 11:53:52 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2684354560
>         +134217728
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:52 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:52 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000022_0
>         
>         14/05/08 11:53:52 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2818572288
>         +134217728
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:52 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:52 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000023_0
>         
>         14/05/08 11:53:52 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:2952790016
>         +134217728
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:52 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:52 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:52 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000024_0
>         
>         14/05/08 11:53:52 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3087007744
>         +134217728
>         
>         14/05/08 11:53:52 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:53 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:53 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000025_0
>         
>         14/05/08 11:53:53 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3221225472
>         +134217728
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:53 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:53 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000026_0
>         
>         14/05/08 11:53:53 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3355443200
>         +134217728
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:53 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:53 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000027_0
>         
>         14/05/08 11:53:53 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3489660928
>         +134217728
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:53 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:53 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000028_0
>         
>         14/05/08 11:53:53 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3623878656
>         +134217728
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:53 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:53 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000029_0
>         
>         14/05/08 11:53:53 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3758096384
>         +134217728
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:53 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:53 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:53 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000030_0
>         
>         14/05/08 11:53:53 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:3892314112
>         +134217728
>         
>         14/05/08 11:53:53 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:54 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:54 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000031_0
>         
>         14/05/08 11:53:54 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4026531840
>         +134217728
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:54 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:54 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000032_0
>         
>         14/05/08 11:53:54 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4160749568
>         +134217728
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:54 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:54 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000033_0
>         
>         14/05/08 11:53:54 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4294967296
>         +134217728
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:54 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:54 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000034_0
>         
>         14/05/08 11:53:54 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4429185024
>         +134217728
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:54 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:54 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000035_0
>         
>         14/05/08 11:53:54 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4563402752
>         +134217728
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:54 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:54 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:54 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000036_0
>         
>         14/05/08 11:53:54 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4697620480
>         +134217728
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:54 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:54 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:54 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000037_0
>         
>         14/05/08 11:53:55 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4831838208
>         +134217728
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:55 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:55 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000038_0
>         
>         14/05/08 11:53:55 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:4966055936
>         +134217728
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:55 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:55 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000039_0
>         
>         14/05/08 11:53:55 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5100273664
>         +134217728
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:55 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:55 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000040_0
>         
>         14/05/08 11:53:55 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5234491392
>         +134217728
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:55 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:55 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000041_0
>         
>         14/05/08 11:53:55 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5368709120
>         +134217728
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:55 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:55 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:55 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000042_0
>         
>         14/05/08 11:53:55 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5502926848
>         +134217728
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:55 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:55 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:55 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000043_0
>         
>         14/05/08 11:53:56 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5637144576
>         +134217728
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:56 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:56 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000044_0
>         
>         14/05/08 11:53:56 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5771362304
>         +134217728
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:56 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:56 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000045_0
>         
>         14/05/08 11:53:56 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:5905580032
>         +134217728
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:56 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:56 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000046_0
>         
>         14/05/08 11:53:56 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6039797760
>         +134217728
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:56 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:56 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000047_0
>         
>         14/05/08 11:53:56 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6174015488
>         +134217728
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:56 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:56 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:56 INFO mapred.LocalJobRunner: Starting task:
>         attempt_local422695915_0001_m_000048_0
>         
>         14/05/08 11:53:56 INFO mapred.Task:  Using
>         ResourceCalculatorProcessTree : [ ]
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Processing split:
>         hdfs://localhost:8020/gdelt/uncompressed/gdelt.tsv:6308233216
>         +134217728
>         
>         14/05/08 11:53:56 INFO mapred.MapTask: Map output collector
>         class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
>         
>         14/05/08 11:53:57 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:57 INFO mapred.MapTask: (EQUATOR) 0 kvi
>         26214396(104857584)
>         
>         14/05/08 11:53:57 INFO mapred.MapTask:
>         mapreduce.task.io.sort.mb: 100
>         
>         14/05/08 11:53:57 INFO mapred.MapTask: soft limit at 83886080
>         
>         14/05/08 11:53:57 INFO mapred.MapTask: bufstart = 0; bufvoid =
>         104857600
>         
>         14/05/08 11:53:57 INFO mapred.MapTask: kvstart = 26214396;
>         length = 6553600
>         
>         14/05/08 11:53:57 INFO mapred.MapTask: Starting flush of map
>         output
>         
>         14/05/08 11:53:57 INFO mapred.LocalJobRunner: map task
>         executor complete.
>         
>         14/05/08 11:53:57 WARN mapred.LocalJobRunner:
>         job_local422695915_0001
>         
>         java.lang.Exception: java.lang.RuntimeException:
>         org.apache.accumulo.core.client.AccumuloSecurityException:
>         Error BAD_AUTHORIZATIONS for user root - The user does not
>         have the specified authorizations assigned
>         
>                         at org.apache.hadoop.mapred.LocalJobRunner
>         $Job.runTasks(LocalJobRunner.java:462)
>         
>                         at org.apache.hadoop.mapred.LocalJobRunner
>         $Job.run(LocalJobRunner.java:522)
>         
>         Caused by: java.lang.RuntimeException:
>         org.apache.accumulo.core.client.AccumuloSecurityException:
>         Error BAD_AUTHORIZATIONS for user root - The user does not
>         have the specified authorizations assigned
>         
>                         at
>         org.apache.accumulo.core.client.impl.TabletServerBatchReaderIterator.hasNext(TabletServerBatchReaderIterator.java:210)
>         
>                         at geomesa.core.data.AccumuloDataStore
>         $$anonfun$readMetadataItem
>         $1.apply(AccumuloDataStore.scala:169)
>         
>                         at geomesa.core.data.AccumuloDataStore
>         $$anonfun$readMetadataItem
>         $1.apply(AccumuloDataStore.scala:157)
>         
>                         at scala.collection.MapLike
>         $class.getOrElse(MapLike.scala:128)
>         
>                         at
>         scala.collection.AbstractMap.getOrElse(Map.scala:58)
>         
>                         at
>         geomesa.core.data.AccumuloDataStore.readMetadataItem(AccumuloDataStore.scala:157)
>         
>                         at
>         geomesa.core.data.AccumuloDataStore.getAttributes(AccumuloDataStore.scala:220)
>         
>                         at
>         geomesa.core.data.AccumuloDataStore.getSchema(AccumuloDataStore.scala:267)
>         
>                         at
>         geomesa.gdelt.GDELTIngestMapper.setup(GDELTIngestMapper.java:53)
>         
>                         at
>         org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142)
>         
>                         at
>         org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
>         
>                         at
>         org.apache.hadoop.mapred.MapTask.run(MapTask.java:340)
>         
>                         at org.apache.hadoop.mapred.LocalJobRunner$Job
>         $MapTaskRunnable.run(LocalJobRunner.java:243)
>         
>                         at java.util.concurrent.Executors
>         $RunnableAdapter.call(Executors.java:471)
>         
>                         at
>         java.util.concurrent.FutureTask.run(FutureTask.java:262)
>         
>                         at
>         java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         
>                         at java.util.concurrent.ThreadPoolExecutor
>         $Worker.run(ThreadPoolExecutor.java:615)
>         
>                         at java.lang.Thread.run(Thread.java:744)
>         
>         Caused by:
>         org.apache.accumulo.core.client.AccumuloSecurityException:
>         Error BAD_AUTHORIZATIONS for user root - The user does not
>         have the specified authorizations assigned
>         
>                         at
>         org.apache.accumulo.core.client.impl.TabletServerBatchReaderIterator.doLookup(TabletServerBatchReaderIterator.java:701)
>         
>                         at
>         org.apache.accumulo.core.client.impl.TabletServerBatchReaderIterator$QueryTask.run(TabletServerBatchReaderIterator.java:361)
>         
>                         at
>         org.apache.accumulo.trace.instrument.TraceRunnable.run(TraceRunnable.java:47)
>         
>                         at
>         java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         
>                         at java.util.concurrent.ThreadPoolExecutor
>         $Worker.run(ThreadPoolExecutor.java:615)
>         
>                         at
>         org.apache.accumulo.trace.instrument.TraceRunnable.run(TraceRunnable.java:47)
>         
>                         at
>         org.apache.accumulo.core.util.LoggingRunnable.run(LoggingRunnable.java:34)
>         
>                         ... 1 more
>         
>         Caused by: ThriftSecurityException(user:root,
>         code:BAD_AUTHORIZATIONS)
>         
>                         at
>         org.apache.accumulo.core.tabletserver.thrift.TabletClientService$startMultiScan_result$startMultiScan_resultStandardScheme.read(TabletClientService.java:8165)
>         
>                         at
>         org.apache.accumulo.core.tabletserver.thrift.TabletClientService$startMultiScan_result$startMultiScan_resultStandardScheme.read(TabletClientService.java:8142)
>         
>                         at
>         org.apache.accumulo.core.tabletserver.thrift.TabletClientService$startMultiScan_result.read(TabletClientService.java:8081)
>         
>                         at
>         org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78)
>         
>                         at
>         org.apache.accumulo.core.tabletserver.thrift.TabletClientService$Client.recv_startMultiScan(TabletClientService.java:294)
>         
>                         at
>         org.apache.accumulo.core.tabletserver.thrift.TabletClientService$Client.startMultiScan(TabletClientService.java:274)
>         
>                         at
>         org.apache.accumulo.core.client.impl.TabletServerBatchReaderIterator.doLookup(TabletServerBatchReaderIterator.java:644)
>         
>                         ... 7 more
>         
>         14/05/08 11:53:57 WARN impl.TabletServerBatchReader:
>         TabletServerBatchReader not shutdown; did you forget to call
>         close()?
>         
>         14/05/08 11:53:58 INFO mapreduce.Job: Job
>         job_local422695915_0001 failed with state FAILED due to: NA
>         
>         14/05/08 11:53:58 INFO mapreduce.Job: Counters: 0
>         
>         Exception in thread "main" java.lang.Exception: Job failed
>         
>                         at
>         geomesa.gdelt.GDELTIngest.runMapReduceJob(GDELTIngest.java:152)
>         
>                         at
>         geomesa.gdelt.GDELTIngest.main(GDELTIngest.java:110)
>         
>                         at
>         sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         
>                         at
>         sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         
>                         at
>         sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         
>                         at
>         java.lang.reflect.Method.invoke(Method.java:606)
>         
>                         at
>         org.apache.hadoop.util.RunJar.main(RunJar.java:212)
>         
> 
>  
> 
> 
> _______________________________________________
> geomesa-dev mailing list
> geomesa-dev@xxxxxxxxxxxxxxxx
> http://locationtech.org/mailman/listinfo/geomesa-dev



Back to the top