Hi all,
I'm attempting to follow the Boostrapping
Geomesa/Accumulo/Spark tutorial located at http://www.geomesa.org/documentation/tutorials/geodocker-geomesa-spark-on-aws.html
and am unable to ingest the GDELT data. It's saying it
"Ingested 0 features" when I run the ingest command.
Here are the commands I ran (correcting for a couple typo's I
found in the tutorial).
$ FILES=$(seq 7 -1 1 | xargs -n 1 -I{} sh -c "date -d'{} days
ago' +%Y%m%d" | xargs -n 1 -I{} echo
s3://gdelt-open-data/events/{}.export.csv | tr '\n' ' ')
$ sudo docker exec accumulo-master geomesa ingest -c
geomesa.gdelt -C gdelt -f gdelt -s gdelt -u root -p secret
$FILES
$ sudo docker exec accumulo-master geomesa export -c
geomesa.gdelt -f gdelt -u root -p secret -m 100
The first command seems to work just fine...echo'ing $FILES
returns a list of valid s3 paths.
The ingest command seems to work successfully (meaning, no
errors), but I get the following and it does not appear to
actually ingest anything.
INFO Creating schema gdelt
INFO Running ingestion in local mode
INFO Ingesting 7 files with 1 thread
[
] 0% complete 0 ingested 0 failed in 00:00:02
INFO Local ingestion complete in 00:00:02
INFO Ingested 0 features with no failures.
The export command simply spits out all of the column headers
without any rows (features).
If anyone can point to what I'm doing wrong here, I would
greatly appreciate it.
Thank you!
--
Brooks Isoldi, Software Developer
Traversed
7164 Columbia Gateway Drive, Suite 120A
Columbia, MD 21046