Skip to main content

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [List Home]
[geomesa-users] Bootstrapping GeoMesa Accumulo and Spark on AWS

Hi all,

I'm attempting to follow the Boostrapping Geomesa/Accumulo/Spark tutorial located at http://www.geomesa.org/documentation/tutorials/geodocker-geomesa-spark-on-aws.html and am unable to ingest the GDELT data.  It's saying it "Ingested 0 features" when I run the ingest command.

Here are the commands I ran (correcting for a couple typo's I found in the tutorial).

$ FILES=$(seq 7 -1 1 | xargs -n 1 -I{} sh -c "date -d'{} days ago' +%Y%m%d" | xargs -n 1 -I{} echo s3://gdelt-open-data/events/{}.export.csv | tr '\n' ' ')
$ sudo docker exec accumulo-master geomesa ingest -c geomesa.gdelt -C gdelt -f gdelt -s gdelt -u root -p secret $FILES
$ sudo docker exec accumulo-master geomesa export -c geomesa.gdelt -f gdelt -u root -p secret -m 100


The first command seems to work just fine...echo'ing $FILES returns a list of valid s3 paths.

The ingest command seems to work successfully (meaning, no errors), but I get the following and it does not appear to actually ingest anything.

INFO  Creating schema gdelt
INFO  Running ingestion in local mode
INFO  Ingesting 7 files with 1 thread
[                                                            ]   0% complete 0 ingested 0 failed in 00:00:02
INFO  Local ingestion complete in 00:00:02
INFO  Ingested 0 features with no failures.

The export command simply spits out all of the column headers without any rows (features).


If anyone can point to what I'm doing wrong here, I would greatly appreciate it.

Thank you!



--
Brooks Isoldi, Software Developer

Traversed
7164 Columbia Gateway Drive, Suite 120A
Columbia, MD 21046

Back to the top