When deploying a stand alone Hadoop you may encounter the bad data node error. I was able to ingest data through hdfs, but when inserting a file larger than one block, I encountered this error. To resolve this set the replication to 1. The default is 3 – the minimum for true high availability fail-over capability. I set up my environment for development only, so
In the hdfs-site.xml file set

dfs.replication
1

Note, if you’re using Ambari, do this through the config section of the hdfs service.