Ladislav Buřita, David Koblížek
University of Defence, Kounicova 65, 662 10 Brno, Czech Republic,
Tomas Bata University in Zlin, Mostní 5139, 760 01 Zlin, Czech Republic,
Vojenský útvar 4854, Pardubice, Czech Republic,
University of Defence, Kounicova 65, 662 10 Brno, Czech Republic.
Abstract. The paper is concerned with scalable pre-processing of data using HADOOP that is a framework based on java for processing of large volumes of data, so called Big Data.. The first part is focused on explaining the main part of the HADOOP system, which includes distributed file system and MapReduce metod. In the second part is described the process of system installation and in the final part is explained the reason of the HADOOP experiment.
Keywords: HADOOP, Big Data, HDFS, MapReduce, Apache