My hadoop
program was initially run in local mode, and now my goal was to run in fully distributed mode.
To do this, you need to provide access to files that are read in the reducer and mapper functions from all computers in the cluster and therefore I asked the question - also because it is not known on which computer the mapper function will run (only one program mapper from the program logic and the program will run with one mapper), it is also necessary to provide access on the entire cluster to the file input to the mapper function.
In this regard, I have a question: Is it possible to directly use hdfs-files, that is, to pre-copy files from the Linux file system to the HDFS file system (thus, I suppose, these files will become available on all computers of the cluster, if it is not then please correct) and then use the HDFS Java API to read these files, in the reducer and mapper functions running on the cluster computers?
If the answer to this question is positive, then please give an example of copying from the Linux file system to the HDFS file system and reading these files in a java program using the HDFS Java API.