Analysing Apache Web Logs using Hadoop Cluster
DOI:
https://doi.org/10.53555/cse.v2i3.542Keywords:
ANALYSING, APACHE, WEB, LOGS, USING, HADOOP, CLUSTERAbstract
Today each and every domain have more impact in web. Without web and its applications nothing can be done. In web, the volume of data that are arriving, storing and retrieving is huge. As log files over the web are outsized, storage becomes a limit wherein effective techniques such as virtual database prove to be ineffectual for the same. Based on the, popularity of the products, websites etc. the trend pattern of that particular product or web site is analyzed.Hadoop offers a large scale spread consignment processing infrastructure that provides adequate data storing, distributive and analogous handling, segregation of process and fault tolerant on occurrences of data loss. Here, the semistructured data is practiced using Hadoop. Map reduce algorithm and HDFS (Hadoop Distributed File Systems) are used for analysing of any web site.
Downloads
References
Q. Yang, “Introduction to the IEEE Transactions on Big Data”, IEEE Transactions on Big data, no. 1, vol. 1, pp. 2-14, January 2015.
S. Chen, Q. Wang, G. Yu and Y. Zhang, “i2MapReduce: Incremental MapReduce for Mining Evolving Big Data”, IEEE transactions on Knowledge and Data Engineering, vol. 27, no. 7, pp. 1906-1919, July 2015.
Hashem, I. Yaqoob, S. Mokhtar, A. Gani and S. Khan, “The rise of “big data” on cloudcomputing: Review and open research issues”, Elsevier, no. 47, pp. 98-115, 2015.
C. Chen, W. Myounggyu , R. Stoleru and G. G. Xie, “Energy-Efficient Fault-Tolerant Data Storage and Processing in Mobile Cloud”, IEEE Transactions on cloud computing, pp. 28-41, March 2015.
D. Dahiphale, R. Karve, A. V. Vasilakos, H. Liu, “An Advanced MapReduce: Cloud MapReduce, Enhancements and Applications”, IEEE Transactions on Network and service Management, pp. 101-115, April 2014.

