Module description :The subject addresses the problems of managing and processing of extremely large data sets in a single-server centralized computing systems and in multi-server clustered and distributed computing systems. The topics related to processing of large data sets in centralized environments include the techniques based on the classical data warehouse technologies such multidimensional data model, data warehouse architecture, data warehouse design both at conceptual and logical levels, and data warehouse processing with appropriate specialised query operations. The topics related to processing of large data sets in distributed environments include the techniques that can be implemented on the clusters of inexpensive computing nodes using MapReduce programming model. The subject introduces the students to the real time analytical processing of large data sets with analytical cluster-based distributed data processing systems. Discussion and hands on exercises related to these topics will equip students to meet the challenges in Big Data environments and appreciate the added challenges of dealing with unstructured data. Students will be presented with opportunities to do hands-on work with appropriate commercial tools.Job Requirement - A Ph.D or Master's Degree in related discipline from a reputable university - At least 2 years of relevant teaching experience at the tertiary level is preferred - 5 years of relevant work experience will be an added advantage - Applicant must be available to teach day classes or night classes or both day and night classesWe regret that only shortlisted candidates will be notified.