R. JaichandranA. Anthony Irudhayarj
Abstract: Global warming is not only the problem of the government or individual organization it is the fundamental problem of every individual. The main cause for global warming is green house gases (GHG). Monitoring and computing the greenhouse gases are a major challenging work. Globally, over the past several decades, human-induced activities like industrial revolution and burning of fossil fuels in power stations, vehicle transport systems and industries contribute significantly to the emission and concentration of GHG in atmosphere. Avoiding their usage may reduce the emission of GHG, but it may not be a practical approach as they are mandatory in modern day-to-day life, alternatively regular monitoring and reporting of GHG parameters may create awareness to individuals and organization for effective and proper use of human induced activities. There are very few works done in developing embedded systems for computing GHG. We have implemented a prototype system for sensing and computing the level of existence of GHG parameters (like CO2, CO, temperature and humidity) in atmosphere using environmental sensors and advanced microcontrollers and energy efficient wireless technologies. The Prototype supports quality in terms of low cost, energy efficiency, flexibility and user friendliness. Data is collected, consistency models are define for analyzing the quality of data and the level of GHG in the deployed environment is computed. The results show that the prototype is capable for monitoring and computation of GHG in the deployed environment and can be applied at all levels of organization for creating awareness, performing scientific studies and to forecast remediation policies by the authorities to individuals and organization in controlling GHG parameters.
Keywords : Wireless sensor network; greenhouse gases; parts per million.
Atallah Mahmoud AL-Shatnawi, Farah Hanna AL-Zawaideh, Safwan AL-Salaimeh, Khairuddin Omar
Abstract: This paper provides and discusses an overview about the offline Arabic Optical Character Recognition (AOCR) system. It also provides and discuses the challenges that must be considered in designing or choosing a cretin method for the AOCR system. Recognition of Arabic characters is more difficult than Latin or Chinese language .The typical AOCR system consists of five components: image acquisition, preprocessing, segmentation, feature extraction and classification (recognition). Each of those contributes to the final recognition rate to improve of the AOCR. In this paper, the challenges in recognition the Arabic written text are explained and discussed. As well as, the operations of offline AOCR system stages are discussed and studied in detail. The operations of AOCR preprocessing stage are also provided and discussed. The AOCR stages drawbacks and advantages are discussed in details.
Keywords : Arabic Text; Character recognition; Pre-processing; Segmentation; Feature Extraction; Classification.
Zahra Sheikhnajdy, Mehran Mohsenzadeh, Mashalah Abbasi Dezfuli
Abstract: Schema matching is a critical step in many applications, such as data warehouse loading, Online Analytical Process (OLAP), Data mining, semantic web and schema integration. This task is defined for finding the semantic correspondences between elements of two schemas. Recently, schema matching has found considerable interest in both research and practice. In this paper, some approaches for supporting semantic schema matching compared and then we suggest three solutions for improving semantic schema matching problem. Keywords : schema matching; element level matcher; structural level matcher; semantic ambiguities; step word; word sense disambiguation.
Priyanka Jangde, Gajendra Singh Chandel, Durgesh Kumar Mishra
Abstract: Secure Multiparty Computation (SMC) allows parties to compute the combine result of their individual data without revealing their data to others. Secure sum computation is one of the important tool of the SMC. On SMC many eminent researchers give their protocols especially in secure sum computation, researchers show their interest. In this paper we provide unique hybrid protocol for secure sum computation which is the combination of Ideal and Real Model. This protocol provides zero data leakage means it is completely secure. If two or more than two parties including third party (TP) become malicious cannot hack or trace the data of any other parties, who were participating in this computation. With the help of Hybrid model we are enhancing the security of the computation and maintaining the privacy of the data. In this paper, we analyzed the computational and communicational complexity and found that both the complexities are O (n).
Keywords : Secure Multiparty Computation (SMC); Third Party (TP); Secure Sum Protocol; Hybrid.
Neetu Sharma, Amit Sharma, Vijay Singh Rathore, D.K Mishra
Abstract: In this paper we introduce a real time congestion control mechanism to multicast the multimedia applications. The key challenge in the design of this mechanism is to calculate the rate of feedback messages by various receivers; appropriate feed back suppression, scalable round trip time measurements & ensuring that feedback delays in the control loop do not adversely affect fairness towards competing flows. Major contributions are the feedback mechanism and provide an end-to-end multicast congestion control schemes. We improve the well-known approach for multicasting, by using exponentially weighted random timers by biasing feedback for low-rate receiver and preventing the response implosion. We evaluate the design using simulation and demonstrate the results for multicast transmission of multimedia data.
Keywords : GECM; ROTT; Congestion; ECN; TFMCC; TFRC; RBRC.
Abstract: Tell now we have seen only the use case, collaboration diagrams, and VLC/UI design process. The main goal of VLC/UI is to display and obtain needed information from the source. Whereas the Low Level Design has not yet clear for web based interactions. In this paper we proposed a new type of design using sequence diagram for web based design on browser side. UI also has to tell us how to complete the interfacing task. Say it uses Apache as web server to interact between Web-based main GUI on client side. The application of this concept is applicable to entire web based software development, Website development. Real time application like banking maintenance, library maintenance…etc our concept gives more reliable result.
Keywords : VLC/UI; ALM-MON; DLD; XSL.
Chandra Sekhar Bhagi
Abstract: Any organization typically has data on Customers, Financials (Chart of accounts, profits, Cost), Products, Business Partners (Suppliers, Distributors), Employees, Locations, Sales Contacts, Physical assets, Claims or Policies (Insurance). These data items or business entities are referred as Master Data. The process and technology involved in acquiring Master Data from multiple domains across enterprise and thereby maintaining a single consistent view is Master Data Management (MDM). MDM is getting popularity in Health & Life sciences, Medical Device Manufacturing, Financial Services, Insurance, Manufacturing & Technology, Retail, Consumer Packaged Goods, Telecommunications, Information services & Media, Aerospace, Defense and Government. The current article illustrates the technical solution to implement enterprise Master Data Management in domain specific companies.
Keywords : Master Data Management; Enterprise Master Data Management; MDM Technical Architecture; MCI; SOA.
Zahid Ansari, M.F. Azeem, Waseem Ahmed, A.Vinaya Babu
Abstract: Clustering techniques are widely used in “Web Usage Mining” to capture similar interests and trends among users accessing a Web site. For this purpose, web access logs generated at a particular web site are preprocessed to discover the user navigational sessions. Clustering techniques are then applied to group the user session data into user session clusters, where inter-cluster similarities are minimized while the intra cluster similarities are maximized. Since the application of different clustering algorithms generally results in different sets of cluster formation, it is important to evaluate the performance of these methods in terms of accuracy and validity of the clusters, and also the time required to generate them, using appropriate performance measures. This paper describes various validity and accuracy measures including Dunn’s Index, Davies Bouldin Index, C Index, Rand Index, Jaccard Index, Silhouette Index, Fowlkes Mallows and Sum of the Squared Error (SSE). We conducted the performance evaluation of the following clustering techniques: k-Means, k-Medoids, Leader, Single Link Agglomerative Hierarchical and DBSCAN. These techniques are implemented and tested against the Web user navigational data. Finally their performance results are presented and compared.
Keywords : component; web usage mining; clustering techniques, cluster validity indices, performance evaluation.
Sunar Arif Hussain, K.E.Sreenivasa Murthy
Abstract: MANETS promise an unprecedented opportunity to monitor physical environments via inexpensive wireless embedded devices. Given the sheer amount of sensed data, efficient taxonomy of them becomes a critical task in many sensor network applications. The Bayesian classifier is a fundamental taxonomy technique. We introduce two classifiers: Naive Bayes and a classifier based on class decomposition using K-means clustering. We consider two complementary tasks: model computation and scoring a data set. We study several layouts for tables and several indexing alternatives. In this paper, we propose a novel decision-tree-based hierarchical scattered taxonomy approach, in which local classifiers are built by individual sensors and merged along the routing path forming a spanning tree. The classifiers are iteratively enhanced by combining strategically generated pseudo data and new local data, eventually converging to a global classifier for the whole network. We also introduce some control factors to facilitate the effectiveness of our approach. Through extensive analysis, we study the impact of the introduced control factors, and demonstrate that our approach maintains high taxonomy accuracy with very low storage and communication overhead. The approach also addresses a critical issue of heterogeneous data distribution among the sensors.
Keywords : Scattered Data Storage; Sensor Networks; Key Sharing; Key Management.
Rajesh Mehra, Virendra Arya, Rajpati Yadav
Abstract: In this paper an efficient design and implementation of ITU-R BT.601 video filter has been presented for digital television receivers. The proposed video filter has been realized using MAC algorithm. The implementation is based on efficient utilization of embedded multipliers and look up table (LUT) of the target device to improve speed, area efficiency and power consumption. It is an efficient method because the use of embedded resources not only increases the speed but also saves the general purpose resources of the target device. The proposed video filter has been designed and simulated using Matlab, synthesized with Xilinx Synthesis Tool (XST), and implemented on Spartan 3E based 3s500efg320-5 FPGA device. The developed video filter structure can operate at an estimated frequency of 28.758 MHz by utilizing 12 multipliers and 245 LUTS of target FPGA device to provide cost effective solution for mobile and wireless communication systems.
Keywords-FPGA; ITU-R; LUT; Video Filter; XST.