International Journal of Engineering and Techniques - Volume 1 Issue 3, May – June 2015 OPEN ACCESS
RESEARCH ARTICLE
Effective Implementation of using Big Data in E-Health Insurance Kalaignanam.K1, Aishwarya.M2, Vasantharaj.K3, Kumaresan.A4, Ramya.A5 1, 2, 3, 4, 5
(Computer Science and Engineering, SKP Engineering College, Tiruvannamalai, Tamil Nadu, India)
Abstract: Big data is to be implemented in as full way in real-time; it is still in a research. People need to know what to do with enormous data. Insurance agencies are actively participating for the analysis of patient's data which could be used to extract some useful information. Analysis is done in term of discharge summary, drug & pharma, diagnostics details, doctor’s report, medical history, allergies & insurance policies which are made by the application of map reduce and useful data is extracted. We are analysing more number of factors like disease Types with its agreeing reasons, insurance policy details along with sanctioned amount, family grade wise segregation. Keywords: Big data, Stemming, Map reduce Policy and Hadoop. I.
INTRODUCTION
In [1,2] data mining there is no data storage space for storing more datas.so to overcome this disadvantage we propose a big data .Data mining (the analysis step of the "Knowledge Discovery in Databases" process, or KDD), an interdisciplinary subfield of computer science, is the computing process of patterns discovering to the data sets in larger involving methods at the intersection of artificial intelligence, machine learning, statistics, and database systems. From a data set, the process of extracting information is the overall goal and transformation of the task into a comprehensible structure for the usage in further. Apart from the raw analysis step, where it involves database, data management aspects, the aspect of Data model and inference considerations are more highly manageable and more importance is preferable, interestingness metrics,
ISSN: 2395-1303
complexity considerations, postprocessing discovered structures, visualization, and online updating. Huge amount of data does not store in data mining so we propose a scheme called big data. Digital healthcare solutions have assured to transmute the whole healthcare process to become well-organized, less expensive and higher quality [3, 4]. In the context of eHealth, numerous flows have generated slightly less than 1,000 peta bytes of data now (and may reach about 12 ZBs by 2020 in our own estimates) from various sources such as electronic medical records (EMR) systems, mobilized health records (MHR), personal health records (PHR), mobile health care monitors, genetic sequencing and predictive analytics as well as a large array of biomedical sensors and smart devices. The electronic medical record (EMR) initiative has resulted all types of patient’s data streams at the doctor’s office, insurance companies and hospital. A
http://www.ijetjournal.org
Page 61
International Journal of Engineering and Techniques - Volume 1 Issue 3, May – June 2015
single sole patient stay generates thousands of data elements, including diagnoses, procedures, medications, medical supplies, digital image, lab results and billing. These need to be validated, processed and integrated into a large data pools to enable meaningful analysis. Multiplying this by the entire patient stays across the health processing systems and combining it with the large number of points where data is generated and stored and the scope of the big data challenge begins to emerge [3]. Outside the care provider facilities, external health data are patient generated (including social media, selfquantification including use of smartphones/wearable sensor information on patients' heart rate, brain activity, sleep patterns, temperature, muscle motion, and numerous other clinically useful data points), high-throughput and system wide measurement systems of many biological body parts and their states, as well as other health-related information being carried among associates parties of insurance, government reporting and so on data mining overcome this big data huge amount data to be stored in big data. Using traditional data processing applications the data sets so complex and large becomes very difficult to process the data processing applications refer this paper [5,6]. The additional information derivable from single large set of related data analysis is a trend to larger data sets, the same total amount of data the separate smaller sets are compared to, Allowing correlations to find "business trends spot, diseases preventions, crime combat and goes on. Fig. 1 which includes the E-Health insurance setup.
The limitations are encountered regularly by the scientists in many areas due to large data sets, inclusion of e-Science in general finalizes. Internet search, finance and business informatics are affected by the limitations. Data sets grow in size as the part of data sets grows increasingly in size because they are gathered by informationsensing communication devices, remote sensing, and wireless sensor networks and so on. To store information doubly in the manner of pre-capita capacity for every 40 months since 1980 to 2012, as per every day 2.5 Exabyte’s (2.5×1018) of data were created; as of 2014, 2.3 zettabytes (2.3×1021) of data created every day. The process of determining the step of owning the initiatives of big data is the challenges for larger enterprises which would be the target of whole organization. So, proposing big data as a domain. Big data is a imitative catchy-phrase, used to describe a large mass by considering both organized and unstructured data using traditional database and software techniques. It’s difficult to process since it’s so large. In most enterprise scenarios the data are vast, fast or capacity gets exceeded. Big data also helps organization to improve operations, speed and some more intelligent technical decisions. While the term may seem to refer the strength of data isn’t always the case. Big data term especially used by vendors sometimes consider to the technology as tools and process which the enterprise handles the large amount of data and storage action. The big data originates with Web search organization which queries very large distributed collection of loosely-organized data. II.
INSURA NCE
PATIENT Disease analysis
Fig: 1 Architecture diagram of E-Health Insurance
ISSN: 2395-1303
RELATED WORK
In [5] Recent technological advancements have led to a flood of data from characteristic domains (e.g., health care and scientific sensors, user-generated data, Internet and financial companies over the two decades of past. Big data was coined to capture the reality existence trend. The definition of big data and big data tests are
http://www.ijetjournal.org
Page 62
International Journal of Engineering and Techniques - Volume 1 Issue 3, May – June 2015
discussed. Additionally the presence of Hadoop framework used for addressing big data challenges. Finally, the evaluation benchmarks outline and research directions strength for the big data systems. The concept of big data and the big data value chain highlights covers the whole big data lifecycle. It consists 4 phases: Generation of data, Acquisition of data, data storage, and data analysis. In the Generation phase of big data, listing of big data sources and data attributes are discussed. In the acquisition phase of big data, the data collection technologies, methods are investigated. In the storage phase of big data, enormous cloud-based No SQL storages introduced, and some key features compared in the action of helping the design decisions of big data. Programming models coupled with data storage approaches and which plays a major role in analytics of big data. In the data analytics phase of big data, it investigates analytic methods based on characteristics of data. RESEARCH IN FUTURE In big data system many challenges are in need of further attention in research. Listing of open issues which covers the whole lifecycle of big data, from the platform of big data and the application scenario of processing models: comparing to DBMSs Hadoop stays in the big data analytics which is far from the mature process and it is over forty years old. Firstly, Hadoop must integration of Hadoop with real time high volume data collection & transmission and also provides faster processing. Secondly, while hiding the complex background execution the Hadoop provides a concise in user programming. Finally, security and privacy is H. Hu et al.: Toward Scalable Systems for Big Data Analytics. Taking [6], Intelligence of SP theory and its awareness in the SP machine may differ with the benefits, which grant the big data management and analysis. The introduction of SP System in this paper which aids to control the issues of variation in big data; universal framework for the representation and processing
ISSN: 2395-1303
diverse kinds of knowledge, which helps to reduce the formalisms and formats of diverse for knowledge are processed. In [3], identifies major benefits in interconnection services which are critical in universal deployment. Then it provides information about e-health details regarding AON (application oriented network). Our association management solutions allow service providers to build sophisticated application flows using identifications of subsystems in contrast of identifying all TCP access end points. A list of e-Health network service profiles are presented in this paper which include communication services, processing flows and integrated management. Those futuristic workflows will simplify networking service integration. Our security architecture addressed emerging security needs (i.e., universal tracking of secure associations of entities, multiparty collaborations in “e-Health transactions�, and end-to-end security control).QoS operational management became an integral part of AON solution through rapid build-up of new e-Health associations. As a native support in our AON solution, QoS monitoring provides QoS provisioning as well as security audits. Considering [2], focusing on the eHealthcare application cloud- which enables characteristics and also find the proximity in close which is in the new architecture of eHealthcare and the environment of cloud. Challenges in adaptation of a pure cloud solution are optimized for the e-Healthcare in digital. It also introduces our e-Healthcare Cloud Computing and Networking Solutions architecture design. This design benefits approach include adaptation of critical health information technology into a cloud environment while allowing long term evolution of interconnected cloud solutions for e-Healthcare applications. The overall contributions of our research are in our innovative e-Healthcare application hosting, cloud service access adaptations and interconnection architecture. Additional new
http://www.ijetjournal.org
Page 63
International Journal of Engineering and Techniques - Volume 1 Issue 3, May – June 2015
features are explored in our security and operational designs to overcome the new eHealthcare cloud computing challenges and regulatory requirements such as security policies, access adaptors with SLA and QoS constraints, inter cloud connectivity, and resource distribution into cloud hosting applications. The key contribution of our paper is our global consideration of macro level issues that have never been systematically addressed in current national trials or regional planning programs. We can now provide the solution framework architecture to guide the Interoperability, the Security, and the QoS management. The interoperability layers solution is our main contribution and can be readily applied to the CLOUD computing environment. [7] Says, that the growing popularity and development serious threat to the data mining issues. Viewing the issues of privacy that relates to data mining from a perspective and which investigates how to protect information in a sensitive way. .By differentiating the different users with respect to the security of sensitive information is reported, also to provide some useful insights into the study of PPDM. III.
EXISTING SYSTEM
In [8] the electronic medical record (EMR) initiative has resulted data streams from various sectors like insurance agencies, hospital and office of doctor’s. A single patient stay generates thousands of data elements, including diagnoses, procedures, medications, medical supplies, digital image, lab results and billing. This large amount of data is not properly and efficiently validated, processed and integrated into a large data pools to enable meaningful analysis. Big data is not implemented in as full way in real-time; it is still in a Research. People do not know what to do with Enormous Data. Additional external data are in the e-Health communication and support infrastructure sources including the National Health Information Network (NHIN), Health Information Exchanges (HIE), Health
ISSN: 2395-1303
Information Organizations (HIO) and Regional Health Information Organizations (RHIO). An environment that supports eHealth applications from individual outlets and test facilities as well as insurance providers and government agencies rate tremendous data points interconnected with the national health information networks. E-Health Big Data Service Environments A. Data Sources and Types Data types include structured data, semistructure.HL 7 (Health Level 7) messaging standards supply well defined structured data to carry e-Health information. Examples include transaction sections such as Admission Discharge Transfer, Summary of Care and so on. Various IDs could be involved. International Disease Codes supply standard diagnostic coding schemes for clinical visits. DICOM (Digital Imaging and Communications in Medicine) provides semi-structured data for radiology image exchanges over IP networks [17]. While the metadata could be structured, some internal image stores may not be fully standardized. HIPAA transactions for insurance claims and other processes are based on American Standard Institute accredited committees X12 messaging standards. Those data are well structured.ISO/IEEE suite of protocols and messaging standards for digital health monitoring and diagnostics devices [19] provide wellstructured data. However, they do not preclude vendor devices from introducing additional optional files or other data set extensions. Internal logging and security audit records (such as Java Messaging Service format from application-domain monitoring nodes) could be similar in Health IT systems, but there have been no standards. They have to be treated as semistructured data as far as end-to-end system flows are concerned. A patient-support system can gather, store and deliver medical information (e.g., electronic patient and medical records) to patients and doctors remotely, including medical history, allergies, vaccinations, appointments, and invoices as well as inquiry supports. Inside the vendor systems, there are no guarantees that all data sets are standard compliant. A clinicalsupport system can gather, store and retrieve
http://www.ijetjournal.org
Page 64
International Journal of Engineering and Techniques - Volume 1 Issue 3, May – June 2015
patient medical information for internal use by physicians and healthcare workers delivering services at the point of care. The admin and supervisory personnel can have access to backend processing of resources, insurance claims and billings. A number of exchanged networks have been under development to ensure sublevel standards. But the system level data have not been standardized. Disadvantage in the existing system are very large amount of data is analyzed, No research on unstructured and data less accurate result. IV.
Fig: 3 Healthcare Management System
SYSTEM OF MAP REDUCING
Fig.4 Patients insurance policy Application
Fig: 2 The Admin and Login application of Doctors, Patient and Insurance
[9, 10, 11]Insurance agencies are actively participating for the analysis of patient's data and used to extraction some useful information. The fig. 2 includes the login and admin of registration entry Analysis of discharge summary, drug & pharmacy, diagnostics details, doctor’s report, medical history, allergies & insurance policies are made and Useful Data is extracted.
In the MODIFICATION, we are analysing more number of factors like disease types with its corresponding reasons, insurance policy Details with Sanctioned Amount, Family Grade wise Segregation. Fig. 3 says the details of the patients in the hospital regarding their healthcare management system. The application form for the patient’s insurance policy shown in the fig 4.
Fig.5 Information of the Patient in the Hospital through the Admin
ISSN: 2395-1303
http://www.ijetjournal.org
Page 65
International Journal of Engineering and Techniques - Volume 1 Issue 3, May – June 2015
morphological; conflation is a query evaluation in which search engines treats words with the same stem with the synonym. Fig. 5 also has the information of the patient in the hospital and those information’s are seen through the login of Hospital admin application. Stemmers or stemming algorithms or stemming programming.
Fig.6 storage of patient details in the Database
Fig.7 fetching of patient details from Hospital Database
Fig.8 Detailed list of patients insurance policy
Benefits of the proposed system are more accurate data, improved business decisions, improved marketing targeting strategy and increased revenue due to increased customer and base and decreased costs. Stemming algorithm is the term used in linguistic morphology and information retrieval which describes the process for reducing words in inflected to their stem word. The stem is identical to the root of
ISSN: 2395-1303
Map Reduce is a substructure for processing parallelizable problems from a place of vast datasets using a massive number of computers (nodes), everything assigned to as a cluster or a grid. In stored data the processing occurs as a one of two in an unstructured file system or in a structured database. Fig.6 includes the storage of patient details in the Database. Map Reduce can take benefits of data location, in the storage strength the processing can take place it on or near in sequence to lower the separation over which it must be transferred. The reduction operations and the distributed processing of the map are allowed by the Map Reduce. Fig.7 the details of patient information’s are fetched from the hospital Database. Providing independent process of mapping operation are done in the form of others, the number of independent data sources are limited in practice and the maps are performed in parallel then near every sources the and/or number of CPUs are there. Finally the insurance are sanctioned after the confirmation of policy registration by the patient are analysed in the proposed system which is said in the fig.8. Reduction phases are performed by the set of ‘reducers’ similarly, At the selfsame time the identical key are given to the identical trimmer which shares the outputs of the map, The function is associative in the reduction setup. While this process often appear ineffectual contrast to algorithms that are more consecutive, Usage of Map Reduce to the large server farm where it handles ‘commodity’ servers in a less comparative to the larger datasets also sort’s pet byte of data. The partial failures are recovering some possibility which is offered by the parallelism process or during the storage operation: The work gets rescheduled when one map per or reducer fails –the input data is still available in the assumption.
http://www.ijetjournal.org
Page 66
International Journal of Engineering and Techniques - Volume 1 Issue 3, May – June 2015
V.
CONCLUSION
We have presented a new BDeHS (Big Data for Health Service) approach to bring healthcare flows and data/message Management intelligence into the ecosystems for better stream Processing, operational management and regulatory compliance. By asserting security control into the service layer with native eHealth features our solutions ensure regulatory compliance. VI.
REFERENCE
[1]
Xindong wu, Xingquan zhu and Gong-qing wu, Wei ding Knowledge and data engineering,”Data mining with big data,” IEEE Transaction on Volume: 26, Issue: 1, publication year 2014.109. [2] Shan.A.H, Gopalakrishnan.g and A;Liebel.U; Big data (big Data) “data mining and sharing tool for high content Screening larger scale biological image data,” IEEE International on 2014. [3] W. Liu and E.K. Park, “e-Health AON (Application Oriented Network),”IEEE International Conference on Computer Communication Networks, BMAN Workshop, Nausa, Bahamas, July 2013. [4] E.K. Park and W. Liu, “e-Healthcare Cloud Computing Application Solutions,”IEEE ICNC2013, International Conference on Computing, Networking and Communications, San Diego, CA, January 2013.
[7] Information Security in Big Data: Privacy and Data Mining LEI XU, CHUNXIAO JIANG, (Member, IEEE), JIAN WANG, (Member, IEEE), JIAN YUAN, (Member, IEEE), AND YONG REN, (Member, IEEE) Department of Electronic Engineering, Tsinghua University, Beijing 100084, China Corresponding author: C. Jiang. [8]” Big Data as an e-Health Service” Dr.W.Liu , school of science and technology, georia gwinnett college.Dr.E.R.Park, vp for research and dean of graduate studies californina state university-chico . [9]“Security–A Big Question for Big Data “ Roger Schell University of Southern California, USA. [10] Big Data Analytics for Drug Discovery Keith C.C. Chan The Hong Kong Polytechnic University. [11] Smart Data - How you and I will exploit Big Data for personalized digital health and many other activities Amit Sheth, Kno.e.sis Wright State University. [12] cost Minimization for Big Data Processing in Geo-Distributed Data Centers LIN GU, (Student Member, IEEE), DEZE ZENG, (Member, IEEE), PENG LI, (Member, IEEE) AND SONG GUO, (Senior Member, IEEE) University of Aizu, Fukushima 965-8580, Japan CORRESPONDING AUTHOR: S. GUO(sguo@u-aizu.ac.jp).
[5] Toward Scalable Systems for Big Data Analytics: A Technology Tutorial HAN HU1, YONGGANG WEN2, (Senior Member, IEEE), TAT-SENG CHUA1, AND XUELONG LI3, (Fellow, IEEE). [6] Big Data and the SP Theory of Intelligence J. GERARD WOLFF, (Member, IEEE) CognitionResearch.org, Menai Bridge, U.K.
ISSN: 2395-1303
http://www.ijetjournal.org
Page 67