バイオエンジニアリングおよびバイオエレクトロニクスのジャーナル オープンアクセス

抽象的な

Machine Learning 2018: Computer-aided diagnosis In a cloud environment based on a multi-agent system: Abbas M Al-Bakry - University of Information Technology and Communications, Iraq

Abbas M Al-Bakry

In this speech we address solutions for the problems of the low accurate decision; low availability especially in maintains procedures and the scalability in online Computer-Aided Diagnosis (CADs). Most CADs became available online and provide a high importance medical services which develop the health of human beings. CADs are to increase the detection of disease by reducing the false-negative rate due to observational oversights. The online CADs face three major problems:
 (1) The CADs cannot diagnose some diseases because the symptoms of these diseases are not available in the knowledge bases of these systems,
(2) problem is the availability of CADs is depended on the webserver which hosted them. The web servers may possible to stop for maintenance that will imply stopping the CADs systems.
(3) The problem is scalability related to the cost if their admins want to expand them to cover more medical problems. In this lecture, we proposed a new framework to solve the above problems. The framework is composed of a multi-agents system to work on the environment of cloud computing. The framework consists of three Sections: SaaS components, PaaS components, and IaaS components. Each section has its own algorithms and procedures. To evaluate the resulted framework we make a survey in for 150 persons from the medical health sector, students, specialists, physicians and others. The results pointed to a good ratio of acceptance from the users.

Huge information may be a field that treats ways to analyze, methodically extricate data from, or something else bargains with information sets that are as well expansive or complex to be managed with by conventional data-processing application computer programs. Information with numerous cases (columns) offers more noteworthy factual control, whereas information with higher complexity (more attributes or columns) may lead to the next wrong disclosure rate. Huge information challenges incorporate capturing information, information capacity, information investigation, look, sharing, exchange, visualization, questioning, overhauling, data protection and information source. Enormous information was initially related with three key concepts: volume, assortment, and speed. When we handle huge information, we may not test but basically watch and track what happens. In this manner, enormous information regularly incorporates information with sizes that surpass the capacity of the conventional program to handle inside a satisfactory time and esteem. Current utilization of the term enormous information tends to allude to the utilize of prescient analytics, client behavior analytics, or certain other progressed information analytics strategies that extricate esteem from information, and rarely to a specific estimate of the information set. "There's small question that the amounts of information presently accessible are without a doubt expansive, but that's not the foremost pertinent characteristic of this modern information ecosystem." Examination of information sets can discover modern relationships to "spot commerce patterns, anticipate maladies, combat wrongdoing, and so on." Researchers, trade administrators, professionals of pharmaceutical, promoting and governments alike frequently meet challenges with huge data-sets in zones counting Web looks, fintech, urban informatics, and commerce informatics. Researchers experience confinements in e-Science work, counting meteorology, genomics, connectomics, complex material science recreations, science and natural inquire about.

Information sets develop quickly, to a certain degree since they are progressively accumulated by cheap and various information-sensing Web of things gadgets such as portable gadgets, airborne (inaccessible detecting), program logs, cameras, amplifiers, radio-frequency recognizable proof (RFID) perusers and remote sensor systems. The world's innovative per-capita capacity to store data has generally multiplied each 40 months since the 1980s; as of 2012, each day 2.5 exabytes (2.5×260 bytes) of information are generated. Based on an IDC report forecast, the worldwide information volume was anticipated to develop exponentially from 4.4 zettabytes to 44 zettabytes between 2013 and 2020. By 2025, IDC predicts there will be 163 zettabytes of data. One address for expansive undertakings is deciding who ought to claim big-data activities that influence the whole organization.The term has been in utilize since the 1990s, with a few giving credit to John Mashey for popularizing the term. Enormous data ordinarily incorporates information sets with sizes past the capacity of commonly utilized computer program devices to capture, clergyman, oversee, and prepare information inside a middle of the road slipped by time. Enormous information reasoning includes unstructured, semi-structured, and organized information, in any case the most center is on unstructured data.Big data "measure" could be a continually moving target, as of 2012 extending from many dozen terabytes to numerous zettabytes of data.Huge information requires a set of procedures and advances with modern shapes of integration to uncover insights from data-sets that are differing, complex, and of an enormous scale. "Variety", "veracity" and different other "Vs" are included by a few organizations to depict it, a modification challenged by a few industry authorities. Enormous information storehouses have existed in many forms, regularly built by enterprises with an uncommon requirement. Commercial sellers generally advertised parallel database administration frameworks for enormous information starting within the 1990s. For numerous a long time, WinterCorp distributed the biggest database report. The promotional source?] Teradata Enterprise in 1984 promoted the parallel preparing DBC 1012 system. Teradata frameworks were the primary to store and analyze 1 terabyte of information in 1992. Difficult disk drives were 2.5 GB in 1991 so the definition of huge information continuously evolves concurring to Kryder's Law. Teradata introduced the primary petabyte class RDBMS based framework in 2007. As of 2017, there are many dozen petabyte lesson Teradata social databases introduced, the biggest of which surpasses 50 PB. Frameworks up until 2008 were 100% organized social information. Since at that point, Teradata has included unstructured information sorts counting XML, JSON, and Avro.Huge information analytics definition: Enormous information analytics makes a difference in businesses and organizations make way better choices by uncovering data that would have something else been hidden. Meaningful bits of knowledge approximately the patterns, relationships and designs that exist inside huge information can be troublesome to extricate without endless computing control. But the strategies and advances utilized in huge information analytics make it conceivable to memorize more from huge information sets. This incorporates information of any source, estimate, and structure. The prescient models and measurable calculations of information visualization with huge information are more progressed than essential commerce insights questions. Answers are about the moment compared to conventional trade insights strategies. Enormous information is as it was getting greater with the development of counterfeit insights, social media and the Web of Things with a bunch of sensors and gadgets. Information is measured within the “3Vs” of assortment, volume, and speed. There’s more of it than ever sometime recently in genuine time. This exuberant surge of information is aimless and unusable on the off chance that it can’t be examined. But the huge information analytics demonstrates employments machine learning to look at content, insights, and dialect to discover already mysterious experiences. All information sources can be mined for forecasts and value. Business applications run from client personalization to extortion discovery utilizing huge information analytics. They moreover lead to more productive operations. Computing control and the capacity to computerize are fundamental for enormous information and commerce analytics. The approach of cloud computing has made this conceivable.

 

免責事項: この要約は人工知能ツールを使用して翻訳されており、まだレビューまたは確認されていません