GET THE APP

International Journal of Advancements in Technology

International Journal of Advancements in Technology
Open Access

ISSN: 0976-4860

+44 1478 350008

Research Article - (2022)Volume 13, Issue 3

Multilayered Map Reduce Framework to Link Cybernetic Vulnerabilities and Cybernetic Laws from E-News Articles

Sudhandradevi P* and Bhuvaneswari V
 
*Correspondence: Sudhandradevi P, Department of Mechanical Engineering, Indian Institute of Technology Patna, Bihar, India, Tel: 9050999578, Email:

Author info »

Abstract

The growth of technology evaluation and the influence of smart gazettes, which have a very complex structure, the amount of data in an organization, E-commerce, and ERP explode. When data is processed as described, it becomes the engine of every individual. According to projections from 2025, social media, IoT, streaming data, and geodata will generate 80% of unstructured data, and there will be 4.8 billion tech enthusiasts. The most popular social media trend allows users to access publicly available data. Hackers are highly qualified in both the web space and the dark web, and the rise of complexity and digitization of this public access will cause loopholes in legislation. The major goal of this study is to gather information about the cyber vulnerability of electronic news. Data collection, text standardization, and feature extraction were all part of the initial step. In the second step, Map reduce was used to obtain demographic insights using a multi-layered categorization strategy. Cybercrime is classified using a classifier technique, and the model has a 53 percent accuracy rate. Phishing is a result of cyber weaknesses, and it has been discovered in a higher number of metropolitan cities. Men, rather than women, make up the majority of crime victims. Individuals should be made aware of secure access to websites and media, according to the findings of the study. People should be aware of cyber vulnerabilities, as well as cyber laws enacted under the IPC, the IT Act 2000, and CERT-in.

Keywords

Cyber vulnerability; CERT-In; DT classifier; Indian penal code; Map reduce; Social media; Text mormalization

Introduction

The Internet has reduced the entire planet to a single little shell. The use of digital surfaces has increased as a result of technological advancements, digital repositories, and social media. The world has entered the big data era; data is rapidly accumulating, and it is estimated that by 2025, the total volume of big data will have increased from 4.4 zettabytes to 44 zettabytes, or 44 trillion gigabytes. The volume of data has grown at a faster rate from many sources, rendering traditional processing systems ineffective. Intrusion detection, according to helps users secures their systems in cyberspace by alerting them [1]. For security analysis, big data will analyze heterogeneous data and a large volume of data, as well as block network traffic. According to the author, big data uses frameworks like Hadoop and Spark to handle the problem of storing and manipulating heterogeneous data [1,2].

Social media is a type of computer-based technology that allows people to share information and ideas in a virtual setting. It is intended to let users to instantly exchange content via smart devices, such as personal data, images, videos, documents, blogging, social games, and social networks. According to projections, the number of social media users in the United States will reach 257 million by 2023. Burnap et al. [3], used machine learning categories to describe cybernetic hate speech on Twitter. Ethnicity, religion, and more generic responses are used to categorize textual data. For anticipating the anticipated spread of cyber hate speech on Twitter, the data is trained and tested using classification algorithms to categories the optimal utilizing a combination of probabilistic, rule-based, geographical based classifiers [3].

According to a study on global tolls, cybercrime would cost $6 trillion in damages by 2021. The potential for cyberattacks should be used to organize the creation of anti-cybercrime policies. The processing of data in a secure manner is a vital aspect of digitalization. Data breaches have grown in importance as a type of cybercrime that can cause harm to people in either a trustworthy or untrustworthy setting. Unauthorized access, harmful code, denial of service, and theft all contribute to computer crime. Heinous content, Internet fraud, and misconduct are examples of traditional crimes. Access to control technologies, cryptography, surveillance, assurance tools, India’s legislative approach, and the IPC’s core purpose of increasing security awareness in India are all technological solutions [4]. Briefs by traditional crime, online fraud, email spoofing, and virus/worm assault, Trojan virus, web hacking, hacking, and computer abuse are all examples of cybercrime [5]. Phishing, email spoofing, and phone phishing are the experimental analyses of the numerous cyberattacks provided in the research. A phone phishing investigation targets 50 employees and their bank account information. In this percentage, 32% of employees used electronic banking to share their credentials. Employees revealed their passwords in 16 percent of cases, while 52 percent refused to do so. According to the report, emails are hacked due to two concerns: they are misleading and consumers overestimate the security and confidentiality of emails [5].

Provided a security analysis solution for cyber security in the context of big data [6], Because of the continuous expansion of data, the type and frequencies of cyber-attacks are increasing at an exponential rate. The worldwide network is connected to a secure system in every country. In 2000, 45 million people were affected by cybercrime, according to a poll. Cybercrime includes spamming, botnets, Denial of Service (DoS), phishing, malware, and website threats. As a result of big data, cyber attackers’ hacking skills have improved. They split the data into passive and active users and implemented security analytics using BDA technology [6].

Users are completely ignorant of the occurrence of cybercrime. Preventing cybercrime and attacks is the responsibility of law enforcement agencies such as the Computer Emergency Readiness Team (CERT-In) and the Government of India. The Indian Penal Code (IPC) sections of the IT Act 2000 provide legal cyber punishment/penalty in India for certain cybercrime attacks. Cybercrime, issues, trends, and difficulties, as well as solutions, were the focus of Kandpal et al. [7]. According to polls, cybercrime increased by 57.1 percent in 2012.61.0 percent of IT Act offenders and 45.2 percent of IPC Sections offenders were between the ages of 18 and 30. According to CERT-in 78, the 308371 government and other websites were hacked between 2011 and 2013. Updating computers, choosing a secure password, protecting computers, being social media savvy, and securing a wireless network are all important skills to reduce cybercrime threats [8].

Methods

A multi-layered technique was used to implement the architecture. Data are gathered in the first layer from various news sources and IPC Laws. Data were extracted and processed in a structured format in the second layer. The Crime corpus is created for storage in the third tier. Cybercrime is classified into demographic data such as gender, geographic attributes, cyber offences, and criminal legislation in the fourth layer. The study was released as an analytical report charts and statistics reports at the bottom layer. The critical reflections on big data security by aim to identify the risks and problems of data security as well as the social and economic benefits of big data [9]. They definitively define the dangers of big data as a loss of confidentiality, integrity, and availability. The problem will be remedied by improving human capital through know-how and inventions, relational capital through improved client relationships, and structural capital through management process modifications. Big data was primarily concerned with technological, sociological, and ethical ramifications, as well as moral judgments [9]. Depicts the multi- layer architecture of cybercrime vulnerability (Figure 1).

architecture

Figure 1: Multi-layered architecture.

Using the Map reduce technique, a multi-layered architecture was used to associate the labels with the document label. Data are divided into two categories: socioeconomic cybercrime and domain-based cybercrime, using a tiered method. Table 1 shows multi-layered data extraction [9].

Data sources Socioeconomic cybercrime Domain-based cybercrime
Article report Cyber offence Cyber article label IPC crime label
Geographic location Geographic parameters State
Gender Cyber victim Cyber criminals
Crime report (IPC) Crime laws Cert-in, IPC/IT act section Punishment

Table 1: Data extraction: Multi-layered approach.

The objective of this research is to create a framework for Cyber Crime Attacks (CCA). Cybercrime attacks are classified using the cyber vulnerabilities found in articles and cyber legislation found in the IPC/IT sections (CVCL). Figure 2 describes the methodology of Cyber Crime Attacks (CCA) [10].

Framework

Figure 2: Framework: Cyber Crime Attacks (CCA).

i) Techniques for building a database of cyber-vulnerabilities and laws.

ii) The multi-layered cyber vulnerability approach considers demographic aspects as well as crimes.

iii) Classify cybercrime by category and test the model using the classifier technique.

Data acquisition

Cybercrime news is gathered from e-news stories and presented systematically. Most of the news pieces are unstructured and clear.

E-news dataset

A manual data collection of 100 instances of newspaper diaries is performed. Article Labels, Headlines, Content, States, Year, Gender, and URL are the attributes of the dataset.

Cyber law dataset

The Cyber law dataset is a collection of cyber law sections and punishments obtained from India’s cybercrime initiatives. The Cyber law dataset contains 78 cases, with Crime labels as the first attribute, followed by sections and sanctions. As per India rules, the crime section and punishments are CERTIn and the Indian Penalty Code (IPC) IT Act 2000. Table 2 shows the statistical reports for both the datasets.

Reports Feature extraction E-news dataset Cyber law dataset
Data acquisition report Number of sources 2 2
Sources The Indian express, The Hindu CERT-in, IPC/IT Act 2000
Domain-level reports Crime labels Email spoofing, clone finger prints, facebook, debit card, social media, twitter, hacking, whatsapp, online crime, ransomware, harassment, internet fraud, data-wiping virus Phishing, profile hacking, cyber bullying, web jacking, online scams, software piracy, cyber-stalking
Dataset instance reports Number of instances 100 78
Number of attributes 7 3
Attributes Article label, headlines, content, district, states, year, url Crime label, IPC sections, punishment

Table 2: Data extraction: Multi-layered approach.

Extraction of relations

In the context of the newspaper, punctuation marks, special characters, and images are determined. To provide a more user- friendly and distinct representation for higher-level modelling, which is also legible by humans. It’s tokenized, with stop words eliminated, morphological normalization, and collocation. The text pre-processing phase is illustrated in Figure 3.

Extractions

Figure 3: Extractions of relations.

The text is standardized in this part using NLP algorithms. To achieve a consistent structure, the first step is to remove the punctuation expressions and convert them to a lower case. They’re listed below.

Tokenization

Tokenization divides the text into bits based on how it is used in the context. The “Sentence tokenization” or “Word Tokenization” techniques can be used to tokenize the context. Word tokenization is done with this task (Figure 4).

technique

Figure 4: Pre-processing technique.

The document is cleaned using the pre-processing technique in step #1. Because the punctuation marks in the document take up unnecessary memory, they are eliminated. The documents have been changed to lower case for unique identification. Converting to a lower expression can sometimes modify the context’s semantic meaning. Part-of-speech refines this issue by revealing the world’s syntactic behavior.

Stop words removal

Stop word removal is used to locate data that isn’t relevant. The information containing words has no sense and is also noisy. Dimension is utilized to greatly minimize the tokenization document.

In #2, unique space characters are used to tokenize words. Following tokenization, a few common terms arise, which add noise to the study and have little meaning. These words are known as stopwords or empty words. Removing stopwords also allows for a large reduction in the number of tokens in documents, as well as a reduction in the feature dimension is represented (Figure 5).

Stop

Figure 5: Stop words or empty words.

Morphological normalization

The method is known as morphomes, and it seeks to find stem words. Stemming is a normalization approach that removes common suffixes from a term’s output and returns the term’s underlying word. Lemmatization is the correct application of the word’s morphological analys is vocabulary (Figure 6).

Stemming

Figure 6: Stemming and Lemmatization of the word morphological analysis.

Building up from smaller meaning-bearing pieces is what normalization is all about. The first document in #3 contains two stem words, such as “constables” and “nabbed.” ‘constabl’ and ‘nab’ are the results of eliminating the suffix. The word ‘constabl’ is incorrect. It will attempt to return the dictionary format as ‘constable’ after removing the inflectional endings.

Crime corpus of relations: Label extraction

The keyword is retrieved from the n-gram approach in collocation to discover the most relevant terms and better insight into the examined material.

Collocation: Bigram (two-gram) approach and trigram (three- gram) approach

After pre-processing the newspaper articles, a corpus was constructed for labeling cybercrime categories with article label domains. The frequency of document terms is calculated for all documents, and important terms are identified using the n-gram method. IDF (t, D)=logN/|d€D:t€D|, where N is the number of documents in the corpus (N=|D|), T is the number of occurrences, and |d € D:t€D| is the number of document terms (Figure 7).

Bigram

Figure 7: Bigram (two-gram) approach and trigram (three-gram) approach.

Narrow down the news content based on a certain theme, such as crime labels or article label, to have a better understanding of the behavioral insights. Collocation aids in the retrieval of two or more words that are extremely likely to occur together. News the documents contain content that is closely related to the term ‘cyber,’ such as ‘cybercrime,’ ‘cyberspace,’ ‘cyber criminals,’ ‘cyber security,’ and so on. The bi-gram (two-gram) or tri-gram (three- gram) approaches might be used depending on the findings. The sentence in 1, 2, 4 documents does not give any significant phrase in the #4 bi-gram technique. To infer more insights, always prefer the tri-gram approach based on behavior analysis.

Association of cyber news-visualizing bigrams approach

The relationships between the news terms are visualized using a ‘network’ or ‘graph’ to find them all at once. In #5, phrases connected to cybercrime, such as ‘cybercrime,’ ‘police,’ ‘bank,’ and ‘crime,’ are nodes that are frequently followed by others. #5 shows how to see a graph (Figure 8).

Association

Figure 8: Association of cyber news-visualizing bigrams approach.

Cybercrime binary classification

The current traffic between People-centric and Techno-centric has clearly demonstrated the complexity of cybercrime and its impact. These two words are linked to the cyber-factor vs. the human element of criminal crimes. Figure 9 shows the categories of cybercrime based on cyber enabled and cyber dependent [11].

Cybercrime

Figure 9: Cybercrime label classification.

Semantic mapping model-corpus creation

To execute semantic modelling, associate the label with the map reduce technique. The Cyber vulnerable label corpa is mapped with IPC Section [Law Punishment] using a map reduces technique. Shows the key (Crime labels) and value (Accusation) pair (Figure 10).

mapreduce

Figure 10: Multi-Layered Architecture: mapreduce Approach.

A multi-layered architecture is used to model the dataset. To determine the Crime vulnerabilities, the qualities are modelled from a domain perspective. The various outcomes of the semantic match are discussed in the following sections.

#6 mapping of cyber elements and crime label

In case #6, 87 percent of cyber vulnerabilities are classified as “techno centric”, and they occur as a result of gadgets such as mobile phones, laptops, and smart devices. “Software piracy, web jacking, phishing, and cyber bullying” are the key contributions of technology aspects, while “online scams”, Profile hacking, and cyber-stalking” are the main contributions of people-based factors (Figure 11).

Mapping

Figure 11: Mapping of cyber elements and crime label.

#7 Mapping of Cyber Labels, Article Labels

#7 implies that the majority of susceptible crime is reported as “Phishing”. According to the data, Phishing and kinds accounted for 38% of all offences. Email spoofing and debit card access are both common vulnerabilities. Furthermore, because of the prevalence of social media apps such as Facebook, WhatsApp, Twitter, and Instagram, 23 percent of offences were reported as “Cyber Bullying” (Figure 12).

cyber

Figure 12: Mapping of cyber labels, article labels

#8 state-wise mapping of cyber elements

#8 It reports on the geographical study of cyber susceptibility victims in India. It alludes to the fact that the majority of crimes in India occur as a result of technological influences. In comparison, the western zone of India was hit by 48 percent of crimes, while the south was hit by 23 percent. India’s most affected states are Maharashtra and Delhi. Both are affected by technology and are centered on people (Figure 13).

State

Figure 13: State-wise mapping of cyber elements.

#9 mapping of states and district

#9 refers to Maharashtra, which is heavily impacted by cybercrime in India’s Western zone. The central zone of Delhi is the second most afflicted, and Karnataka is the most affected state in India’s southern zone. In the northern section of India, the number of crimes is the lowest (Figure 14).

district

Figure 14: Mapping of states and district.

Results

Classification of relations: decision tree classifier

To classify cybercrime, the methodology is tested using a decision tree classifier. Various metrics, such as Gini and Entropy, are used to validate the decision tree. Table 3 shows the decision tree, which is divided into two phases.

Building phase Operational phase
Import the dataset Make predictions
Pre-process the dataset Measure performance
Split the data based on the train/test ratio  -
Build the model for classifier  -

Table 3: Phases of decision tree.

Decision tree algorithm-pseudo code

i) Assign all the training data to the root node

ii) Partition input data recursively based on selected attributes

iii) Test attributes at each node are selected based on a heuristic or statistical measure

iv) Conditions to stop partition

v) If all samples belong to the same class, then no further partitioning

Decision tree: Outcome

In this dataset, the crime label (IPCV_CL) is a subset based on domain expertise (B_LAB). Here the dependent variable is crime label (IPCV_CL) and the independent variable is Article_label (CAL). Article_label is fetched from articles and linked to Crime vulnerability activity (IPCV_CL). The data is pre-processed by removing null values and missing values. The structure of the dataset is given in below Figure 15.

Dataset

Figure 15: Dataset. Note: Crime label (IPCV_CL); Domain expertise (B_LAB).

Split of train/test

Random sampling was used to divide the dataset into the train and test sets. A random id is generated so that the value can be split in the ratio at random. 70% of the data is used to train the model, while 30% of the data is used to improve the forecast (Figure 16).

Split

Figure 16: Split of Train/Test

Randomization of train/test split

Create two data frames depending on the train and test split, using random sampling. To begin, construct the model using the training data. 70 observations are in the train set (data train) and 30 observations are in the test set (data test) based on the ratio. The probability of the randomization method is given in Figure 17 [12].

Randomization

Figure 17: Randomization of train/test split.

Classification of cyber vulnerabilities

Create a criminal element-based model. It depicts the percentage of the tool that is a major victim of cyberattack. This node checks for “Phishing,” “Web Jacking,” “Software Piracy,” “Cyber- Stalking,” “Cyber Bullying,” “Online Scams,” and “Profile Hacking” in the element. If true, then shift to the left child node of the root node. It will delve into the features and determine which ones have an impact on the risk of criminal factors. Figure 18 shows how the article label is categorized once the decision tree model is built based on the criminal label.

DT

Figure 18: DT classifier: Cybercrime labels [IPC label vs. Article label].

Accuracy test

The confusion matrix is used to evaluate the classification accuracy. The true negative is accounted for. Each column represents a predicted target, whereas each row represents an actual target. It calculated the projected value for test data using the train data as input. In #14 there is a comparison between the expected and actual value. The model’s correctness is assessed using a fine-tuning technique, and this model achieves a 53 percent accuracy level (Figure 19).

Accuracy

Figure 19: Accuracy test.

Discussion

The purpose of this paper is to examine the crime that was reported in the news article. The unstructured data retrieved from the news story was converted to structured data using multi-level crime categorization label mapping of the IPC legal code and a demographic of location and gender. Based on the analysis of the dataset (CVCL), the following conclusion can be drawn. The most common criminal article is “Phishing,” which is followed by Bullying and Cyber Stalking. “Jacking, webspace, and the last few are allegations of profile hacking and data theft,” according to the typical criminal record.

Qualitative insights-cyber news

The basic qualitative insights from the news given in the article are depicted using a word cloud. The cloud is visualized based on the word occurrence. Dark typefaces are used to highlight words with a high frequency. Small typefaces are used for the terms with the fewest occurrences. Figure 20 shows one example of this. The most frequently used terms in a news item or ‘crime,’ ‘bank,’ ‘account,’ and ‘online.’

Qualitative

Figure 20: Qualitative insights: Cyber news.

Domain wise inference

Email spoofing is determined to be the most common type of crime, and social networking sites are also used to fund cyberattacks. Financial services are the target of cyberattacks. This project aims to map cyber vulnerabilities in accordance with the IPC, IT Act 2000, as well as to comprehend the penalties and sections of the “IPC-Section 465”.

Location-wise analysis

Maharashtra is one of the most vulnerable states in India to cybercrime. The central zone of Delhi is the second most afflicted, whereas Karnataka is the most hit in the southern zone. Figure 21 depicts a crime analysis by location.

vulnerabilities

Figure 21: Cyber vulnerabilities-India

Gender wise analysis

In comparison to women, most men are reported to be quite vulnerable. It has received a lot of attention in Mumbai and New Delhi. Andhra Pradesh, Tamil Nadu, and Kerala are the southern states with the lowest reported incidences.

Year-wise analysis

Because of the tendencies of social networking and digitalized technologies in India, cyber susceptible crimes have increased in 2018.

Conclusion

Cybercrime vulnerability to the IPC Section is recommended as a specific goal in this article. The dataset was built using news stories extracted from top news publications that dealt with cybercrime. For the years 2012-2018, 100 articles have been downloaded. Cybercrime label mapping, IPC Sections, and demographics are all classified using a framework using a methodology. The Decision Tree Classifier was used to examine the results of the crime classification approach, and it was discovered that 100% of the crime labels are classed as Article labels.

Limitations

• The dataset has a minimum size that needs to be expanded, which is one of the study work’s drawbacks.

• The corpus developed for the research effort can be used to create an automation web crawler.

• This study does not take into account the mapping of security features in other cloud systems.

Future Scope

On future research, an automatic web crawler will be used.

Acknowledgement

I would like to express my deep gratitude to my research supervisors, for their patient guidance, enthusiastic encouragement and useful critiques of this research work.

Conflict of Interest

I P SUDHANDRADEVI, declare that no funds, grants, or other support were received during the preparation of this manuscript.

References

Author Info

Sudhandradevi P* and Bhuvaneswari V
 
Department of Mechanical Engineering, Indian Institute of Technology Patna, Bihar, India
 

Citation: P Sudhandradevi, V Bhuvaneswari (2022) Multilayered Map Reduce Framework to Link Cybernetic Vulnerabilities and Cybernetic Laws from E-News Articles. Int J Adv Technol. 13:166.

Received: 18-Mar-2022, Manuscript No. IJOAT-22-16296; Editor assigned: 23-Mar-2022, Pre QC No. IJOAT-22-16296 (PQ); Reviewed: 06-Apr-2022, QC No. IJOAT-22-16296; Revised: 11-Apr-2022, Manuscript No. IJOAT-22-16296 (R); Accepted: 11-Apr-2022 Published: 18-Apr-2022 , DOI: 10.35248/0976-4860-22.13.176

Copyright: © 2022 P Sudhandradevi. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

Top