Untitled
Untitled
Untitled
INDEX
Unit Page
Topics and Sub-topics
Number Number
Unit I 1.1 Introduction of AI 01
● Concept
● Scope of AI
● Components of AI
● Types of AI
● Application of AI
1.2 Data Visualization 11
● Data types in data visualization
● Scales map of data values in aesthetics
● Use of coordinate system in data visualization
● Use of colors to represent data values
● Representing - Amounts, Distribution, and Proportions
1.3 Data Storytelling 22
● Introduction
● Ineffectiveness of Graphical representation of data
● Explanatory Analysis
○ Who
○ What
○ How
1.4 Concept of machine learning and deep learning. 25
Unit II 2.1 Internet of Things (IoT ) 29
● Definition
● Characteristics of IoT
● Features and Applications of IoT
● Advantages and Disadvantages of IoT
2.1.2 Design of IoT 34
● Physical design of IoT
● Logical design of IoT
2.1.3 IoT Protocols
2.1.4 Sensors and actuators used in IoT 36
39
2.2 Introduction to 5G Network
46
5-G characteristics and application areas.
NGN architecture: Features, Functional block diagram,
Network components: Media Gateway, Media
Gateway Controller, and Application Server.
NGN Wireless Technology: Telecom network
Spectrum: Types [licensed and unlicensed],
Mobile Network Evolution (2G to 5G), Comparative
features,
NGN Core: Features, Multi-Protocol Label Switching
(MPLS): Concepts, Features and Advantages.
Unit III 3.1 Introduction to Blockchain 67
● Backstory of Blockchain
● What is Blockchain?
3.2 Centralize versus Decentralized System 70
3.3 Layers of Blockchain 72
● Application Layer
● Execution Layer
● Semantic Layer
● Propagation Layer
● Consensus Layer
3.4 Importance of Blockchain 74
● Limitations of Centralized Systems
● Blockchain Adoption So Far
3.5 Blockchain Use and Use Cases 75
Unit IV 4.1 Digital forensics 78
● Introduction to digital forensic
● Digital forensics investigation process 80
● Models of Digital Forensic Investigation - 80
o Abstract Digital Forensics Model (ADFM)
o Integrated Digital Investigation Process (IDIP)
o An extended model for cybercrime investigation
4.2 Ethical issues in digital forensic 81
● General ethical norms for investigators
● Unethical norms for investigation
4.3 Digital Evidences 82
● Definition of Digital Evidence
● Best evidence rule
● Original Evidence
4.4 Characteristics of Digital Evidence 85
● Locard’s Exchange Principle
● Digital Stream of bits
4.5 Types of evidence 87
● Illustrative, Electronics, Documented, Explainable,
Substantial, Testimonial
4.6 Challenges in evidence handling 87
● Authentication of evidence
● Chain of custody
● Evidence validation
4.7 Volatile evidence 92
Unit V 5.1 Ethical Hacking 102
● How Hackers Beget Ethical Hackers
● Defining hacker, Malicious users
● Data Privacy and General Data Protection and
Regulation(GDPR)
5.2 Understanding the need to hack your own systems 112
5.3 Understanding the dangers your systems face 114
● Non Technical attacks
● Network-infrastructure attacks
● Operating-system attacks
● Application and other specialized attacks
5.4 Obeying the Ethical hacking Principles 115
● Working ethically
● Respecting privacy
● Not crashing your systems
5.5 The Ethical Hacking Process 116
● Formulating your plan
● Selecting tools
● Executing the plan
● Evaluating results
● Moving on
5.6 Cyber Security act 120
Unit VI 6.1 Network Hacking 124
Network Infrastructure: 125
● Network Infrastructure Vulnerabilities
● Scanning-Ports
● Ping sweep
● Scanning SNMP
● Grabbing Banners
● MAC-daddy attack
Wireless LANs: 133
● Wireless Network Attacks
6.2 Operating System Hacking 135
● Introduction of Windows and Linux Vulnerabilities
● Buffer Overflow Attack
● 6.3 Applications Hacking 139
Messaging Systems:
● Vulnerabilities
● E-Mail Attacks- E-Mail Bombs
● Banners
● Best practices for minimizing e-mail security risks
Web Applications:
● Web Vulnerabilities
● Directories Traversal and Countermeasures
● Google Dorking
Database system
● Database Vulnerabilities
● Best practices for minimizing database security risks
Emerging Trends in CO and IT (22618)
Contents
1.1 Introduction of AI
● Concept
● Scope of AI
● Components of AI
● Types of AI
● Application of AI
1.2 Data Visualization
● Data types in data visualization
● Scales map of data values in aesthetics
● Use of coordinate system in data visualization
● Use of colors to represent data values
● Representing - Amounts, Distribution, and Proportions
1.3 Data Storytelling
● Introduction
● Ineffectiveness of Graphical representation of data
● Explanatory Analysis
○ Who
○ What
○ How
1.4 Concept of machine learning and deep learning.
1.1 Introduction of AI
A branch of Computer Science named Artificial Intelligence (AI)pursues creating the
computers / machines as intelligent as human beings. John McCarthy the father of Artificial
Intelligence described AI as, “The science and engineering of making intelligent
machines, especially intelligent computer programs”. Artificial Intelligence (AI) is a
branch of science which deals with helping machines find solutions to complex problems
in a more human-like fashion.
Artificial is defined in different approaches by various researchers during its evolution,
such as “Artificial Intelligence is the study of how to make computers do things which at
the moment, people do better.”
There are other possible definitions “like AI is a collection of hard problems which can be
solved by humans and other living things, but for which we don’t have good algorithms for
solving.” e. g., understanding spoken natural language, medical diagnosis, circuit design,
learning, self-adaptation, reasoning, chess playing, proving math theories, etc.
● Data: Data is defined as symbols that represent properties of objects events and their
environment.
● Information: Information is a message that contains relevant meaning, implication, or
input for decision and/or action.
● Knowledge: It is the (1) cognition or recognition (know-what), (2) capacity to
act(know-how), and (3) understanding (know-why) that resides or is contained within
the mind or in the brain.
● Intelligence: It requires ability to sense the environment, to make decisions, and to
control action.
1.1.1 Concept:
Artificial Intelligence is one of the emerging technologies that try to simulate human
reasoning in AI systems The art and science of bringing learning, adaptation and self-
organization to the machine is the art of Artificial Intelligence. Artificial Intelligence is the
ability of a computer program to learn and think. Artificial intelligence (AI) is an area of
computer science that emphasizes the creation of intelligent machines that work and reacts
like humans. AI is built on these three important concepts
Machine learning: When you command your smartphone to call someone, or when you chat
with a customer service chatbot, you are interacting with software that runs on AI. But this
type of software actually is limited to what it has been programmed to do. However, we expect
to soon have systems that can learn new tasks without humans having to guide them. The idea
is to give them a large number of examples for any given chore, and they should be able to
process each one and learn how to do it by the end of the activity.
Deep learning: The machine learning example I provided above is limited by the fact that
humans still need to direct the AI’s development. In deep learning, the goal is for the software
to use what it has learned in one area to solve problems in other areas. For example, a program
that has learned how to distinguish images in a photograph might be able to use this learning
to seek out patterns in complex graphs.
Neural networks: These consist of computer programs that mimic the way the human brain
processes information. They specialize in clustering information and recognizing complex
patterns, giving computers the ability to use more sophisticated processes to analyze data.
AI Approach:
The difference between machine and human intelligence is that the human think / act
rationally compares to machine. Historically, all four approaches to AI have been followed,
each by different people with different methods.
1.1.3 Components of AI
The core components and constituents of AI are derived from the concept of logic, cognition
and computation; and the compound components, built-up through core components are
knowledge, reasoning, search, natural language processing, vision etc.
Level Core Compound Coarse components
Induction Proposition
Knowledge Reasoning Knowledge based systems, Heuristic
Logic Tautology Model
Control Search Search Theorem Proving
Logic
Temporal Learning
Multi Agent system Co-operation,
Cognition Adaptation Belief Desire Intention
Co-ordination AI Programming
Self-organization
Vision
Functional Memory Perception Utterance Natural Language Speech
Processing
The core entities are inseparable constituents of AI in that these concepts are fused at atomic
level. The concepts derived from logic are propositional logic, tautology, predicate calculus,
model and temporal logic. The concepts of cognitive science are of two types: one is
functional which includes learning, adaptation and self-organization, and the other is memory
and perception which are physical entities. The physical entities generate some functions to
make the compound components.
The compound components are made of some combination of the logic and cognition stream.
These are knowledge, reasoning and control generated from constituents of logic such as
predicate calculus, induction and tautology and some from cognition (such as learning and
adaptation). Similarly, belief, desire and intention are models of mental states that are
predominantly based on cognitive components but less on logic. Vision, utterance (vocal) and
expression (written) are combined effect of memory and perceiving organs or body sensors
such as ear, eyes and vocal. The gross level contains the constituents at the third level which
are knowledge-based systems (KBS), heuristic search, automatic theorem proving, multi-
agent systems, Al languages such as PROLOG and LISP, Natural language processing (NLP).
Speech processing and vision are based mainly on the principle of pattern recognition.
AI Dimension: The philosophy of Al in three-dimensional representations consists in logic,
cognition and computation in the x-direction, knowledge, reasoning and interface in the y-
direction. The x-y plane is the foundation of AI. The z-direction consists of correlated systems
of physical origin such as language, vision and perception as shown in Figure.1.2
Cognition:
Computers has become so popular in a short span of time due to the simple reason that
they adapted and projected the information processing paradigm (IPP) of human beings:
sensing organs as input, mechanical movement organs as output and the central nervous
system (CNS) in brain as control and computing devices, short-term and long-term
memory were not distinguished by computer scientists but, as a whole, it was in
conjunction, termed memory.
In further deepening level, the interaction of stimuli with the stored information to
produce new information requires the process of learning, adaptation and self-
organization. These functionalities in the information processing at a certain level of
abstraction of brain activities demonstrate a state of mind which exhibits certain specific
behavior to qualify as intelligence. Computational models were developed and
incorporated in machines which mimicked the functionalities of human origin. The
creation of such traits of human beings in the computing devices and processes
originated the concept of intelligence in machine as virtual mechanism. These virtual
machines were termed in due course of time artificial intelligent machines.
Computation
The theory of computation developed by Turing-finite state automation—was a turning
point in mathematical model to logical computational. Chomsky's linguistic
computational theory generated a model for syntactic analysis through a regular
grammar.
1.1.4 Types of AI
Artificial Intelligence can be divided in various types, there are mainly two types of
main categorization which are based on capabilities and based on functionally of AI.
Following is flow diagram which explain the types of AI.
Types of AI
think, to reason, solve the puzzle, make judgments, plan, learn, and communicate
by its own.
● Super AI is still a hypothetical concept of Artificial Intelligence. Development
of such systems in real is still world changing task.
2. Theory of Mind
● Theory of Mind AI should understand the human emotions, people, beliefs, and
be able to interact socially like humans.
● This type of AI machines are still not developed, but researchers are making
lots of efforts and improvement for developing such AI machines.
3. Self-Awareness
● Self-awareness AI is the future of Artificial Intelligence. These machines will
be super intelligent, and will have their own consciousness, sentiments, and self-
awareness.
● These machines will be smarter than human mind.
● Self-Awareness AI does not exist in reality still and it is a hypothetical concept.
1.1.5 Application of AI
AI has been dominant in various fields such as −
● Gaming: AI plays crucial role in strategic games such as chess, poker, tic-tac-
toe, etc., where machine can think of large number of possible positions based on
heuristic knowledge.
● Natural Language Processing: It is possible to interact with the computer that
understands natural language spoken by humans.
● Expert Systems: There are some applications which integrate machine,
software, and special information to impart reasoning and advising. They provide
explanation and advice to the users.
● Vision Systems: These systems understand, interpret, and comprehend visual
input on the computer. For example,
• A spying aeroplane takes photographs, which are used to figure out spatial
information or map of the areas.
• Doctors use clinical expert system to diagnose the patient.
• Police use computer software that can recognize the face of criminal with the
stored portrait made by forensic artist.
● Speech Recognition: Some intelligent systems are capable of hearing and
comprehending the language in terms of sentences and their meanings while a
human talks to it. It can handle different accents, slang words, noise in the
background, change in human’s noise due to cold, etc.
● Handwriting Recognition: The handwriting recognition software reads the text
written on paper by a pen or on screen by a stylus. It can recognize the shapes of
the letters and convert it into editable text.
● Intelligent Robots: Robots are able to perform the tasks given by a human.
They have sensors to detect physical data from the real world such as light, heat,
temperature, movement, sound, bump, and pressure. They have efficient
processors, multiple sensors and huge memory, to exhibit intelligence. In addition,
they are capable of learning from their mistakes and they can adapt to the new
environment.
1.2.1 Introduction –
Data visualization is the graphical representation of information and data. By
using visual elements like charts, graphs, and maps, data visualization tools
provide an accessible way to see and understand trends, outliers, and patterns in
data. It also provides an excellent way to present data to non-technical audiences
without confusion. The first and foremost objective of data visualization is to
convey data correctly. Whenever we visualize data, we take data values and
convert them in a systematic and logical way into the visual elements that make
up the final graphic. Even though there are many different types of data
visualizations, and on first glance a scatterplot, a pie chart, and a heatmap don’t
seem to have much in common, all these visualizations can be described with a
common language that captures how data values are turned into blobs of ink on
paper or colored pixels on a screen. The key insight is the following: all data
visualizations map data values into quantifiable features of the resulting graphic.
We refer to these features as aesthetics.
1.2.2 Data types in data visualization –
When we consider types of data in data visualization, we consider various types
of data in use as well as aesthetics too. Aesthetics describe every aspect of a given
graphical element. For example, in Figure 1.4 -
A critical component of every graphical element is of course its position, which
describes where the element is located. In standard 2D graphics, we describe
positions by an x and y value, but other coordinate systems and one- or three-
dimensional visualizations are possible. Next, all graphical elements have a shape,
a size, and a color. Even if we are preparing a black-and-white drawing, graphical
elements need to have a color to be visible: for example, black if the background
is white or white if the background is black. Finally, to the extent we are using
lines to visualize data, these lines may have different widths or dash–dot patterns.
There are many other aesthetics may encountered in a data visualization. For
example, if we want to display text, we may have to specify font family, font face,
and font size, and if graphical objects overlap, we may have to specify whether
they are partially transparent.
Figure 1.4 Commonly used aesthetics in data visualization: position, shape, size,
color, line width, line type. Some of these aesthetics can represent both continuous
and discrete data (position, size, line width, color), while others can usually only
represent
50 seconds and 51 seconds, there are arbitrarily many intermediates, such as 50.5
seconds, 50.51 seconds, 50.50001 seconds, and so on. By contrast, number of
persons in a room is a discrete value. A room can hold 5 persons or 6, but not 5.5.
For the examples in Figure 1.4, position, size, color, and line width can represent
continuous data, but shape and line type can usually only represent discrete data.
Next, we’ll consider the types of data we may want to represent in our
visualization. You may think of data as numbers, but numerical values are only
two out of several types of data we may encounter. In addition to continuous and
discrete numerical values, data can come in the form of discrete categories, in the
form of dates or times, and as text (Table 1.1). When data is numerical, we also
call it quantitative and when it is categorical, we call it qualitative. Variables
holding qualitative data are factors, and the different categories are called levels.
The levels of a factor are most commonly without order (as in the example of dog,
cat, fish in Table 1.1 given below, but factors can also be ordered, when there is
an intrinsic order among the levels of the factor (as in the example of good, fair,
poor in Table 1.1).
Table 1.1 Types of variables encountered in Data Visualization Scenario
Types of Appropriate
Example Description
Variables Scale
Quantitative/ 1.3, 5.7, 83, Arbitrary numerical values. These
numerical 1.5 × Continuous can be integers, rational numbers, or
continuous 10–2 real numbers.
Numbers in discrete units. These are
Quantitative/ most commonly but not necessarily
numerical integers. For example, the numbers
discrete 1, 2, 3, 4 Discrete 0.5, 1.0, 1.5 could also be treated as
discrete if intermediate values cannot
exist in the given dataset.
Categories without order. These are
Qualitative/
discrete and unique
categorical
dog, cat, fish Discrete categories that have no inherent
unordered
order. These variables are also called
factors.
Categories with order. These are
Qualitative/ discrete and unique
categorical good, fair, categories with an order. For
ordered Discrete example, “fair” always lies between
poor
“good” and “poor.” These variables
are also called ordered factors.
Continuous Specific days and/or times. Also,
Jan. 5 2018, or
Date or time generic dates, such as July 4 or Dec.
8:03am
Discrete 25 (without year).
The quick
brown fox None, or Free-form text. Can be treated as
Text
jumps over discrete categorical if needed.
the lazy dog.
Let’s consider an example, the below Table 1.2 shows the first few rows of a dataset
providing the daily temperature normal (aver‐ age daily temperatures over a 30-year
window) for four US locations. This table contains five variables: month, day,
location, station ID, and temperature (in degrees Fahrenheit). Month is an ordered
factor, day is a discrete numerical value, location is an unordered factor, station ID
is similarly an unordered factor, and temperature is a continuous numerical value.
Table 1.2 First 8 rows of a dataset listing daily temperature normal for four weather stations
Temperature
Month Day Location Section ID
(F)
USW000148
Jan 1 Chicago 25.6
19
USW000931
Jan 1 San Diego 55.2
07
USW000129
Jan 1 Houston 53.9
18
Death USC0004231
Jan 1 51.0
Valley 9
USW000148
Jan 2 Chicago 25.5
19
USW000931
Jan 2 San Diego 55.3
07
USW000129
Jan 2 Houston 53.8
18
Death USC0004231
Jan 2 51.2
Valley 9
Data source: National Oceanic and Atmospheric Administration (NOAA).
Figure 1.6 Monthly normal mean temperatures for the same example
1.2.3 Use of coordinate system in Data Visualization
To make any sort of data visualization, we need to define position scales, which
deter‐ mine where in graphic different data values are located. We cannot visualize
data without placing different data points at different locations, even if we just arrange
them next to each other along a line. For regular 2D visualizations, two numbers are
required to uniquely specify a point, and therefore we need two position scales. These
two scales are usually but not necessarily the x and y axes of the plot. We also have to
specify the relative geometric arrangement of these scales. Conventionally, the x axis
runs horizontally and the y axis vertically, but we could choose other arrangements.
For example, we could have the y axis run at an acute angle relative to the x axis, or
we could have one axis run in a circle and the other run radially. The combination of
a set of position scales and their relative geometric arrangement is called a coordinate
system.
● Cartesian coordinates –
The most widely used coordinate system for data visualization is the 2D Cartesian
coordinate system, where each location is uniquely specified by an x and a y value.
The x and y axes run orthogonally to each other, and data values are placed in an even
spacing along both axes. The two axes are continuous position scales, and they can
represent both positive and negative real numbers. To fully specify the coordinate
system, we need to specify the range of numbers each axis covers. Any data values
between these axis limits are placed at the appropriate respective location in the plot.
Maharashtra State Board of Technical Education P a g e 14 | 151
Emerging Trends in CO and IT (22618)
Figure 1.8 Daily temperature normals for Huston using different aspect ratio
A Cartesian coordinate system can have two axes representing two different units. This
situation arises quite commonly whenever we’re mapping two different types of
variables to x and y. For example, consider below image, if we plot temperature versus
days of the year. The y axis of is measured in degrees Fahrenheit, with a grid line every
at 20 degrees, and the x axis is measured in months, with a grid line at the first of every
third month. Whenever the two axes are measured in different units, we can stretch or
compress one relative to the other and maintain a valid visualization of the data. Which
version is preferable may depend on the story we want to convey. A tall and narrow
figure emphasizes change along the y axis and a short and wide figure does the opposite.
Ideally, we want to choose an aspect ratio that ensures that any important differences in
position are noticeable.
● Nonlinear Axes –
In a Cartesian coordinate system, the grid lines along an axis are spaced evenly both in
data units and in the resulting visualization. We refer to the position scales in these
Maharashtra State Board of Technical Education P a g e 15 | 151
Emerging Trends in CO and IT (22618)
are in Death Valley, Houston, and San Diego from late fall to early spring. In the
Cartesian coordinate system, this fact is obscured because the temperature values in late
December and in early January are shown in opposite parts of the figure and therefore
don’t form a single visual unit.
Such a scale contains a finite set of specific colors that are chosen to look clearly
distinct from each other while also being equivalent to each other. The second
condition requires that no one color should stand out relative to the others. Also,
the colors should not create the impression of an order, as would be the case with
a sequence of colors that get successively lighter. Such colors would create an
apparent order among the items being colored, which by definition have no order.
Many appropriate qualitative color scales are readily available. Figure 4-1 shows
three representative examples. In particular, the ColorBrewer project provides a
nice selection of qualitative color scales, including both fairly light and fairly dark
colors [Brewer 2017].
Figure 1.11. Example qualitative color scales. The Okabe Ito scale is the default scale
used throughout this book [Okabe and Ito 2008]. The ColorBrewer Dark2 scale is
provided by the ColorBrewer project [Brewer 2017]. The ggplot2 hue scale is the
default qualitative scale in the widely used plotting software ggplot2.
Figure 1.13. Example sequential color scales. The ColorBrewer Blues scale is a monochro‐
matic scale that varies from dark to light blue. The Heat and Viridis scales are multihue
scales that vary from dark red to light yellow and from dark blue via green to light yel‐
low, respectively.
iii. Color as a tool to highlight –
Color can also be an effective tool to highlight specific elements in the data. There may
be specific categories or values in the dataset that carry key information about the story
we want to tell, and we can strengthen the story by emphasizing the relevant figure
elements to the reader. An easy way to achieve this emphasis is to color these figure
elements in a color or set of colors that vividly stand out against the rest of the figure.
This effect can be achieved with accent color scales, which are color scales that contain
both a set of subdued colors and a matching set of stronger, darker, and/or more
saturated colors.
Figure 4-7. Example accent color scales, each with four base colors and three accent col‐ ors.
Accent color scales can be derived in several different ways: (top) we can take an existing color
scale (e.g., the Okabe Ito scale) and lighten and/or partially desaturate some colors while
darkening others; (middle) we can take gray values and pair them with colors; (bottom) we can
use an existing accent color scale (e.g., the one from the ColorBrewer project).
The most common approach to visualizing amounts (i.e., numerical values shown for
some set of categories) is using bars, either vertically or horizontally. However, instead
of using bars, we can also place dots at the location where the corresponding bar would
end.
If there are two or more sets of categories for which we want to show amounts, we can
group or stack the bars. We can also map the categories onto the x and y axes and
show amounts by color, via a heatmap.
ii. Distributions
Histograms and density plots provide the most intuitive visualizations of a distribution,
Maharashtra State Board of Technical Education P a g e 20 | 151
Emerging Trends in CO and IT (22618)
but both require arbitrary parameter choices and can be misleading. Cumulative
densities and quantile-quantile (q-q) plots always represent the data faithfully but can
be more difficult to interpret.
Boxplots, violin plots, strip charts, and since plots are useful when we want to visualize
many distributions at once and/or if we are primarily interested in overall shifts among
the distributions. Stacked histograms and overlapping densities allow a more in-depth
com‐ parison of a smaller number of distributions, though stacked histograms can be
difficult to interpret and are best avoided. Ridgeline plots can be a useful alternative to
violin plots and are often useful when visualizing very large numbers of distributions
or changes in distributions over.
iii. Proportions
Proportions can be visualized as pie charts, side-by-side bars, or stacked bars. As for
amounts, when we visualize proportions with bars, the bars can be arranged either
vertically or horizontally. Pie charts emphasize that the individual parts add up to a
whole and highlight simple fractions. However, the individual pieces are more easily
compared in side-by-side bars. Stacked bars look awkward for a single set of
proportions, but can be useful when comparing multiple sets of proportions.
When visualizing multiple sets of proportions or changes in proportions across
conditions, pie charts tend to be space-inefficient and often obscure relationships.
Grouped bars work well as long as the number of conditions compared is moderate, and
stacked bars can work for large numbers of conditions. Stacked densities are appropriate
when the proportions change along a continuous variable.
When proportions are specified according to multiple grouping variables, mosaic plots,
tree maps, or parallel sets are useful visualization approaches. Mosaic plots assume that
every level of one grouping variable can be combined with every level of another
grouping variable, whereas tree maps do not make such an assumption. Tree maps work
well even if the subdivisions of one group are entirely distinct from the subdivisions of
another. Parallel sets work better than either mosaic plots or tree maps when there are
more than two grouping variables.
1.3 Data Storytelling
1.3.1 Introduction
Data storytelling is a methodology for communicating information, tailored to a specific
audience, with a compelling narrative. It is the last ten feet of your data analysis and
arguably the most important aspect. Data storytelling is the concept of building a
compelling narrative based on complex data and analytics that help tell your story and
influence and inform a particular audience.
● The benefits of data storytelling
✔ Adding value to your data and insights.
✔ Interpreting complex information and highlighting essential key points for the
audience.
✔ Providing a human touch to your data.
✔ Offering value to your audience and industry.
✔ Building credibility as an industry and topic thought leader.
1.3.2. Ineffectiveness of Graphical representation of data
Data visualization plays a significant role in determining how receptive your audience
is to receiving complex information. Data visualization helps transform boundless
amounts of data into something simpler and digestible. Here, you can supply the visuals
needed to support your story. Effective data visualizations can help:
● Reveal patterns, trends, and findings from an unbiased viewpoint.
● Provide context, interpret results, and articulate insights.
● Streamline data so your audience can process information.
● Improve audience engagement.
data. As you create your data story, it is important to combine the following three
elements to write a well-rounded anecdote of your theory and the resulting actions you’d
like to see from users.
1. Build your narrative
As you tell your story, you need to use your data as supporting pillars to your insights.
Help your audience understand your point of view by distilling complex information
into informative insights. Your narrative and context are what will drive the linear
nature of your data storytelling.
2. Use visuals to enlighten
Visuals can help educate the audience on your theory. When you connect the visual
assets (charts, graphs, etc.) to your narrative, you engage the audience with otherwise
hidden insights that provide the fundamental data to support your theory. Instead of
presenting a single data insight to support your theory, it helps to show multiple pieces
of data, both granular and high level, so that the audience can truly appreciate your
viewpoint.
3. Show data to support
Humans are not naturally attracted to analytics, especially analytics that lack
contextualization using augmented analytics. Your narrative offers enlightenment,
supported by tangible data. Context and critique are integral to the full interpretation of
your narrative. Using business analytic tools to provide key insights and understanding
to your narrative can help provide the much-needed context throughout your data story.
By combining the three elements above, your data story is sure to create an emotional
response in your audience. Emotion plays a significant role in decision-making. And by
linking the emotional context and hard data in your data storytelling, you’re able to
influence others. When these three key elements are successfully integrated, you have
created a data story that can influence people and drive change.
o Your audience - The more specific you can be about who your audience is, the
better position you will be in for successful communication. Avoid general
audiences, such as “internal and external stakeholders” or “anyone who might be
interested”—by trying to communicate to too many different people with disparate
needs at once, you put yourself in a position where you can’t communicate to any
one of them as effectively as you could if you narrowed your target audience.
Sometimes this means creating different communications for different audiences.
Identifying the decision maker is one way of narrowing your audience. The more
you know about your audience, the better positioned you’ll be to understand how
to resonate with them and form a communication that will meet their needs and
yours.
o You - It’s also helpful to think about the relationship that you have with your
audience and how you expect that they will perceive you. Will you be encountering
each other for the first time through this communication, or do you have an
established relationship? Do they already trust you as an expert, or do you need to
work to establish credibility? These are important considerations when it comes to
determining how to structure your communication and whether and when to use
data, and may impact the order and flow of the overall story you aim to tell.
1.3.4.2. What -
o Action - What do you need your audience to know or do? This is the point where
you think through how to make what you communicate relevant for your audience
and form a clear understanding of why they should care about what you say. You
should always want your audience to know or do something. If you can’t concisely
articulate that, you should revisit whether you need to communicate in the first place.
o Mechanism - How will you communicate to your audience? The method you will
use to communicate to your audience has implications on a number of factors,
including the amount of control you will have over how the audience takes in the
information and the level of detail that needs to be explicit. We can think of the
communication mechanism along a continuum, with live presentation at the left and
a written document or email at the right, as shown in Figure 1.1. Consider the level
of control you have over how the information is consumed as well as the amount of
detail needed at either end of the spectrum.
1.3.4.3. How -
Finally—and only after we can clearly articulate who our audience is and what we
need them to know or do—we can turn to the data and ask the question: What data is
available that will help make my point? Data becomes supporting evidence of the
story you will build and tell.
1.4 Concept of machine learning and deep learning
1.4.1 Machine Learning:
● Machine learning is a branch of science that deals with programming the systems
in such a way that they automatically learn and improve with experience. Here, learning
Maharashtra State Board of Technical Education P a g e 24 | 151
Emerging Trends in CO and IT (22618)
means recognizing and understanding the input data and making wise decisions based
on the supplied data.
● It is very difficult to cater to all the decisions based on all possible inputs. To
tackle this problem, algorithms are developed. These algorithms build knowledge from
specific data and past experience with the principles of statistics, probability theory,
logic, combinatorial optimization, search, reinforcement learning, and control theory.
The developed algorithms form the basis of various applications such as:
● Vision processing
● Language processing
● Forecasting (e.g., stock market trends)
● Pattern recognition
● Games
● Data mining
● Expert systems
● Robotics
Machine learning is a vast area and it is quite beyond the scope of this tutorial to cover
all its features. There are several ways to implement machine learning techniques,
however the most commonly used ones are supervised and unsupervised learning.
1.4.2. Supervised Learning: Supervised learning deals with learning a function from
available training data. A supervised learning algorithm analyzes the training data and
produces an inferred function, which can be used for mapping new examples. Common
examples of supervised learning include:
● classifying e-mails as spam,
● labeling webpages based on their content, and
● voice recognition.
There are many supervised learning algorithms such as neural networks, Support Vector
Machines (SVMs), and Naive Bayes classifiers. Mahout implements Naive Bayes
classifier.
1.4.3. Unsupervised Learning: Unsupervised learning makes sense of unlabeled data
without having any predefined dataset for its training. Unsupervised learning is an
extremely powerful tool for analyzing available data and look for patterns and trends.
It is most commonly used for clustering similar input into logical groups. Common
approaches to unsupervised learning include:
● k-means
● self-organizing maps, and
● hierarchical clustering
Each algorithm in deep learning goes through the same process. It includes a hierarchy
of nonlinear transformation of input that can be used to generate a statistical model as
output. Consider the following steps that define the Machine Learning process
● Identifies relevant data sets and prepares them for analysis.
● Chooses the type of algorithm to use
● Builds an analytical model based on the algorithm used.
● Trains the model on test data sets, revising it as needed.
● Runs the model to generate test scores.
Deep learning has evolved hand-in-hand with the digital era, which has brought about
an explosion of data in all forms and from every region of the world. This data, known
simply as big data, is drawn from sources like social media, internet search engines, e-
commerce platforms, and online cinemas, among others. This enormous amount of data
is readily accessible and can be shared through fintech applications like cloud
computing.
However, the data, which normally is unstructured, is so vast that it could take decades
for humans to comprehend it and extract relevant information. Companies realize the
incredible potential that can result from unraveling this wealth of information and are
increasingly adapting to AI systems for automated support.
References:
● https://www.tutorialspoint.com/artificial_intelligence/artificial_intelligence
_overview. htm
● https://www.javatpoint.com/introduction-to-artificial-intelligence
● https://www.tutorialspoint.com/tensorflow/tensorflow_machine_learning_d
eep_learni ng.htm
● Story telling with data by Cole Nissbuamer Knafilc – Wiley Publication -
ISBN 9781119002253
● Fundamentals of Data Visualization, A primer on making informative and
compelling figures by Claus O Wilke - O’Reilly Publication – March 2019
b. General audience
c. Specific audience
d. Data specific audience
Content
2.1 Internet of Things (IoT )
● Definition
● Characteristics of IoT
● Features of IoT
● Advantages and Disadvantages of IoT
2.1.2 Design of IoT
● Physical design of IoT
● Logical design of IoT
2.1.3 IoT Protocols
2.1.4 Sensors and actuators used in IoT
2.2 Introduction to 5G Network
● 5-G characteristics and application areas.
● NGN architecture: Features, Functional block diagram, Network components:
Media Gateway, Media Gateway Controller, and Application Server.
● NGN Wireless Technology: Telecom network Spectrum: Types [licensed
and unlicensed], Mobile Network Evolution (2G to 5G), Comparative features,
● NGN Core: Features, Multi-Protocol Label Switching (MPLS): Concepts,
Features and Advantages.
● Smart Parking: make the search for parking space easier and convenient for
drivers. Smart parking are powered by IoT systems that detect the no. of empty
parking slots and send information over internet to smart application back ends.
● Smart Lighting: for roads, parks and buildings can help in saving energy.
● Smart Roads: Equipped with sensors can provide information on driving
condition, travel time estimating and alert in case of poor driving conditions,
traffic condition and accidents.
● Structural Health Monitoring: uses a network of sensors to monitor the
vibration levels in the structures such as bridges and buildings.
● Surveillance: The video feeds from surveillance cameras can be aggregated in
cloud based scalable storage solution.
● Emergency Response:IoT systems for fire detection, gas and water leakage
detection can help in generating alerts and minimizing their effects on the critical
infrastructures.
c. Environment:
● Weather Monitoring: Systems collect data from a no. of sensors attached and
send the data to cloud based applications and storage back ends. The data
collected in cloud can then be analyzed and visualized by cloud based
applications.
● Air Pollution Monitoring: System can monitor emission of harmful gases
(CO2, CO, NO, NO2 etc.,) by factories and automobiles using gaseous and
meteorological sensors. The collected data can be analyzed to make informed
decisions on pollutions control approaches.
● Noise Pollution Monitoring: Due to growing urban development, noise levels
in cities have increased and even become alarmingly high in some cities. IoT
based noise pollution monitoring systems use a no. of noise monitoring systems
that are deployed at different places in a city. The data on noise levels from the
station is collected on servers or in the cloud. The collected data is then
aggregated to generate noise maps.
● Forest Fire Detection: Forest fire can cause damage to natural resources,
property and human life. Early detection of forest fire can help in minimizing
damage.
● River Flood Detection: River floods can cause damage to natural and human
resources and human life. Early warnings of floods can be given by monitoring
the water level and flow rate. IoT based river flood monitoring system uses a no.
of sensor nodes that monitor the water level and flow rate sensors.
d. Retail:
● Inventory Management: IoT systems enable remote monitoring of inventory
using data collected by RFID readers.
● Smart Payments: Solutions such as contact-less payments powered by
technologies such as Near Field Communication(NFC) and Bluetooth.
Maharashtra State Board of Technical Education P a g e 31 | 151
Emerging Trends in CO and IT (22618)
b. Disadvantages of IoT
● Security: As the IoT systems are interconnected and communicate over
networks. The system offers little control despite any security measures, and it
can be lead the various kinds of network attacks.
● Privacy: Even without the active participation on the user, the IoT system
provides substantial personal data in maximum detail.
● Complexity: The designing, developing, and maintaining and enabling the
large technology to IoT system is quite complicated.
● Flexibility: Many are concerned about the flexibility of an IoT system to
integrate easily with another. They worry about finding themselves with several
conflicting or locked systems.
● Compliance: IoT, like any other technology in the realm of business, must
comply with regulations. Its complexity makes the issue of compliance seem
incredibly challenging when many consider standard software compliance a
battle.
2.4 GHz band, 802.11n operates in the 2.4/5 GHz bands, 802.11ac operates in
the 5 GHz band and 802.11ad operates in the 60 GHz band. These standards
provide data rates from 1 Mb/s to upto 6.75 Gb/s.
3. 802.16-WiMax: IEEE 802.16 is a collection of wireless broadband standards,
including extensive descriptions for the link layer (also called WiMax).
WiMaxstandards provide data rates from 1.5 Mb/s to 1 Gb/s. The recent update
(802.16m) provides data rates of 100 Mbit/s for mobile stations and 1 Gbit/s for
fixed stations.
4. 802.15.4-LR-WPAN: IEEE 802.15.4 is a collection of standards for low-rate
wireless personal area networks (LR-WPANs). These standards form the basis
of specifications for high level communication protocols such as ZigBee. LR-
WPAN standards provide data rates from 40 Kb/s 250 Kb/s. These standards
provide low-cost and low-speed communication for power constrained devices.
5. 2G/3G/4G - Mobile Communication: There are different generations of
mobilecommunication standards including second generation (2G including
GSM and CDMA), third generation (3G - including UMTS and CDMA2000)
and fourth generation (4G - including LTE). IoT devices based on these
standards can communicate over cellular networks. Data rates for these standards
range from 9.6 Kb/s (for 2G) to upto 100 Mb/s (for 4G) and are available from
the 3GPP websites.
b. Network/Internet Layer Protocols:
The network layers are responsible for sending of IP datagrams from the source network
to the destination network. This layer performs the host addressing and packet routing.
The datagrams contain the source and destination addresses which are used to route
them from the source to destination across multiple networks. Host identification is
done using hierarchical IP addressing schemes such as IPv4 or IPv6.
1. IPv4: Internet Protocol version 4 (IPv4) is the most deployed Internet protocol
that is used to identify the devices on a network using a hierarchical addressing
scheme. IPv4 uses a 32-bit address scheme that allows total of 232 or
4,294,967,296 addresses. IPv4 has been succeeded by IPv6. The IP protocols
establish connections on packet networks, but do not guarantee delivery of
packets. Guaranteed delivery and data integrity are handled by the upper layer
protocols (such as TCP).
2. IPv6: Internet Protocol version 6 (IPv6) is the newest version of Internet
protocol and successor to IPv4, IPv6 uses 128-bit address scheme that allows
total of 2128 or 3.4 x 1038 addresses.
3. 6LOWPAN: 6LOWPAN (IPv6 over Low Power Wireless Personal Area
Networks) brings IP protocol to the low-power devices which have limited
processing capability. 6LOWPAN operates in the 2.4 GHz frequency range and
provides data transfer rates of 250 Kb/s. 6LOWPAN works with the 802.15.4
link layer protocol and defines compression mechanisms for IPv6 datagrams
over IEEE 802.15.4-based networks.
Maharashtra State Board of Technical Education P a g e 36 | 151
Emerging Trends in CO and IT (22618)
2. Humidity sensors: The amount of water vapour in air, or humidity, can affect
human comfort as well as many manufacturing processes in industries. So
monitoring humidity level is important. Most commonly used units for humidity
measurement are relative humidity (RH), dew/frost point (D/F PT) and parts per
million (PPM) as shown in Figure 2.5.
3. Motion sensors: Motion sensors are not only used for security purposes but also
in automatic door controls, automatic parking systems, automated sinks,
automated toilet flushers, hand dryers, energy management systems, etc. You
use these sensors in the IoT and monitor them from your smartphone or
computer. HC-SR501 passive infrared (PIR) sensor is a popular motion sensor
for hobby projects as shown in Figure 2.6.
5. Smoke sensors: Smoke detectors have been in use in homes and industries for
quite a long time as shown in Fig 2.8. With the advent of the IoT, their
application has become more convenient and user-friendly. Furthermore, adding
a wireless connection to smoke detectors enables additional features that increase
safety and convenience.
1. Servo motors: A Servo is a small device that incorporates a two wire DC motor
as shown in Fig.2.14, a gear train, a potentiometer, an integrated circuit, and a
shaft (output spine). The shaft can be positioned to specific angular positions by
sending the servo a coded signal. Of the three wires that stick out from the
servo casing, one is for power, one is for ground, and one is a control input line.
It uses the position-sensing device to determine the rotational position of the
shaft, so it knows which way the motor must turn to move the shaft to the
commanded position.
2. Stepper Motor: Stepper motors as shown in Fig.2.15 are DC motors that move
in discrete steps. They have multiple coils that are organized in groups called
“phases”. By energizing each phase in sequence, the motor will rotate, one step
at a time. With a computer controlled stepping, you can achieve very precise
positioning and/or speed control.
● The logistics and shipping industry can be use smart 5G technology for goods
tracking, fleet management, centralized database management, staff scheduling,
and real-time delivery tracking and reporting.
(i) Efficient use of RFID tags
(ii) Accelerated packing and labeling
(iii) Use of smart tracking devices for accurate monitoring of temperature, shock,
light exposure, humidity, etc…
(iv) Realtime GPS location tracking and reporting
(v) Efficient monitoring minimizes theft risk and misplacing of items
(vi) Realtime delivery tracking and reporting
(vii) Self-driving cars and drones for future goods delivery
● Smart city applications like traffic management, Instant weather update, local
area broadcasting, energy management, smart power grid, smart lighting of the
Maharashtra State Board of Technical Education P a g e 46 | 151
Emerging Trends in CO and IT (22618)
13) NGN compliant with all regulatory requirements, for example concerning
emergency communications, security, privacy, lawful interception, etc.
14) Reliability: To improve reliability, every communication device is highly reliable.
Provision of redundant configuration for communication circuits and equipment’s
is maintained
15) NGN is Layered Architecture having four layers:
a) Access Layer
b) Transport /Core Layer
c) Control Layer
d) Service Layer
e. NGN Architecture
Fig. 2.21 shows the functional block diagram of the NGN architecture. The NGN
functional architecture supports the UNI, NNI, ANI and SNI reference points. The NGN
architecture supports the delivery of multimedia services and content delivery services,
including video streaming and broadcasting. An aim of the NGN is to serve as aPSTN
and ISDN replacement. The NGN architecture defines a Network-Network Interface
(NNI), User-Network Interface (UNI), and an Application Network Interface (ANI).
The NGN functions are divided into:
● Service stratum (layer) functions
● Transport stratum (layer) functions.
To provide services, several functions in both the service stratum and the transport
stratum are needed. The delivery of services/applications to the end-user is provided by
utilizing the application support functions and service support functions, and related
control functions.
c) Wireless access (e.g., [b-IEEE 802.11] and [b-IEEE 802.16] technologies and
3G RAN access)
d) Optical access.
4) Gateway functions: The gateway functions provide capabilities to interwork with
end-user functions and/or other networks, including other types of NGN and many
existing networks, such as the PSTN/ISDN, the public Internet, and so forth.
Gateway functions can be controlled either directly from the service control
functions or through the transport control functions
5) Resource and admission control functions (RACF): Within the NGN
architecture, the resource and admission control functions (RACF) act as the
arbitrator between service control functions and transport functions for QoS. The
RACF provides an abstract view of transport network infrastructure to service
control functions (SCF) and makes service stratum functions agnostic to the details
of transport facilities, such as network topology, connectivity, resource utilization
and QoS mechanisms/technology.
6) Network attachment control functions (NACF): The network attachment control
functions (NACF) provide registration at the access level and initialization of end-
user functions for accessing NGN services. They also announce the contact point
of NGN functions in the service stratum to the end user.
The NACF provides the following functionalities:
a. Dynamic provisioning of IP addresses and other user equipment configuration
parameters.
b. By endorsement of the user, auto-discovery of user equipment capabilities and
other parameters.
c. Authentication of end user and network at the IP layer (and possibly other
layers). Regarding the authentication, mutual authentication between the end
user and the network attachment is performed.
d. Authorization of network access, based on user profiles.
e. Access network configuration, based on user profiles.
ii. Service stratum functions
Functional grouping in the service stratum includes:
a) The service control and content delivery functions including service user profile
functions.
b) The application support functions and service support functions.
1) Service control functions (SCF): The service control functions (SCF) include
resource control, registration and authentication- authorization functions at the
service level for both mediated and non-mediated services. They can also include
functions for controlling media resources, i.e., specialized resources and
gateways at the service-signaling level.
2) Content delivery functions (CDF): The content delivery functions (CDF)
receive content from the application support functions and service support
Maharashtra State Board of Technical Education P a g e 52 | 151
Emerging Trends in CO and IT (22618)
functions, store, process, and deliver it to the end-user functions using the
capabilities of the transport functions, under control of the service control
functions.
3) Application support functions and service support functions (ASF&SSF):
The application support functions and service support functions (ASF&SSF)
include functions such as the gateway, registration, authentication and
authorization functions at the application level. These functions are available to
the "applications" and "end-user" functional groups. Through the UNI, the
application support functions and service support functions provide reference
points to the end-user functions. Application interactions with the application
support functions and service support functions are handled through the ANI
reference point.
4) End-user functions: No assumptions are made about the diverse end-user
interfaces and end-user networks that may be connected to the NGN access
network. End-user equipment may be either mobile or fixed.
5) Management functions: These functions provide the capabilities to manage the
NGN in order to provide NGN services with the expected quality, security, and
reliability. All the NGN components are centralized controlled.
or services that are not directly hosted on the MGC. Typically, the MGC routes
calls to the appropriated AS for features the MGC does not support.
ii. Unlicensed radio bands: Unlicensed radio bands have been allocated to certain
users by the government or any individual can use it, but to be able to use and
broadcast on these bands, you do not need to have a license; you only need to
create compliant devices that are to be used. Regulations exist around these
bands. Some of the types of unlicensed radio bands are given in Table 2.2:
Standard Bodies: IEEE 802.11 networks have several choices of wireless bands that
are available to them to use, without the requirement to lease the frequencies from the
government. Following groups and standards bodies have helped to develop standards
so that all users can be good neighbors with others who use those radio bands.
Maharashtra State Board of Technical Education P a g e 55 | 151
Emerging Trends in CO and IT (22618)
4. The fourth generation (4G) integrates 3G with fixed internet to support wireless
mobile internet, which is an evolution to mobile technology and it overcome the
limitations of 3G. It also increases the bandwidth and reduces the cost of resources.
5. 5G stands for 5th Generation Mobile technology and is going to be a new revolution
in mobile market which has changed the means to use cell phones within very high
bandwidth. User never experienced ever before such high value technology which
includes all type of advance features and 5G technology will be most powerful and
in huge demand in near future.
i. 1G – First generation mobile communication system
The first generation of mobile network was deployed in Japan by Nippon Telephone
and Telegraph company (NTT) in Tokyo during 1979. In the beginning of 1980s, it
gained popularity in the US, Finland, UK and Europe. This system used analogue
signals and it had many disadvantages due to technology limitations.
Disadvantages of 1G system
Disadvantages of 2G system
● Low data rate
● Limited mobility
● Less features on mobile devices
● Limited number of users and hardware capability
iii. 3G – Third generation communication system
Third generation mobile communication started with the introduction of UMTS –
Universal Mobile Terrestrial / Telecommunication Systems. UMTS has the data rate of
384kbps and it support video calling for the first time on mobile devices.
After the introduction of 3G mobile communication system, smart phones became
popular across the globe. Specific applications were developed for smartphones which
handles multimedia chat, email, video calling, games, social media and healthcare.
Disadvantages of 3G systems
● Expensive spectrum licenses
● Costly infrastructure, equipment and implementation
● Higher bandwidth requirements to support higher data rate
● Costly mobile devices
Disadvantages of 4G system
● Expensive hardware and infrastructure
● Costly spectrum in most countries, frequency bands are too expensive.
● High end mobile devices compatible with 4G technology required, which is
costly
● Wide deployment and upgrade is time consuming
Faster
Better coverage and no
Secure, broadband
Key Mobility Better Internet Dropped calls ,much
Mass Internet
differentiator experience lower latency, Better
adoption , lower
performance
Latency
NGN Core
MPLS (Multi-protocol label switching) is used at core transport layer in NGN network.
MPLS provides faster switching, propagation delay is less.
a. MPLS Concept
In NGN, a packet of a connectionless network layer protocol travels from one router to the
next, each router makes an independent forwarding decision for that packet. That is, each
router analyzes the packets header, and each router runs a network layer routing algorithm.
Each router independently chooses a next hop for the packet, based on its analysis of the
packet's header and the results of running the routing algorithm.
Packet headers contain considerably more information than is needed simply to choose the next
hop. Choosing the next hop can therefore be thought of as the composition of two functions.
i. The first function partitions the entire set of possible packets into a set of Forwarding
Equivalence Classes (FECs).
ii. The second maps each FEC to a next hop.
b. MPLS Basics
1) Labels: A label is a short, fixed length, locally significant identifier which is used to identify
a FEC. The label which is put on a particular packet represents the Forwarding Equivalence
Class to which that packet is assigned. Most commonly, a packet is assigned to a FEC based
on its network layer destination address. Each IP destination network has a different label
which has local significance. Label for a destination network changes in each hop.
2) Label Switch Router: A label switch router (LSR) is a router that supports MPLS. It is
capable of understanding MPLS labels and of receiving and transmitting a labeled packet
on a data link. Three kinds of LSRs exist in an MPLS network:
LSP is the path through the MPLS network or a part of it that packets take. The first LSR
of an LSP is the ingress LSR for that LSP, whereas the last LSR of the LSP is the egress
LSR. All the LSRs in between the ingress and egress LSRs are the intermediate LSRs.
LSP is unidirectional. The flow of labeled packets in the other direction right to left
between the same edge LSRs would be another LSP.
MPLS Features:
1 QoS enabled MPLS transport network will provide real time and data transport
application.
2 MPLS increases operator revenue.
3 MPLS offers high transport efficiency by using hybrid technology (packetswitching
and circuit switching)
4 MPLS offers high reliability and operational simplicity.
5 MPLS provide control and deterministic usage of network resources, end-to-end control
to engineer network paths and to efficiently utilize network resources.
6 In MPLS network management is simple
7 It ensures smooth interworking of the packet transport network with other
existing/legacy packet networks.
8 MPLS header length is of 32 bits, in which label length is 20 bits, 3 bits for service
quality and 8bits for Timeto live (no of hops in network) and 1 bit for stacking the
labels.
MPLS Advantages
1. Consistent Network Performance: The MPLS allows different Class of Service
classifications to be applied to packets, to help ensure that data loss (packet loss),
transmission delays (latency) and variations in transmission delay (jitter) are kept
without appropriate limits.
2. Obscures Network Complexity: MPLS can effectively hide the underlying
complexity of the network from devices and users that don't need to know about it.
3. Reduced Network Congestion: The MPLS can support Traffic Engineering having
many uses, including the re-routing of comparatively delay-tolerant traffic to slower,
circuitous, under-utilised routes, freeing up capacity on quicker (lower-latency)
overcrowded paths.
4. Increased Uptime: MPLS is that it has the potential to increase uptime as Fast Reroute
that enables traffic to be switched to an alternative path very rapidly
5. Scalability: According to the needs, the MPLS can be scaled up and down
6. Efficiency: MPLS offers much higher quality connections without packet loss and jitter
using it along with VoIP may lead to increased efficiency
7. Reliability: Since MPLS uses labels for forwarding packets, it can be assured that the
packets will be delivered to the right destination.
References:
● https://data-flair.training/blogs/iot-applications/
● https://kainjan1.files.wordpress.com/2018/01/chapter-1_iot.pdf
● https://www.tutorialspoint.com/internet_of_things/internet_of_things_tutorial.p
df
● https://www.iare.ac.in/sites/default/files/lecture_notes/IOT%20LECTURE%20
NOTES_IT.pdf
● https://www2.deloitte.com/content/dam/insights/us/articles/iot-primer-iot-
technologies-applications/DUP_1102_InsideTheInternetOfThings.pdf
● https://techdifferences.com/difference-between-sensors-and-actuators.html
● https://electronicsforu.com/technology-trends/tech-focus/iot-sensors
● https://iotbytes.wordpress.com/basic-iot-actuators/
● https://www.rfpage.com/evolution-of-wireless-technologies-1g-to-5g-in-
mobile-communication /#:~:text=In%20order%20to%20support%20higher
,to%20473.6kbps%20(maximum).
● https://www.itu.int/rec/dologin_pub.asp?lang=e&id=T-RECY.2012-200609-
S!!PDF-E&type=items
● https://pdfs.semanticscholar.org/4dfd/40cc3a386573ee861c5329ab4
c6711210819.pdf
● www.tec.gov.in
● www.trai.gov.in
Sample Multiple Choice Questions:
1. IoT stands for
a. Internet of Technology
b. Intranet of Things
c. Internet of Things
d. Information of Things
2. Which is not the Feature of IoT
a. Connectivity
b. Self-Configuring
c. Endpoint Management
d. Artificial Intelligence
3. Devices that transforms electrical signals into physical movements.
a. Sensors
b. Actuators
c. Switches
d. display
4. IPv4 uses a ______address scheme.
a. 8-bit
b. 16-bit
c. 32-bit
d. 64 bit
5. Data Distribution Service (DDS) is a _______standard for device-to-device or
machine-to-machine communication
a. data-centric middleware
b. data-centric hardware
c. data-centric Software
d. None of above
Content
3.1 Introduction to Blockchain
● Backstory of Blockchain
● What is Blockchain?
3.2 Centralize versus Decentralized System
3.3 Layers of Blockchain
● Application Layer
● Execution Layer
● Semantic Layer
● Propagation Layer
● Consensus Layer
3.4 Importance of Blockchain
● Limitations of Centralized Systems
● Blockchain Adoption So Far
3.5 Blockchain Use and Use Cases
What is Blockchain?
Blockchain is a common, unchallengeable digital ledger that allows the process
of recording transactions and tracking assets in a business network. Where an asset may
be tangible such as a house, car, cash, land or intangible such as intellectual property,
patents, copyrights, branding. Virtually everything of value can be tracked and traded
on a blockchain network, decreasing risk and cutting costs for all involved.
Business runs on information can says better where faster it’s received and the
more accurate it is. Blockchain is perfect which provides immediate, shared and
completely transparent information stored on an immutable digital ledger that can be
accessed only by permissioned network members. A blockchain network can track
orders, payments, accounts, production etc. Hence members share a single view of the
fact, other can see all details of a transaction end to end, giving greater sureness, as well
as new efficiencies and opportunities.
● Blockchain is a peer-to-peer system of transacting values with no trusted third parties
in between.
● It is a shared, decentralized, and open ledger of transactions. This ledger database is
replicated across a large number of nodes.
● This ledger database is an append-only database and cannot be changed or altered. It
means that every entry is a permanent entry. Any new entry on it gets reflected on all
copies of the databases hosted on different nodes.
● There is no need for trusted third parties to serve as intermediaries to verify, secure,
and settle the
● transactions.
● It is another layer on top of the Internet and can coexist with other Internet
technologies.
● Just the way TCP/IP was designed to achieve an open system, blockchain technology
was designed to enable true decentralization. In an effort to do so, the creators of
Bitcoin open-sourced it so it could inspire many decentralized applications.
A typical blockchain may look as shown in Figure 3.1.
The genesis block is the first block in any blockchain referred as Block 0 and
there is no previous block for reference. The Genesis Block has previous hash value set
to 0 to indicate no data was processed before the Genesis Block. The hash of genesis
Maharashtra State Board of Technical Education P a g e 67 | 151
Emerging Trends in CO and IT (22618)
block is added to all new transactions in a new block and used to create its unique hash.
This process is repeated until all the new blocks are added to a blockchain.
Every node on the blockchain network has an identical copy of the blockchain
shown in Figure 3.1, where every block is a list of transactions. There are two major
parts in every block. The “header” part pointer links back to the previous block in the
chain and every block header contains the hash of the previous block so that no one can
modify any transaction in the previous block. The other part of a block contents the
validated list of transactions, their amounts, timestamp, the addresses of the parties
involved, and some another details.
A centralized system has a centralized control with all administrative rights and are easy
to design, maintain, enforce trust, and administrate, but suffer from many intrinsic
limitations, as given below:
1. A centralized system has a central point of failure; hence they are less stable.
2. A centralized system is more vulnerable to attack and hence less secured.
3. Centralization of power may lead to unethical operations.
4. Most of the time, scalability is difficult.
Decentralized System
A decentralized system does not have a centralized control and every node has equal
authority. Such systems are difficult to design, maintain, govern, or impose trust. But,
decentralized system do not suffer from the limitations of conventional centralized
systems. Decentralized systems offer the following advantages:
1. A decentralized system is more stable and fault tolerant as they do not have a
central point of failure.
2. Attack resistant, as no central point to easily attack and hence more secured
3. Symmetric system with equal right to all nodes, so less scope of unethical
operations and usually independent in nature
A distributed system may also be decentralized system, for example, blockchain. But,
the task is not subdivided and delegated to nodes unlike common distributed systems,
because there is no master that in blockchain. A typical decentralized and distributed
system is efficiently a peer-to-peer system as shown in Figure 3.3.
A propagation layer is used in the peer-to-peer communications between the nodes that
allow them to discover each other and get synchronized with another node in a network.
When a transaction is done, then gets broadcasted to all other nodes in the network.
Also, when a node advises a block, it is immediately get broadcast in the entire network
so that other nodes can use this newly created block and work on it. Hence, the
propagation of the block or a transaction in the network is defined in this layer and
confirms the stability of the complete network. In the asynchronous Internet network,
there are often latency issues for transaction or block propagation. However, depending
upon the network capacity or bandwidth, the propagation may occur instantly or it may
take a longer.
5. Consensus Layer
Consensus layer is the first layer for most of the blockchain systems and main purpose
is to make sure that all the nodes must get approve on a common state of the shared
ledger. Consensus layer also deals with the safety and security of the blockchain. There
are many consensus algorithms which can be applied for generation of cryptocurrencies
like Bitcoin and Ethereum, they use proof-of-work mechanism to select a node
randomly out of various nodes present on the network that can propose a new block.
Once a new block is created, the block is propagated to all the other nodes to check if
the new block is valid or not with the transactions in it and based on the consensus from
all other nodes the new block gets added on to the blockchain.
Content
4.1 Digital forensics
● Introduction to digital forensic
● Digital forensics investigation process
● Models of Digital Forensic Investigation –
o Abstract Digital Forensics Model (ADFM)
o Integrated Digital Investigation Process (IDIP)
o An extended model for cybercrime investigation
4.2 Ethical issues in digital forensic
● General ethical norms for investigators
● Unethical norms for investigation
4.3 Digital Evidences
● Definition of Digital Evidence
● Best Evidence Rule
● Original Evidence
4.4 Characteristics of Digital Evidence
● Locard’s Exchange Principle
● Digital Stream of bits
4.5 Types of Evidence : Illustrative, Electronics, Documented, Explainable, Substantial,
Testimonial
4.6 Challenges in evidence handling
o Authentication of evidence
o Chain of custody
o Evidence validation
4.7 Volatile evidence
Identification
Preparation
Approach Strategy
Preservation
Collection
Examination
Analysis
Presentation
Returning Evidence
Physical
Readiness Deployment Crime Review
Investigation
Digital
Crime
Investigatio
● Detection and Notification phase; where the incident is detected and then
appropriate people notified.
● Confirmation and Authorization phase; which confirms the incident and
obtains authorization for legal approval to carry out a search warrant.
3. Physical Crime Investigation phase The goal of these phases is to collect and
analyze the physical evidence and reconstruct the actions that took place during
the incident.
It includes six phases:
● Preservation phase; which seeks to preserve the crime scene so that evidence can
be later identified and collected by personnel trained in digital evidence
identification.
● Survey phase; that requires an investigator to walk through the physical crime
scene and identify pieces of physical evidence.
● Documentation phase; which involves taking photographs, sketches, and videos
of the crime scene and the physical evidence. The goal is to capture as much
information as possible so that the layout and important details of the crime scene
are preserved and recorded.
● Search and collection phase; that entails an in-depth search and collection of the
scene is performed so that additional physical evidence is identified and hence
paving way for a digital crime investigation to begin
● Reconstruction phase; which involves organizing the results from the analysis
done and using them to develop a theory for the incident.
● Presentation phase; that presents the physical and digital evidence to a court or
corporate management.
4. Digital Crime Investigation phaseThe goal is to collect and analyze the digital
evidence that was obtained from the physical investigation phase and through
any other future means. It includes similar phases as the Physical Investigation
phases, although the primary focus is on the digital evidence. The six phases are:
● Preservation phase; which preserves the digital crime scene so that evidence
can later be synchronized and analyzed for further evidence.
● Survey phase; whereby the investigator transfers the relevant data from a
venue out of physical or administrative control of the investigator to a
controlled location.
● Documentation phase; which involves properly documenting the digital
evidence when it is found. This information is helpful in the presentation
phase.
● Search and collection phase; whereby an in-depth analysis of the digital
evidence is performed. Software tools are used to reveal hidden, deleted,
swapped and corrupted files that were used including the dates, duration, log
file etc. Low-level time lining is performed to trace a user’s activities and
identity.
● Reconstruction phase; which includes putting the pieces of a digital puzzle
together, and developing investigative hypotheses.
● Presentation phase; that involves presenting the digital evidence that was
found to the physical investigative team.
It is noteworthy that this DFPM facilitates concurrent execution of physical and
digital investigation.
5. Review phase this entails a review of the whole investigation and identifies areas
of improvement. The IDIP model does well at illustrating the forensic process,
and also conforms to the cyber terrorism capabilities which require a digital
investigation to address issues of data protection, data acquisition, imaging,
extraction, interrogation, ingestion/normalization, analysis and reporting. It also
highlights the reconstruction of the events that led to the incident and emphasizes
reviewing the whole task, hence ultimately building a mechanism for quicker
forensic examinations.
III. An Extended Model of Cybercrime Investigation (EMCI)
Storage of
Awareness Evidence
Authorization Examination
of evidence
Planning Hypothesis
Notification
Presentation
of hypothesis
Search for
identify evidence Proof/
Defense of
hypothesis
Collection
of
evidence
Dissemination
of information
Transport
of Evidence
There are many sources of digital evidence; the topic is divided into three major forensic
categories of devices where evidence can be found: Internet-based, stand-alone
computers or devices, and mobile devices. These areas tend to have different evidence-
gathering processes, tools and concerns, and different types of crimes tend to lend
themselves to one device or the other.
Some of the popular electronic devices which are potential digital evidence are: HDD,
CD/DVD media, backup tapes, USB drive, biometric scanner, digital camera, smart
phone, smart card, PDA, etc.
4.3.2 Forms of digital evidence: Text message, emails, pictures, videos and internet
searches are the most common types of Digital evidences.
The digital evidence is used to establish a credible link between the attacker,
victim, and the crime scene. Some of the information stored in the victim’s system can
be potential digital evidence, such as IP address, system log-in & remote log-in details,
browsing history, log files, emails, images, etc.
● Web History
● Cache files
● Cookies
● Registry
● Unallocated Space
● Slack Space
● Web/E-Mail server access Logs
● Domain access Logs
store either the best evidence or original evidence for every investigation in the evidence
safe.
4.4 Characteristics of Digital Evidence:
Characteristics of digital evidences can help and challenge investigators during an
investigation.The main goals in any investigation are to follow the trails that offenders
leave during the commission of a crime and to tie perpetrators to the victims and crime
scenes. Although witnesses may identify a suspect, tangible evidence of an individual’s
involvement is usually more compelling and reliable. Forensic analysts are employed
to uncover compelling links between the offender, victim, and crime scene.
Figure 4.4:
Evidence transfer in the physical and digital dimensions helps investigators
establish connections between victims, offenders, and crime scenes.
In computer intrusions, the attackers will leave multiple traces of their presence
throughout the environment, including in the fi le systems, registry, system logs, and
network-level logs. Furthermore, the attackers could transfer elements of the crime
scene back with them, such as stolen user passwords or PII in a file or database. Such
evidence can be useful to link an individual to an intrusion.
In an e-mail harassment case, the act of sending threatening messages via a Web-
based e-mail service such as Hotmail can leave a number of traces. The Web browser
used to send messages will store fi les, links, and other information on the sender’s hard
drive along with date-time–related information. Therefore, forensic analysts may find
an abundance of information relating to the sent message on the offender’s hard drive,
including the original message contents. Additionally, investigators may be able to
obtain related information from Hotmail, including Web server access logs, IP
addresses, and possibly the entire message in the sent mail folder of the offender’s e-
mail account.
Suppose that, as the examiner, you obtain metadata for a piece of evidence. However,
you are unable to extract meaningful information from it. The fact that there is no
meaningful information within the metadata does not mean that the evidence is
insufficient. The chain of custody in this case helps show where the possible evidence
might lie, where it came from, who created it, and the type of equipment that was
used. That way, if you want to create an exemplar, you can get that equipment, create
the exemplar, and compare it to the evidence to confirm the evidence properties.
It is possible to have the evidence presented in court dismissed if there is a missing link
Maharashtra State Board of Technical Education P a g e 88 | 151
Emerging Trends in CO and IT (22618)
If there is a break in the chain of custody brought to the attention of the court,
then the court has to decide whether the breach is so severe as to meet exclusion of the
item from trial. Alternatively, the court can decide that the Trier (trial judge or jury)
need to decide the value of the evidence. To prevent a breach, a forensic investigation
should follow a written policy, so that necessary deviations of the policy can be argued.
The policy itself should take all reasonable (or arguably reasonable) precautions against
tampering.
For example, assume that a PDA is seized from a suspected drug dealer. In the
case of an PDA, there is no hard drive image to mirror, that is, the examination will
have to be done on the powered-on original. The PDA can lose data, for example by
disconnecting it from its battery. On seizure, the device should not be switched on. If it
is seized switched on, it should be switched off in order to preserve battery power. It
needs to be put into an evidence bag that does not allow access to the PDA without
breaking the seal (no clear plastic bag!). The evidence needs to be tagged with all
pertinent data, including the serial number of the PDA and the circumstances of the
seizure. The PDA should never be returned to the accused at the scene, because the
device can lose data if reset. To maintain the data in the PDA, it needs to be kept in a
continuously charged mode. It should only be used to extract evidence by a competent
person who can testify in court. As long as the PDA could be evidence, it needs to be
kept in an evidence locker, with check-out logs, so that it can be determined who had
access to the PDA at any time.
4.6.3 Evidence Validation: The challenge is to ensure that providing or obtaining the
data that you have collected is similar to the data provided or presented in court. Several
years pass between the collection of evidence and the production of evidence at a
judiciary proceeding, which is very common. To meet the challenge of validation, it is
necessary to ensure that the original media matches the forensic duplication by using
MD5 hashes. The evidence for every file is nothing but the MD5 hash values that are
generated for every file that contributes to the case.
The verify function within the Encase application can be used while duplicating
a hard drive with Encase. To perform a forensic duplication using dd , you must record
MD5 hash for both the original evidence media and binary files or the files which
compose the forensic duplication.
Note: Evidence collection calculated by MD5 after 6 months may not be helpful.MD5
hashes should be performed when the evidence is obtained.
4.7 Volatile Evidence: Not all the evidence on a system is going to last very long. Some
evidence is residing in storage that requires a consistent power supply; other evidence
may be stored in information that is continuously changing. When collecting evidence,
you should always try to proceed from the most volatile to the least. Of course, you
should still take the individual circumstances into account—you shouldn’t waste time
extracting information from an unimportant/unaffected machine’s main memory when
an important or affected machine’s secondary memory hasn’t been examined.
You need to respond to the target system at the console during the collection of
volatile data rather than access it over the network. This way the possibility of the
attacker monitoring your responses is eliminated, ensuring that you are running trust
commands. If you are creating a forensic duplication of the targeted system, you should
focus on obtaining the volatile system data before shutting down the system.
To determine what evidence to collect first, you should draw up an Order of
Volatility—a list of evidence sources ordered by relative volatility. An example an
Order of Volatility would be:
Note: Once you have collected the raw data from volatile sources you may be able to
shutdown the system.{Matthew Braid, “Collecting Electronic Evidence After A System
Compromise,” Australian Computer Emergency Response Team}
Registers, Cache: The contents of CPU cache and registers are extremely volatile,
since they are changing all of the time. Literally, nanoseconds make the difference here.
An examiner needs to get to the cache and register immediately and extract that
evidence before it is lost.
Routing Table, ARP Cache, Process Table, Kernel Statistics, Memory: Some of
these items, like the routing table and the process table, have data located on network
devices. In other words, that data can change quickly while the system is in operation,
Maharashtra State Board of Technical Education P a g e 92 | 151
Emerging Trends in CO and IT (22618)
so evidence must be gathered quickly. Also, kernel statistics are moving back and forth
between cache and main memory, which make them highly volatile. Finally, the
information located on random access memory (RAM) can be lost if there is a power
spike or if power goes out. Clearly, that information must be obtained quickly.
Temporary File Systems: Even though the contents of temporary file systems have the
potential to become an important part of future legal proceedings, the volatility concern
is not as high here. Temporary file systems usually stick around for awhile.
Disk: Even though we think that the data we place on a disk will be around forever, that
is not always the case (see the SSD Forensic Analysis post from June 21). However, the
likelihood that data on a disk cannot be extracted is very low.
Remote Logging and Monitoring Data that is Relevant to the System in Question:
The potential for remote logging and monitoring data to change is much higher than
data on a hard drive, but the information is not as vital. So, even though the volatility of
the data is higher here, we still want that hard drive data first.
Case Studies :
Background:
The assistant manager (the complainant) with the fraud control unit of a large business
process outsourcing (BPO) organization filed a complaint alleging that two of its
employees had conspired with a credit card holder to manipulate the credit limit and as
a result cheated the company of INR 0.72 million.
The BPO facility had about 350 employees. Their primary function was to issue the
Maharashtra State Board of Technical Education P a g e 93 | 151
Emerging Trends in CO and IT (22618)
bank's credit cards as well as attend to customer and merchant queries. Each employee
was assigned to a specific task and was only allowed to access the computer system for
that specific task. The employees were not allowed to make any changes in the credit-
card holder's account unless they received specific approvals.
Each of the employees was given a unique individual password. In case they entered an
incorrect password three consecutive times then their password would get blocked and
they would be issued a temporary password.
The company suspected that its employees conspired with the son (holding an add-on
card) of one of the credit card holders. The modus operandi suspected by the client is
as follows.
The BPO employee deliberately keyed in the wrong password three consecutive times
(so that his password would get blocked) and obtained a temporary password to access
the computer system. He manually reversed the transactions of the card so that it
appeared that payment for the transaction has taken place. The suspect also changed the
credit card holder's address so that the statement of account would never be delivered
to the primary card holder.
The team analysed the attendance register which showed that the accused was present
at all the times when the fraudulent entries had been entered in the system. They also
analysed the system logs that showed that the accuser's ID had been used to make the
changes in the system.
The team also visited the merchant establishments from where some of the transactions
had taken place. The owners of these establishments identified the holder of the add-on
card.
Current status:The BPO was informed of the security lapse in the software utilised.
Armed with this evidence the investigating team arrested all the accused and recovered,
on their confession, six mobile phones, costly imported wrist watches, jewels, electronic
items, leather accessories, credit cards, all worth INR 0. 3 million and cash INR 25000.
The investigating team informed the company of the security lapses in their software so
that instances like this could be avoided in the future.
This case won the second runner-up position for the India Cyber Cop Award, for its
investigating officer Mr S. Balu, Assistant Commissioner of Police, Crime, Chennai
Police. The case was remarkable for the excellent understanding displayed by the
investigating team, of the business processes and its use in collecting digital evidence.
Using the same log-in details, the investigating team accessed the Web pages where
these profiles were uploaded. The message had been posted on five groups, one of which
was a public group. The investigating team obtained the access logs of the public group
and the message to identify the IP addresses used to post the message. Two IP addresses
were identified.
The ISP was identified with the help of publicly available Internet sites. A request was
made to the ISPs to provide the details of the computer with the IP addresses at the time
the messages were posted. They provided the names and addresses of two cyber cafes
located in Mumbai to the police.
The investigating team scrutinised the registers maintained by the cyber cafes and found
that in one case the complainant's name had been signed into the register.
The team also cross-examined the complainant in great detail. During one of the
meetings she revealed that she had refused a former college mate who had proposed
marriage.
In view of the above the former college mate became the prime suspect. Using this
information the investigating team, with the help of Mumbai police, arrested the suspect
and seized a mobile phone from him. After the forensic examination of the SIM card
and the phone, it was observed that phone had the complainant’s telephone number that
was posted on the internet. The owner of the cyber cafes also identified the suspect as
the one who had visited the cyber cafes.
Based on the facts available with the police and the sustained interrogation the suspect
confessed to the crime.
Current status:The suspect was convicted of the crime and sentenced to two years of
imprisonment as well as a fine.
The investigators were successful in arresting two people as they laid a trap in a local
bank where the accused had fake accounts for illegally transferring money.
During the investigation the system server logs of the BPO were collected. The IP
addresses were traced to the Internet service provider and ultimately to the cyber cafes
through which illegal transfers were made.
The registers maintained in cyber cafes and the owners of cyber cafes assisted in
identifying the other accused in the case. The e-mail IDs and phone call print outs were
also procured and studied to establish the identity of the accused. The e-mail accounts
of the arrested accused were scanned which revealed vital information to identify the
other accused. Some e-mail accounts of the accused contained swift codes, which were
required for internet money transfer.
All the 17 accused in the case were arrested in a short span of time. The charge sheet
was submitted in the court within the stipulated time. In the entire wire transfer scam,
an amount to the tune of about INR 19 million was transferred, out of this INR 9 million
was blocked in transit due to timely intimation by police, INR 2 million was held in
balance in one of the bank accounts opened by the accused which was frozen. In
addition the police recovered cash, ornaments, vehicles and other articles amounting to
INR 3 million.
During the investigation the investigating officer learned the process of wire transfer,
the banking procedures and weakness in the system. The investigating officer suggested
measures to rectify the weakness in the present security systems of the call centre. This
has helped the local BPO industry in taking appropriate security measures.
Background: The accused in this case was posing to be a genuine railway ticket agent
and had been purchasing tickets online by using stolen credit cards of non residents.
The accused created fraudulent electronic records/ profiles, which he used to carry out
the transactions.The tickets so purchased were sold for cash to other passengers. Such
events occurred for a period of about four months.
The online ticket booking service provider took notice of this and lodged a complaint
with the cyber crime investigation cell.
Investigation: A procedure to find the Digital Evidence
The service provider gave the IP addresses, which were used for the fraudulent online
bookings, to the investigating team. IP addresses were traced to cyber cafes in two
locations.
The investigating team visited the cyber cafŽs but was not able to get the desired logs
as they were not maintained by the cyber cafŽ owners. The investigating team was able
to short list the persons present at cyber cafes when the bookings were made. The
respective owners of the cyber cafes were able to identify two persons who would
regularly book railway tickets.
The investigating team then examined the passengers who had travelled on these tickets.
They stated that they had received the tickets from the accused and identified the
delivery boy who delivered the tickets to them. On the basis of this evidence the
investigating team arrested two persons who were identified in an identification parade.
The investigating officer took the original e-mail from the complainant and extracted
the IP address of the same. From the IP address he could ascertain the Internet service
provider.
The IP address was traced to a cable Internet service provider in the city area of
Hyderabad. The said IP address was allotted to the former husband sometime back and
his house was traced with the help of the staff of ISP.
A search warrant was obtained and the house of the accused was searched. During the
search operation, a desktop computer and a handicam were seized from the premises. A
forensic IT specialist assisted the investigation officer in recovering e-mails (which
Maharashtra State Board of Technical Education P a g e 98 | 151
Emerging Trends in CO and IT (22618)
were sent to the complainant), using a specialised disk search tool as well as
photographs (which had been posted on the Internet) from the computer and the
handicam respectively. The seized computer and the handicam were sent to the forensic
security laboratory for further analysis.
The experts of the forensic security laboratory analysed the material and issued a report
stating that: the hard disk of the seized computer contained text that was identical to
that of the obscene e-mail; the computer had been used to access the matrimonial
websites on which the obscene profiles were posted; the computer had been used to
access the e-mail account that was used to send the obscene e-mail; the handicam seized
from the accused contained images identical to the ones posted on the matrimonial
Websites. Based on the report of the FSL it was clearly established that the accused had:
created a fictitious e-mail ID and had sent the obscene e-mail to the complainant; posted
the profiles of the victim along with her photographs on the matrimonial sites.
Current status:Based on the material and oral evidence, a charge sheet has been filed
against the accused and the case is currently pending for trial.
References
1. http://www.forensicsciencesimplified.org/digital/
2. http://www.forensicsciencesimplified.org/digital/
3. https://www.helpnetsecurity.com/2007/07/20/the-rules-for-computer-
forensics/ as on 28 August 2019
4. Digital Evidence and Computer Crime, Third Edition © 2011 Eoghan Casey.
Published by Elsevier Inc.
5. www.cse.scu.edu/~tschwarz/COEN252_13/LN/legalissues.html
6. The evidences or proof that can be obtained from the electronic source is called
the_______
a. digital evidence
b. demonstrative evidence
c. Explainable Evidence
d. substantial evidence
7. Which of the following is not a type of volatile evidence?
a. Routing Tables
b. Main Memory
c. Log files
d. Cached Data
Contents
5.1 Ethical Hacking
● How Hackers Beget Ethical Hackers
● Defining hacker, Malicious users
● Data Privacy and General Data Protection and Regulation(GDPR)
5.2 Understanding the need to hack your own systems
5.3 Understanding the dangers your systems face
● Nontechnical attacks
● Network-infrastructure attacks
● Operating-system attacks
● Application and other specialized attacks
5.4 Obeying the Ethical hacking Principles
● Working ethically
● Respecting privacy
● Not crashing your systems
5.5 The Ethical hacking Process
● Formulating your plan
● Selecting tools
● Executing the plan
● Evaluating results
● Moving on
5.6 Cyber Security act
● Gather Information: This is the first stage, the learns as much as he can about
the intended victim. The information is gathered from company websites, other
publications and sometimes by talking to the users of the target system.
● Plan Attack: The attackers outline how he/she intends to execute the attack
● Acquire Tools: These include computer programs that an attacker will use when
launching the attack.
● Attack: Exploit the weaknesses in the target system.
● Use acquired knowledge: Information gathered during the social engineering
tactics such as pet names, birthdates of the organization founders, etc. is used in
attacks such as password guessing.
Most techniques employed by social engineers involve manipulating human biases.
To counter such techniques, an organization can;
✔ To counter the familiarity exploit
✔ To counter intimidating circumstances attacks
✔ To counter phishing techniques
✔ To counter tailgating attacks
✔ To counter human curiosity
✔ To counter techniques that exploit human greed
Summary
● Social engineering is the art of exploiting the human elements to gain access to
un-authorized resources.
● Social engineers use a number of techniques to fool the users into revealing
sensitive information.
● Organizations must have security policies that have social engineering
countermeasures.
Hacker’s attitude:
A hacker-cracker separation gives more emphasis to a range of different categories,
such as white hat (ethical hacking), grey hat, black hat and script kiddies. The term
cracker refers to black hat hackers, or more generally hackers with unlawful
intentions.
Hackers are problem solvers. They get extract from understanding a problem and
sorting out a solution. Their motivation to meet challenges is internal. Hackers do what
Maharashtra State Board of Technical Education P a g e 102 | 151
Emerging Trends in CO and IT (22618)
they do because it’s extremely satisfying to solve puzzles and fix the up-until-now
unfixable. The pleasure derived is both intellectual and practical. But one doesn’t have
to be a geek to be a hacker. Being a hacker is a mind-set. In Raymond’s
dissertation, “How to Become a Hacker”, he describes the fundamentals of a hacker
attitude.
These are very same principles apply to being innovative which are explained as below:
The world is full of fascinating problems waiting to be solved.
Innovation happens because hackers like to solve the problem rather than complaining.
If one happen to find these problems fascinating and exciting, then it won’t even feel
like hard work.
No Problem should ever have to be solved twice.
Hackers are perfectionists for clarifying the problem before they start generating ideas.
It’s easy to jump to solutions, but sometimes that means wrong problems are solved. A
little bit of accuracy on the front end of a problem solving process means one tackles
the right and real problem, so one only have to do it once.
Boredom and drudgery(more and more work) are evil.
The best way to lose touch with innovation is to become too repetitive. Innovation
requires constant and vigilant creativity. It may not be broken enough to fix, but there’s
no reason not to squeeze it and cut boredom off at the pass.
Freedom is good.
Hackers need freedom to work upon their ideas.
Attitude is no substitute for competence.
They are open-minded and they see problems as interesting opportunities. Innovators
are seeking to understand a problem more deeply, puzzling at how an unworkable idea
might become workable, increasing their skill set so that they are better problem solvers
and can better execute their ideas.
Hackers are the innovators of the Internet. Overall hackers are who have got that
relentless, curious, problem-solving attitude.
Computer Hacking:
Computer Hackers have been in existence for more than a century. Originally,
"hacker" did not carry the negative implications. In the late 1950s and early 1960s,
computers were much different than the desktop or laptop systems most people are
familiar with. In those days, most companies and universities used mainframe
computers: giant, slow-moving hunks of metal locked away in temperature-controlled
glass cages. It cost thousands of dollars to maintain and operate those machines, and
programmers had to fight for access time.
Because of the time and money involved, computer programmers began looking for
ways to get the most out of the machines. The best and brightest of those programmers
created what they called "hacks" - shortcuts that would modify and improve the
performance of a computer's operating system or applications and allow more tasks
to be completed in a shorter time.
Maharashtra State Board of Technical Education P a g e 103 | 151
Emerging Trends in CO and IT (22618)
Still, for all the negative things hackers have done, they provide a necessary (and even
valuable) service, which is elaborated on after a brief timeline in the history of
computer hacking
● Policy considerations
If it is chosen to make ethical hacking an important part of business’s information risk
management program, one really need to have a documented security testing policy.
Such a policy outlines who’s doing the testing, the general type of testing that is
performed, and how often the testing takes place.
● What is Hacking?
Hacking is identifying weakness in computer systems or networks to exploit its
weaknesses to gain access.
● Example of Hacking:
Computers have become mandatory to run successful businesses. It is not enough to
have isolated computers systems; they need to be networked to facilitate
communication with external businesses.
✔ Using password cracking algorithm to gain access to a system.
✔ This exposes them to the outside world and hacking. Hacking means using
computers to commit fraudulent acts such as fraud, privacy invasion, stealing
corporate/personal data, etc.
✔ Cybercrimes cost many organizations millions of dollars every year.
Businesses need to protect themselves against such attacks.
Definition:
Ethical hacking
✔ Refers to the act of locating weaknesses and vulnerabilities of computer and
information systems by duplicating the intent and actions of malicious hackers.
✔ Known as penetration testing, intrusion testing or red teaming.
An ethical hacker is a security professional who applies their hacking skills for
defensive purposes on behalf of the owners of information systems.
By conducting penetration tests, an ethical hacker looks to answer the following four
basic questions:
1. What information/locations/systems can an attacker gain access?
2. What can an attacker see on the target?
3. What can an attacker do with available information?
4. Does anyone at the target system notice the attempts?
An ethical hacker operates with the knowledge and permission of the organization for
which they are trying to defend. In some cases, the organization will neglect to inform
their information security team of the activities that will be carried out by an ethical
hacker in an attempt to test the effectiveness of the information security team. This is
referred to as a double-blind environment. In order to operate effectively and legally,
an ethical hacker must be informed of the assets that should be protected, potential
threat sources, and the extent to which the organization will support an ethical hacker's
efforts.
✔ Malicious attackers are, generally known as both, hackers and malicious users.
✔ Malicious user means a rogue employee, contractor, intern, or other user who
abuses his or her trusted privileges .It is a common term in security circles.
Users search through critical database systems to collect sensitive information, e-mail
confidential client information to the competition or elsewhere to the cloud, or delete
sensitive files from servers that they probably do not have access.
There’s also the occasional ignorant insider whose intent is not malicious but who
still causes security problems by moving, deleting, or corrupting sensitive
information. Even an innocent “fat-finger” on the keyboard can have terrible
consequences in the business world.
Malicious users are often the worst enemies of IT and information security
professionals because they know exactly where to go to get the goods and don’t need
to be computer savvy to compromise sensitive information. These users have the
access they need and the management trusts them, often without question. In short
they take the undue advantage of the trust of the management.
Hackers are classified according to the intent of their actions.
The following list classifies hackers according to their intent.
Symbol Description
Symbol Description
5.1.3 Data Privacy and General Data Protection and Regulation (GDPR):
Data privacy is a guideline for how data should be collected or handled, based on its
sensitivity and importance. Data privacy is typically applied to personal health
information (PHI) and personally identifiable information (PII). This includes financial
information, medical records, social security or ID numbers, names, birthdates, and
contact information.
Data privacy concerns apply to all sensitive information that organizations handle,
including that of customers, shareholders, and employees. Often, this information plays
a vital role in business operations, development, and finances.
Data privacy helps ensure that sensitive data is only accessible to approved parties. It
prevents criminals from being able to maliciously use data and helps ensure that
organizations meet regulatory requirements.
Data privacy defines who has access to data, while data protection provides tools and
policies to actually restrict access to the data.
Data protection principles help protect data and make it available under any
circumstances. It covers operational data backup and business continuity/disaster
recovery (BCDR) and involves implementing aspects of data management and data
availability.
Maharashtra State Board of Technical Education P a g e 109 | 151
Emerging Trends in CO and IT (22618)
● Data availability—ensuring users can access and use the data required to perform
business even when this data is lost or damaged.
● Data lifecycle management—involves automating the transmission of critical data to
offline and online storage.
● Information lifecycle management—involves the valuation, cataloging, and protection
of information assets from various sources, including facility outages and disruptions,
application and user errors, machine failure, and malware and virus attacks.
5.1.4 GDPR:
The GDPR is a legal standard that protects the personal data of European Union (EU)
citizens and affects any organization that stores or processes their personal data, even if
it does not have a business presence in the EU.
Because there are hundreds of millions of European Internet users, the standard affects
almost every company that collects data from customers or prospects over the Internet.
GDPR non-compliance carries severe sanctions, with fines up to 4% of annual revenue
or €20 million.
GDPR legislators aimed to define data privacy as a basic human right, and standardize
the protection of personal data while putting data subjects in control of the use and
retention of their data.
There are two primary roles in the GDPR: the GDPR Data Controller is an entity that
collects or processes personal data for its own purposes, and a GDPR Data
Processor is an entity that holds or processes this type of data on behalf of another
organization.
“Personal data”, according to the legal definition of the GDPR legislation, is any
information about an identified or identifiable person, known as a data subject.
Personal data includes any information that can be used, alone or in combination with
other information, to identify someone.
This includes: name, address, ID or passport number, financial info, cultural details, IP
addresses, or medical data used by healthcare professionals or institutions.
Other special data you may not process or store: Race or ethnicity, sexual
orientation, religious beliefs, political beliefs of memberships, health data (unless the
explicit concern is granted or there is substantial public interest).
● Collecting data from children — requires parental consent until children are between
13-16 years old.
● Data portability and access — data subjects must be able to access their data as stored
by the Data Controller, know-how and why it is being processed, and where it is being
sent.
● Correcting and objecting to data — data subjects should be able to correct incorrect
or incomplete data, and data controllers must notify all data recipients of the change.
They should also be able to object to the use of their data, and Data Controllers must
comply unless they have a legitimate interest that overrides the data subject’s interest.
● Right to erasure — data subjects can ask data controllers to “forget” their personal
data. Organizations may be permitted to retain the data, for example, if they need it to
comply with a legal obligation or if it is in the public interest, for example in the case
of scientific or historical research.
● Automated decision-making — data subjects have the right to know that they were
subject to an automated decision based on their private information, and can request
that the automated decision is reviewed by a person, or contest the automated decision.
● Notification of breaches — if personal data under the responsibility of a data controller
is exposed to unauthorized parties, the controller must notify the Data Protection
Authority in the relevant EU country within 72 hours, and in some cases also needs to
inform individual data subjects.
● Transferring data outside the EU — if personal data is transferred outside the EU,
the data controller should ensure there are equivalent measures to protect the data and
the rights of data subjects.
1. Data discovery—a first step in data protection, this involves discovering which
data sets exist in the organization, which of them are business critical and which
contains sensitive data that might be subject to compliance regulations.
2. Data loss prevention (DLP)—a set of strategies and tools that you can use to
prevent data from being stolen, lost, or accidentally deleted. Data loss prevention
solutions often include several tools to protect against and recover from data loss.
3. Storage with built-in data protection—modern storage equipment provides
built-in disk clustering and redundancy.
4. Backup—creates copies of data and stores them separately, making it possible
to restore the data later in case of loss or modification. Backups are a critical
strategy for ensuring business continuity when original data is lost, destroyed, or
damaged, either accidentally or maliciously.
Maharashtra State Board of Technical Education P a g e 111 | 151
Emerging Trends in CO and IT (22618)
This is the only proven method of greatly hardening your systems from attack. If
weaknesses are not identified, it’s a matter of time before the vulnerabilities are
exploited.
As hackers expand their knowledge, one should also gain the required knowledge of
it. You must think like them to protect your systems from them. As the ethical hacker,
one must know activities hackers carry out and how to stop their efforts. One should
know what to look for and how to use that information to spoil hackers’ efforts.
One cannot protect the systems from everything. The only protection against
everything is to unplug computer systems and lock them away so no one can touch
them , not even you.
That’s not the best approach to information security. What’s important is to protect
your systems from known vulnerabilities and common hacker attacks. It’s impossible
to support all possible vulnerabilities on all systems. One can’t plan for all possible
attacks, especially the ones that are currently unknown.
However, the more combinations you try — the more you test whole systems instead
of individual units ,the better your chances of discovering vulnerabilities that affect
everything as a whole.
Building the Foundation for Ethical Hacking:
One should not forget about insider threats from malicious employees. One’s overall
goals as an ethical hacker should be as follows:
✔ Hack your systems in a non-destructive fashion.
✔ Enumerate vulnerabilities and, if necessary, prove to upper management that
vulnerabilities exist.
✔ Apply results to remove vulnerabilities and better secure your systems.
on production systems during regular business hours. One can even create an account
or system lockout condition by social engineering, changing a password, not realizing
that doing so might create a system lockout condition.
One should not stop with one security hole. This can lead to a false sense of
security. One should keep going to see what else he/she can discover. It’s not
like to keep hacking until the end of time or until one crash all his/ her systems.
Simply pursue the path he/she is going down until he//she can’t hack it any
longer.
One of the goals may be to perform the tests without being detected.
For example, one may be performing his/her tests on remote systems or on a
remote office, and he/she doesn’t want the users to be aware of what they are
doing. Otherwise, the users may be on to him/her and be on their best
behaviour.
Extensive knowledge of the systems is not needed for testing . Just a basic
understanding is required to protect the tested systems.
Understanding the systems which are being tested shouldn’t be difficult if one
is hacking his/her own in-house systems. If hacking a customer’s systems, one
may have to dig deeper. In fact, Most people are scared of these assessments.
Base the type of test one will perform on his/her organization’s or customer’s
needs.
Amendments:
A major amendment was made in 2008. It introduced Section 66A which penalized
sending "offensive messages". It also introduced Section 69, which gave authorities the
power of "interception or monitoring or decryption of any information through any
computer resource". Additionally, it introduced provisions addressing
- pornography, child porn, cyber terrorism and voyeurism. The amendment was passed
on 22 December 2008 without any debate in Lok Sabha. The next day it was passed by
the Rajya Sabha. It was signed into law by President Pratibha Patil, on 5 February 2009.
Offences:
List of offences and the corresponding penalties
References
● https://www.dynamicchiropractic.com/mpacms/dc/article.php?id=18078)
● Hacking For Dummies, 5th Edition By Kevin Beaver
● http://cdn.ttgtmedia.com/searchNetworking/downloads/hacking_for_dummie
s
● http://wiki.cas.mcmaster.ca/index.php/Ethical_Hacking
● https://www.dummies.com/programming/networking/what-is-a-malicious-
user/
● https://www.guru99.com/what-is-hacking-an-introduction.html#2
● http://cdn.ttgtmedia.com/searchNetworking/downloads/hacking_for_dummie
s.pdf
● 2600 — The Hacker Quarterly magazine (www.2600.com)
● (IN)SECURE Magazine (www.net-security.org/insecuremag.php)
● Hackin9 (http://hakin9.org)
● PHRACK (www.phrack.org/archives/)
● https://learning.oreilly.com/library/view/hacking-for-dummies/
9781118380956/06_9781118380956-ch02.html
● https://www.quora.com/What-knowledge-is-required-to-become-an-ethical-
hacker
4) Ethical hacking will allow to________ all the massive security breaches.
a. remove
b. measure
c. reject
d. None of these
6.1.3 Scanning-Ports:
● A port scanner is a software tool that basically scans the network to see who’s
there. Port scanners provide basic views of how the network is laid out. They can
help identify unauthorized hosts or applications and network host configuration
errors that can cause serious security vulnerabilities.
● The big-picture view from port scanners often uncovers security issues that may
otherwise go unnoticed. Port scanners are easy to use and can test systems
regardless of what operating systems and applications they’re running. The tests
can be performed very quickly without having to touch individual network hosts,
which would be a real pain otherwise.
● Port-scan tests take time. The length of time depends on the number of hosts you
have, the number of ports you scan, the tools you use, and the speed of your
network links. Also, perform the same tests with different utilities to see whether
you get different results. Not all tools find the same open ports and
vulnerabilities. This is unfortunate, but it’s a reality of ethical hacking tests.
● If your results don’t match after you run the tests using different tools, you may
want to explore the issue further. If something doesn’t look right such as a
strange set of open ports it probably isn’t. Test it again; if you’re in doubt, use
another tool for a different perspective.
● As an ethical hacker, you should scan all 65,535 UDP and 65,535 TCP ports on
each network host that’s found by your scanner. If you find questionable ports,
look for documentation that the application is known and authorized. For speed
and simplicity, you can scan commonly hacked ports.
Port
Nos. Service Protocols
7 Echo TCP, UDP
19 Chargen TCP, UDP
20 FTP data (File Transfer Protocol) TCP
21 FTP control TCP
22 SSH TCP
23 Telnet TCP
25 SMTP (Simple Mail Transfer Protocol) TCP
37 Daytime TCP, UDP
53 DNS (Domain Name System) UDP
69 TFTP (Trivial File Transfer Protocol) UDP
79 Finger TCP, UDP
80 HTTP (Hypertext Transfer Protocol) TCP
110 POP3 (Post Office Protocol version 3) TCP
111 SUN RPC (remote procedure calls) TCP, UDP
RPC/DCE end point mapper for Microsoft
135 networks TCP, UDP
137,
138, NetBIOS over TCP/IP TCP, UDP
139
161 SNMP (Simple Network Management Protocol) TCP, UDP
220 IMAP (Internet Message Access Protocol) TCP
443 HTTPS (HTTP over SSL) TCP
512,
513, Berkeley r commands (such as rsh, rexec, and
514 rlogin) TCP
1214 Kazaa and Morpheus TCP, UDP
1433 Microsoft SQL Server TCP, UDP
1434 Microsoft SQL Monitor TCP, UDP
3389 Windows Terminal Server TCP
5631,
5632 pcAnywhere TCP
6346,
6347 Gnutella TCP, UDP
12345,
12346,
12631, NetBus
12632,
20034,
TCP
20035
27444 Trinoo UDP
27665 Trinoo TCP
31335 Trinoo UDP
31337 Back Orifice UDP
34555 Trinoo UDP
A Serious Threat
● Any times there are open ports on one's personal computer, there is potential for
the loss of data, the occurrence of a virus, and at times, even complete system
compromise.
● It is essential for one to protect his or her virtual files, as new security risks
concerning personal computers are discovered every day.
● Computer protection should be the number one priority for those who use
personal computers.
● Port scanning is considered a serious threat to one's PC, as it can occur without
producing any outward signs to the owner that anything dangerous is taking
place.
Firewall Protection
- Protection from port scanning is often achieved through the use of a firewall. A
firewall monitors incoming and outgoing connections through one's personal
computer.
- One technique used by firewall technology is the opening of all the ports at one
time. This action stops port scans from returning any ports. This has worked in
many situations in the past, however, most experts agree it is best to have all
open ports investigated individually.
- Another approach is to filter all port scans going to one's computer. An individual
can also choose to port scan his or her own system, which enables one to see the
personal computer through the eyes of a hacker.
- Firewalls are the best protection one can invest in with regard to port scanning.
Firewalls deny outside access to an individual's personal computer. With this
type of protection, a personal computer is essentially hidden from unwelcome
visitors and is also protected from a variety of other hacking techniques. With
firewall software, an individual is assured that his or her sensitive and personal
information remains protected.
● A ping sweep of all your network subnets and hosts is a good way to find out
which hosts are alive and kicking on the network.
● A ping sweep is when you ping a range of addresses using Internet Control
Message Protocol (ICMP) packets.
● Dozens of Nmap command-line options exist, which can be overwhelming when
you just want to do a basic scan.
● You can just enter nmap on the command line to see all the options available.
● These command-line options can be used for an Nmap ping sweep:
- sP tells Nmap to perform a ping scan.
- ntells Nmap to not perform name resolution. You may want to omit this if you
want to resolve hostnames to see which systems are responding. Name resolution
may take slightly longer, though.
- -T 4 option tells Nmap to perform an aggressive (faster) scan.
- 192.168.1.1-254 tells Nmap to scan the entire 192.168.1.x subnet.
Vulnerabilities (SNMP)
- The problem is that most network hosts run SNMP that isn’t hardened or
patched to prevent known security vulnerabilities. The majority of network
devices have SNMP enabled and don’t even need it.
- If SNMP is compromised, a hacker can gather such network information as
ARP tables and TCP connections to attack your systems. If SNMP shows up in
port scans, you can bet that a hacker will try to compromise the system.
● Countermeasures (SNMP)
- Preventing SNMP attacks can be as simple as A-B-C:
- Always disable SNMP on hosts if you’re not using it period.
- Block the SNMP port (UDP port 161) at the network perimeter.
Maharashtra State Board of Technical Education P a g e 130 | 151
Emerging Trends in CO and IT (22618)
- Change the default SNMP community string from public to another value
that’s more difficult to guess. This makes SNMP harder to hack.
● Hackers can use ARP Protocol that is running on the network to make their
systems seem as your system or another allowed host on your network.
● A too much number of ARP (Address Resolution Protocol) requests can be a
sign of an ARP poisoning or spoofing attack on your network. Anyone can run
a program, such as dsniff tool or Cain & Abel tool, can modify the ARP tables,
which are responsible for saving IP addresses to media access control (MAC)
address mappings on network hosts.
● That makes the victim machines to think they require to forward traffic to the
hacker’s computer rather than to the correct destination machine when
communicating on the network. And this is a type of man-in-the-middle (MITM)
attacks. Spoofed ARP responses can be sent to a switch, which returns the switch
to broadcast mode and basically turns it into a hub. When this happens, a hacker
can sniff every packet going through the switch and capture anything and
everything from the network.
ARP spoofing
✔ An excessive amount of ARP requests can be a sign of an ARP poisoning
attack (or ARP spoofing) on your network.
✔ What happens is that a client running a program such as the UNIX-based
dsniff or the UNIX- and DOS/Windows-based ettercap can change the
ARP tables the tables that store IP addresses to media access control
(MAC) mappings on network hosts.
✔ This causes the victim computers to think they need to send traffic to the
attacker’s computer, rather than the true destination computer, when
communicating on the network. This is often referred to as a Man-in-the-
Middle (MITM) attack.
MAC-address spoofing
✔ MAC-address spoofing tricks the switch into thinking you (actually, your
computer) are someone else. You simply change your MAC address and
masquerade as another user
✔ You can use this trick to test such access control systems as your IDS,
fire-wall, and even operating-system login controls that check for specific
MAC addresses.
Countermeasures(MAC-daddy attack)
✔ A few countermeasures on your network can minimize the effects of a
hacker attack against ARP and MAC addresses on your network.
- You can prevent MAC-address spoofing if your switches can enable port security
to prevent automatic changes to the switch MAC address tables.
- No realistic countermeasures for ARP poisoning exist. The only way to prevent
ARP poisoning is to create and maintain static ARP entries in your switches for
every host on the network. This is definitely something that no network
administrator has time to do.
Maharashtra State Board of Technical Education P a g e 132 | 151
Emerging Trends in CO and IT (22618)
Detection
✔ You can detect these two types of hacks through either an IDS or a stand-
alone MAC address monitoring utility.
✔ Arp watch is a UNIX-based program alerts you via e-mail if it detects
changes in MAC addresses associated with specific IP addresses on the
network.
✔ Passive Capturing
✔ Encrypted traffic
- Wireless traffic can be captured directly out of the airwaves, making this
communications medium susceptible to malicious eavesdropping.
- Unless the traffic is encrypted, it’s sent and received in clear text just like on a
standard wired network.
- On top of that, the 802.11 encryption protocol, Wired Equivalent Privacy (WEP),
has its own weakness that allows hackers to crack the encryption keys and
decrypt the captured traffic.
✔ Rogue networks
- Watch out for unauthorized Access Points and wireless clients attached to your
network that are running in ad-hoc mode.
- Using NetStumbler or your client manager software, you can test for Access
Points that don’t belong on your network.
- You can also use the network monitoring features in a WLAN analyzer such as
AiroPeek.
- Walk around your building or campus to perform this test to see what you can
find.
- Physically look for devices that don’t belong a well-placed Access Point or
WLAN client that’s turned off won’t show up in your network analysis tools.
- Search near the outskirts of the building or near any publicly accessible areas.
- Scope out boardrooms and the offices of upper level managers for any
unauthorized devices. These are places that are typically off limits but often are
used as locations for hackers to set up rogue Access Points.
✔ Physical-security problems
- Various physical-security vulnerabilities can result in physical theft, the
reconfiguration of wireless devices, and the capturing of confidential
information.
- You should look for the security vulnerabilities when testing your systems such
as Access Points mounted on the outside of a building and accessible to the
public,Poorly mounted antennas or the wrong types of antennas that broadcast
too strong a signal and that are accessible to the public.
Maharashtra State Board of Technical Education P a g e 134 | 151
Emerging Trends in CO and IT (22618)
- You can view the signal strength in NetStumbler or your wireless client manager.
✔ Vulnerable wireless workstations
- Wireless workstations have tons of security vulnerabilities from weak passwords
to unpatched security holes to the storage of WEP(Wired Equivalent Privacy)
keys locally.
- One serious vulnerability is for wireless clients using the Orinoco wireless card.
- The Orinoco Client Manager software stores encrypted WEP keys in the
Windows Registry even for multiple networks.
✔ Default configuration settings
- Similar to wireless workstations, wireless Access Points have many known
vulnerabilities.
- The most common ones are default SSIDs (Service Set IDentifier) and admin
passwords. The more specific ones occur only on certain hardware and software
versions that are posted in vulnerability databases and vendor Web sites.
- The one vulnerability that stands out above all others is that certain Access
Poinits, including Linksys, D-Link, and more, are susceptible to a vulnerability
that exposes any WEP key(s), MAC(Media Access Control) address filters, and
even the admin password! All that hackers have to do to exploit this is to send a
broadcast packet on UDP port 27155 with a string of gstsearch.
6.2.1 Windows:
✔ The Microsoft Windows OS is the most widely used OS in the world.
✔ It’s also the most widely hacked, because Microsoft doesn’t care as much
about security as other OS vendors? The answer is no.Numerous security
mistakes were unnoticed especially in the Windows NT days but because
Microsoft products are so pervasive throughout networks. Microsoft is the
easiest vendor to pick on, and often its Microsoft products that end up in the
crosshairs of hackers. This is the same reason for many vulnerability alerts
on Microsoft products. The one positive about hackers is that they’re driving
the requirement for better security!
✔ There are variants of vulnerabilities that have been around for a long time in
UNIX and Linux, such as the RPC vulnerabilities that the Blaster worm used.
Most Windows attacks are prevented if the patches were properly applied.
Thus, poor security management is often the real reason Windows attacks are
successful
- Much vulnerability have been published for windows operating system.
- Some of the common vulnerabilities found in all versions of windows are:
DoS, Remote Code Execution, Memory Corruption, Overflow, Sql
Injection, XSS, Http Response Splitting, Directory Traversal, Bypass
something Gain Information /Privileges, CSRF File Inclusion etc.
- The maximum number of vulnerabilities detected were of Gaining
Privileges by which the confidentiality and integrity was highly impacted.
● Windows Vulnerabilities
✔ Due to the ease of use of Windows, many organizations have moved to
the Microsoft platform for their networking needs.
✔ Many businesses especially the small to medium sized ones depend solely
on the Windows OS for network usage.
✔ Many large organizations run critical servers such as Web servers and
database servers on the Windows platform.
✔ If security vulnerabilities aren’t addressed and managed properly, they
can bring a network or an entire organization to its knees.
✔ When Windows and other Microsoft software are attacked especially by
a widespread Internet-based worm or virus hundreds of thousands of
organizations and millions of computers are affected.
✔ Many well-known attacks against Windows can lead to
- Leakage of confidential information, including files being copied and credit card
numbers being stolen
- Passwords being cracked and used to carry out other attacks
- Systems taken completely offline by DoS attacks
- Entire databases being corrupted or deleted when insecure Windows-based
systems are attacked, serious things can happen to a tremendous number of
computers around the world.
- Autoplay feature came in Windows XP. This feature checks removable media/
devices then identifies and launches appropriate application based on its
contents. This feature is useful for authentic users but is a gateway for an
attacker.
- Clipboard vulnerability can allow attacker to get access to the sensitive clipboard
data. In windows clipboard is common for all applications. This may lead to
Maharashtra State Board of Technical Education P a g e 136 | 151
Emerging Trends in CO and IT (22618)
● Linux Vulnerabilities
✔ Vulnerabilities and hacker attacks against Linux are affecting a growing
number of organizations especially e-commerce companies and ISPs that
rely on Linux for many of their systems.
✔ When Linux systems are hacked, the victim organizations can experience
the same side effects as if they were running Windows, including:
- Leakage of confidential intellectual property and customer
information
- Passwords being cracked
- Systems taken completely offline by DoS attacks
- Corrupted or deleted databases
● Use new operating systems. It's time to throw out legacy programs with expired
support systems. Newer code comes with more protections.
● Watch the language. Programs written in COBOL, Python, and Java are likely
safer than others.
● Add space. Some programs allow for executable space protections. When
enabled, a hacker can't execute code inserted via an overflow attack.
● Lean on developers. System administrators often complain that
developers ignore their bug reports. Be persistent. When you spot a problem
leading to a buffer overflow, keep talking about it until someone fixes it.
● Apply your patches. When developers do find out about buffer overflow
problems, they fix them with code.
Maharashtra State Board of Technical Education P a g e 138 | 151
Emerging Trends in CO and IT (22618)
● Email Bombs
✔ E-mail bombs can crash a server and provide unauthorized administrator
access.
✔ They attack by creating DoS conditions against your e-mail software and
even your network and Internet connection by taking up so much bandwidth
and requiring so much storage space.
✔ An email bomb is a form of Internet abuse which is perpetrated through the
sending of massive volumes of email to a specific email address with the
goal of overflowing the mailbox and overwhelming the mail server hosting
the address, making it into some form of denial of service attack.
✔ An email bomb is also known as a letter bomb.
✔ Different email bomb attacks are as attachment overloading attack,
connection attack, autoresponder attack.
b. Bandwidth blocking
- An attacker can crash your e-mail service or bring it to a crawl by filling the
incoming Internet connection with junk. Even if your system automatically
identifies and discards obvious attachment attacks, the bogus messages eat
resources and delay processing of valid messages
2. Connection Attack
✔ A hacker can send a huge amount of e-mails simultaneously to addresses on
your network.
✔ These connection attacks can cause the server to give up on servicing any
inbound or outbound TCP requests.
✔ This can lead to a complete server lockup or a crash, often resulting in a
condition where the attacker is allowed administrator or root access to the
system!
✔ This attack is often carried out as spam attack.
3. Autoresponders Attack
✔ This is an interesting attack to find two or more users on the same or different e-
mail systems that have autoresponder configured.
✔ Autoresponder is that annoying automatic e-mail response you often get back
from random users when you’re subscribing to a mailing list.
✔ A message goes to the mailing list of subscribers and then users have their e-mail
configured to automatically respond back, saying they’re out of the office or, on
vacation.
Maharashtra State Board of Technical Education P a g e 141 | 151
Emerging Trends in CO and IT (22618)
Countermeasures (Banners)
There is not a 100 percent secure way of disguising banner information.
Following are some banner security tips for SMTP, POP3, and IMAP servers:
- Change your default banners to cover up the information.
- Make sure that you’re always running the latest software patches.
- Harden your server as much as possible by using well-known best practices
Directory Traversals
✔ A directory traversal is a really basic attack, but it can turn up interesting
information about a Web site.
✔ This attack is basically browsing a site and looking for clues about the server’s
directory structure.
✔ Properly controlling access to web content is crucial for running a secure web
server.
✔ Directory traversal or Path Traversal is an HTTP attack which allows attackers
to access restricted directories and execute commands outside of the web
server’s root directory.
✔ Web servers provide two main levels of security mechanisms
Root directory
- The root directory is the top-most directory on the server file System.
- User access is confined to the root directory, meaning users are unable to access
directories or files outside of the root
Countermeasures (Directory Traversal Attack)
✔ There are two main countermeasures to having files compromised via
Malicious directory traversals:
o Don’t store old, sensitive, or otherwise nonpublic files on your Web
server.
- The only files that should be in your /htdocs or Document Root folder are those
that are needed for the site to function properly.
- These files should not contain confidential information that you don’t want the
world to see.
o Ensure that your Web server is properly configured to allow public
access only to those directories that are needed for the site to function.
- Minimum necessary privileges are key here, so provide access only to the
bare-minimum files and directories needed for the Web application to
perform properly.
Google Dorking:
Google dorking is a hacking technique that makes use of Google's advanced
search services to locate valuable data or hard-to-find content. Google dorking
is also known as "Google hacking."
At the surface level, Google dorking involves using specific modifiers to search
data. For example, instead of searching the entire Web, users can click on tags
like "image" or "site" to collect images or find information about a specific site.
Users can utilize other commands like "filetype" and "datarange" to get other
specific search results.
Although benign types of Google dorking simply use the resources that are
available from Google, some forms of it are concerning to regulators and security
specialists because they could indicate hacking or cyber attack reconnaissance.
Hackers and other cyber-criminals can use these types of Google Dorking to
obtain unauthorized data or to exploit security vulnerabilities in websites, which
is why this term is gaining a negative connotation from the security community.
Understanding Google Dorks and How Hackers Use Them:
The idea of using Google as a hacking tool or platform certainly isn’t novel, and hackers
have been leveraging this incredibly popular search engine for years. Google Dorks had
their roots in 2002 when a man named Johnny Long started using custom queries to
search for elements of certain websites that he could leverage in an attack. At its core,
that’s what Google Dorks are – a way to use the search engine to pinpoint websites that
have certain flaws, vulnerabilities, and sensitive information that can be taken
advantage of. As a side note, some people refer to Google Dorks as Google Hacking
(they’re more or less synonymous terms).
Google Dorks can uncover great information such as email addresses and lists, login
credentials, sensitive files, website vulnerabilities, and even financial information (e.g.,
Payment card data).
● intitle – This allows a hacker to search for pages with specific text in their HTML
title. So intitle: “login page” will help a hacker scour the web for login pages.
● allintitle – Similar to the previous operator, but only returns results for pages that
meet all of the keyword criteria.
● inurl – Allows a hacker to search for pages based on the text contained in the URL
(i.e., “login.php”).
● allinurl – Similar to the previous operator, but only returns matches for URLs that
meet all the matching criteria.
● filetype – Helps a hacker narrow down search results to specific files such as PHP,
PDF, or TXT file types.
● ext – Very similar to filetype, but this looks for files based on their file extension.
● intext – This operator searches the entire content of a given page for keywords
supplied by the hacker.
● allintext – Similar to the previous operator but requires a page to match all of the
given keywords.
● site – Limits the scope of a query to a single website.
can help even the smallest of businesses secure their database enough to make
an attacker move on to an easier target.
Use a WAF
- Employ web application firewalls.
- The misconception here might be that protecting the web server has nothing to
do with the database.
- Nothing could be further from the truth. In addition to protecting a site against
cross-site scripting vulnerabilities and web site vandalism, a good application
firewall can thwart SQL injection attacks as well.
- By preventing the injection of SQL queries by an attacker, the firewall can help
keep sensitive information stored in the database away from prying eyes.
- Web sites that are rich with third-party applications, widgets, components and
various other plug-ins and add-ons can easily find themselves a target to an
exploit that should have been patched.
References:
1. Hacking for Dummies (5th Edition), Kevin Beaver CISSP, Wiley Publishing
Inc.
ISBN: 978-81-265-6554-2
2. CISSP for Dummies(5th Edition),Lawrence C. Miller, Peter H. Gregory, ISBN:
978-1-119-21023-8
3. http://www.applicure.com/blog/database-security-best-practice
4. https://thecybersecurityplace.com/database-hacking-its-prevention
5. https://www.valencynetworks.com/blogs/cyber-attacks-explained-database-
hacking
6. https://www.acunetix.com/websitesecurity/directory-traversal
7. https://www.veracode.com/security/directory-traversal
8. https://www.hackingloops.com/google-dorks
2. Which of the following tool is used for Network Testing and port Scanning
a. NetCat
b. SuperScan
c. NetScan
d. All of Above
3. Banner grabbing is often used for
a. White Hat Hacking
b. Black Hat Hacking
c. Gray Hat Hacking
d. Script Kiddies
7) In _____, your hacker corrupts data within the ____, and that code change forces
your system to overwrite important data.
a. Stack Based, heap
b. Stack Based, stack
c. Heap-based, heap
d. Heap-based, stack