IoT NOtes
IoT NOtes
IoT NOtes
IIOT ANALYTICS
Big Data Analytics and Software Defined Networks, Machine Learning and Data Science,
Julia Programming, Data Management with Hadoop
Big data analytics is the process of collecting, examining, and analyzing large amounts of data to
discover market trends, insights, and patterns that can help companies make better business
decisions. This information is available quickly and efficiently so that companies can be agile in
crafting plans to maintain their competitive advantage.
Technologies such as business intelligence (BI) tools and systems help organizations take the
unstructured and structured data from multiple sources. Users (typically employees) input queries
into these tools to understand business operations and performance. Big data analytics uses the
four data analysis methods to uncover meaningful insights and derive solutions.
So, what makes data “big”?
B
Example of big data analytics
For example, big data analytics is integral to the modern health care industry. As you can imagine,
thousands of patient records, insurance plans, prescriptions, and vaccine information need to be
managed. It comprises huge amounts of structured and unstructured data, which can offer
important insights when analytics are applied. Big data analytics does this quickly and efficiently
so that health care providers can use the information to make informed, life-saving diagnoses.
Why is big data analytics important?
Big data analytics is important because it helps companies leverage their data to identify
opportunities for improvement and optimization. Across different business segments, increasing
efficiency leads to overall more intelligent operations, higher profits, and satisfied customers. Big
data analytics helps companies reduce costs and develop better, customer-centric products and
services.
Data analytics helps provide insights that improve the way our society functions. In health care, big
data analytics not only keeps track of and analyzes individual records, but plays a critical role in
measuring COVID-19 outcomes on a global scale. It informs health ministries within each nation’s
government on how to proceed with vaccinations and devises solutions for mitigating pandemic
outbreaks in the future.
Benefits of big data analytics
There are quite a few advantages to incorporating big data analytics into a business or organization.
These include:
Cost reduction: Big data can reduce costs in storing all the business data in one place. Tracking
analytics also helps companies find ways to work more efficiently to cut costs wherever possible.
Product development: Developing and marketing new products, services, or brands is much
easier when based on data collected from customers’ needs and wants. Big data analytics also
helps businesses understand product viability and keep up with trends.
Strategic business decisions: The ability to constantly analyze data helps businesses make better
and faster decisions, such as cost and supply chain optimization.
Customer experience: Data-driven algorithms help marketing efforts (targeted ads, as an
example) and increase customer satisfaction by delivering an enhanced customer experience.
Risk management: Businesses can identify risks by analyzing data patterns and developing
solutions for managing those risks.
Big data in the real world
Big data analytics helps companies and governments make sense of data and make better,
informed decisions.
There are four main types of big data analytics that support and inform different business
decisions.
1. Descriptive analytics
Descriptive analytics refers to data that can be easily read and interpreted. This data helps create
reports and visualize information that can detail company profits and sales.
Diagnostics analytics helps companies understand why a problem occurred. Big data technologies
and tools allow users to mine and recover data that helps dissect an issue and prevent it from
happening in the future.
Example: A clothing company’s sales have decreased even though customers continue to add
items to their shopping carts. Diagnostics analytics helped to understand that the payment page was
not working properly for a few weeks.
3. Predictive analytics
Predictive analytics looks at past and present data to make predictions. With artificial intelligence
(AI), machine learning, and data mining, users can analyze the data to predict market trends.
Example: In the manufacturing sector, companies can use algorithms based on historical data to
predict if or when a piece of equipment will malfunction or break down.
4. Prescriptive analytics
Example: Within the energy sector, utility companies, gas producers, and pipeline owners identify
factors that affect the price of oil and gas in order to hedge risks.
Tools used in big data analytics
Harnessing all of that data requires tools. Thankfully, technology has advanced so that there are
many intuitive software systems available for data analysts to use.
Hadoop: An open-source framework that stores and processes big data sets. Hadoop is able to
handle and analyze structured and unstructured data.
Spark: An open-source cluster computing framework used for real-time processing and analyzing
data.
Data integration software: Programs that allow big data to be streamlined across different
platforms, such as MongoDB, Apache, Hadoop, and Amazon EMR.
Stream analytics tools: Systems that filter, aggregate, and analyze data that might be stored in
different platforms and formats, such as Kafka.
Distributed storage: Databases that can split data across multiple servers and have the capability
to identify lost or corrupt data, such as Cassandra.
Predictive analytics hardware and software: Systems that process large amounts of complex
data, using machine learning and algorithms to predict future outcomes, such as fraud detection,
marketing, and risk assessments.
Data mining tools: Programs that allow users to search within structured and unstructured big
data.
NoSQL databases: Non-relational data management systems ideal for dealing with raw and
unstructured data.
Data warehouses: Storage for large amounts of data collected from many different sources,
typically using predefined schemas.
Familiarizing yourself with big data analytics tools
Knowing how to use industry-standard tools like the ones mentioned above is essential. Whether
you're pursuing a promotion or interested in obtaining a more advanced role, you can gain
experience in commonly used tools online. Here are a few relevant options to consider:
Introduction to Big Data with Spark and Hadoop
Google Cloud Big Data and Machine Learning Specialization
Data Engineering, Big Data, and Machine Learning on GCP Specialization
PostgreSQL for Everybody Specialization
Advance your career with big data analytics skills
If you want to expand your big data analytics skill set to meet your career goals, you have options.
For example, you can learn from an industry leader while earning a credential for your resume
with IBM's Introduction to Data Analytics. And, in just six months or less, you can learn in-
demand, job-ready skills like data cleaning, analysis, and visualization with the Google Data
Analytics Professional Certificate. You'll also gain hands-on experience with spreadsheets, SQL
programming, and Tableau.
SDN represents a substantial step forward from traditional networking, in that it enables the
following:
The key difference between SDN and traditional networking is infrastructure: SDN is software-
based, while traditional networking is hardware-based. Because the control plane is software-
based, SDN is much more flexible than traditional networking. It allows administrators to control
the network, change configuration settings, provision resources, and increase network capacity
— all from a centralized user interface, without the need for more hardware.
There are also security differences between SDN and traditional networking. Thanks to greater
visibility and the ability to define secure pathways, SDN offers better security in many ways.
However, because software-defined networks use a centralized controller, securing the controller
is crucial to maintaining a secure network.
How does Software-Defined Networking (SDN) work?
Here are the SDN basics: In SDN (like anything virtualized), the software is decoupled from the
hardware. SDN moves the control plane that determines where to send traffic to software, and
leaves the data plane that actually forwards the traffic in the hardware. This allows network
administrators who use software-defined networking to program and control the entire network
via a single pane of glass instead of on a device by device basis.
There are three parts to a typical SDN architecture, which may be located in different physical
locations:
Many of today’s services and applications, especially when they involve the cloud, could not
function without SDN. SDN allows data to move easily between distributed locations, which is
critical for cloud applications.
Additionally, SDN supports moving workloads around a network quickly. For instance, dividing
a virtual network into sections, using a technique called network functions virtualization (NFV),
allows telecommunications providers to move customer services to less expensive servers or
even to the customer’s own servers. Service providers can use a virtual network infrastructure to
shift workloads from private to public cloud infrastructures as necessary, and to make new
customer services available instantly. SDN also makes it easier for any network to flex and scale
as network administrators add or remove virtual machines, whether those machines are on-
premises or in the cloud.
Finally, because of the speed and flexibility offered by SDN, it is able to support emerging trends
and technologies such as edge computing and the Internet of Things, which require transferring
data quickly and easily between remote sites.
How is SDN different from Traditional Networking?
The key difference between SDN and traditional networking is infrastructure: SDN is software-
based, while traditional networking is hardware-based. Because the control plane is software-
based, SDN is much more flexible than traditional networking. It allows administrators to control
the network, change configuration settings, provision resources, and increase network capacity—
all from a centralized user interface, without adding more hardware.
There are also security differences between SDN and traditional networking. Thanks to greater
visibility and the ability to define secure pathways, SDN offers better security in many ways.
However, because software-defined networks use a centralized controller, securing the controller
is crucial to maintaining a secure network, and this single point of failure represents a potential
vulnerability of SDN.
What are the different models of SDN?
While the premise of centralized software controlling the flow of data in switches and routers
applies to all software-defined networking, there are different models of SDN.
Data Science is the study of data cleansing, preparation, and analysis, while machine learning is
a branch of AI and subfield of data science. Data Science and Machine Learning are the two
popular modern technologies, and they are growing with an immoderate rate. But these two
buzzwords, along with artificial intelligence and deep learning are very confusing term, so it is
important to understand how they are different from each other. In this topic, we will understand
the difference between Data Science and Machine Learning only, and how they relate to each
other.
Data Science and Machine Learning are closely related to each other but have different
functionalities and different goals. At a glance, Data Science is a field to study the approaches to
find insights from the raw data. Whereas, Machine Learning is a technique used by the group of
data scientists to enable the machines to learn automatically from the past data. To understand
the difference in-depth, let's first have a brief introduction to these two technologies.
Note: Data Science and Machine Learning are closely related to each other but cannot be
treated as synonyms.
Data science, as its name suggests, is all about the data. Hence, we can define it as, "A field of
deep study of data that includes extracting useful insights from the data, and processing that
information using different tools, statistical models, and Machine learning algorithms." It is a
concept that is used to handle big data that includes data cleaning, data preparation, data analysis,
and data visualization.
A data scientist collects the raw data from various sources, prepares and pre-processes the data,
and applies machine learning algorithms, predictive analysis to extract useful insights from the
collected data.
For example, Netflix uses data science techniques to understand user interest by mining the data
and viewing patterns of its users.
Machine learning is a part of artificial intelligence and the subfield of Data Science. It is a
growing technology that enables machines to learn from past data and perform a given task
automatically. It can be defined as:
Machine Leaning allows the computers to learn from the past experiences by its own, it uses
statistical methods to improve the performance and predict the output without being explicitly
programmed.
The use of machine learning in data science can be understood by the development process or
life cycle of Data Science. The different steps that occur in Data science lifecycle are as follows:
1. Business Requirements: In this step, we try to understand the requirement for the
business problem for which we want to use it. Suppose we want to create a
recommendation system, and the business requirement is to increase sales.
2. Data Acquisition: In this step, the data is acquired to solve the given problem. For the
recommendation system, we can get the ratings provided by the user for different
products, comments, purchase history, etc.
3. Data Processing: In this step, the raw data acquired from the previous step is
transformed into a suitable format, so that it can be easily used by the further steps.
4. Data Exploration: It is a step where we understand the patterns of the data, and try to
find out the useful insights from the data.
5. Modeling: The data modeling is a step where machine learning algorithms are used. So,
this step includes the whole machine learning process. The machine learning process
involves importing the data, data cleaning, building a model, training the model, testing
the model, and improving the model's efficiency.
6. Deployment & Optimization: This is the last step where the model is deployed on an
actual project, and the performance of the model is checked.
The below table describes the basic differences between Data Science and ML:
It deals with understanding and finding hidden It is a subfield of data science that enables the machin
patterns or useful insights from the data, which learn from the past data and experiences automatically.
helps to take smarter business decisions.
It is used for discovering insights from the data. It is used for making predictions and classifying the re
for new data points.
It is a broad term that includes various steps to It is used in the data modeling step of the data science a
create a model for a given problem and deploy the complete process.
model.
A data scientist needs to have skills to use big data Machine Learning Engineer needs to have skills such
tools like Hadoop, Hive and Pig, statistics, computer science fundamentals, programming skills
programming in Python, R, or Scala. Python or R, statistics and probability concepts, etc.
It can work with raw, structured, and unstructured It mostly requires structured data to work on.
data.
Data scientists spent lots of time in handling the ML engineers spend a lot of time for managing
data, cleansing the data, and understanding its complexities that occur during the implementation
patterns. algorithms and mathematical concepts behind that.
As we all know Hadoop is a framework written in Java that utilizes a large cluster of
commodity hardware to maintain and store big size data. Hadoop works on MapReduce
Programming Algorithm that was introduced by Google. Today lots of Big Brand Companies
are using Hadoop in their Organization to deal with big data, eg. Facebook, Yahoo, Netflix,
eBay, etc. The Hadoop Architecture Mainly consists of 4 components.
MapReduce
HDFS(Hadoop Distributed File System)
YARN(Yet Another Resource Negotiator)
Common Utilities or Hadoop Common
Let’s understand the role of each one of this component in detail.
1. MapReduce
MapReduce nothing but just like an Algorithm or a data structure that is based on the YARN
framework. The major feature of MapReduce is to perform the distributed processing in
parallel in a Hadoop cluster which Makes Hadoop working so fast. When you are dealing with
Big Data, serial processing is no more of any use. MapReduce has mainly 2 tasks which are
divided phase-wise:
In first phase, Map is utilized and in next phase Reduce is utilized.
Here, we can see that the Input is provided to the Map() function then it’s output is used as an
input to the Reduce function and after that, we receive our final output. Let’s understand What
this Map() and Reduce() does.
As we can see that an Input is provided to the Map(), now as we are using Big Data. The Input
is a set of Data. The Map() function here breaks this DataBlocks into Tuples that are nothing
but a key-value pair. These key-value pairs are now sent as input to the Reduce(). The
Reduce() function then combines this broken Tuples or key-value pair based on its Key value
and form set of Tuples, and perform some operation like sorting, summation type job, etc.
which is then sent to the final Output Node. Finally, the Output is Obtained.
The data processing is always done in Reducer depending upon the business requirement of
that industry. This is How First Map() and then Reduce is utilized one by one.
Let’s understand the Map Task and Reduce Task in detail.
Map Task:
RecordReader The purpose of recordreader is to break the records. It is responsible for
providing key-value pairs in a Map() function. The key is actually is its locational
information and value is the data associated with it.
Map: A map is nothing but a user-defined function whose work is to process the Tuples
obtained from record reader. The Map() function either does not generate any key-value
pair or generate multiple pairs of these tuples.
Combiner: Combiner is used for grouping the data in the Map workflow. It is similar to a
Local reducer. The intermediate key-value that are generated in the Map is combined with
the help of this combiner. Using a combiner is not necessary as it is optional.
Partitionar: Partitional is responsible for fetching key-value pairs generated in the Mapper
Phases. The partitioner generates the shards corresponding to each reducer. Hashcode of
each key is also fetched by this partition. Then partitioner performs it’s(Hashcode)
modulus with the number of reducers(key.hashcode()%(number of reducers)).
Reduce Task
Shuffle and Sort: The Task of Reducer starts with this step, the process in which the
Mapper generates the intermediate key-value and transfers them to the Reducer task is
known as Shuffling. Using the Shuffling process the system can sort the data using its key
value.
Once some of the Mapping tasks are done Shuffling begins that is why it is a faster process
and does not wait for the completion of the task performed by Mapper.
Reduce: The main function or task of the Reduce is to gather the Tuple generated from
Map and then perform some sorting and aggregation sort of process on those key-value
depending on its key element.
OutputFormat: Once all the operations are performed, the key-value pairs are written into
the file with the help of record writer, each record in a new line, and the key and value in a
space-separated manner.
2. HDFS
DataNode: DataNodes works as a Slave DataNodes are mainly utilized for storing the data in a
Hadoop cluster, the number of DataNodes can be from 1 to 500 or even more than that. The
more number of DataNode, the Hadoop cluster will be able to store more data. So it is advised
that the DataNode should have High storing capacity to store a large number of file blocks.
High Level Architecture Of Hadoop
File Block In HDFS: Data in HDFS is always stored in terms of blocks. So the single block of
data is divided into multiple blocks of size 128MB which is default and you can also change it
manually.
Let’s understand this concept of breaking down of file in blocks with an example. Suppose you
have uploaded a file of 400MB to your HDFS then what happens is this file got divided into
blocks of 128MB+128MB+128MB+16MB = 400MB size. Means 4 blocks are created each of
128MB except the last one. Hadoop doesn’t know or it doesn’t care about what data is stored
in these blocks so it considers the final file blocks as a partial record as it does not have any
idea regarding it. In the Linux file system, the size of a file block is about 4KB which is very
much less than the default size of file blocks in the Hadoop file system. As we all know
Hadoop is mainly configured for storing the large size data which is in petabyte, this is what
makes Hadoop file system different from other file systems as it can be scaled, nowadays file
blocks of 128MB to 256MB are considered in Hadoop.
Replication In HDFS Replication ensures the availability of the data. Replication is making a
copy of something and the number of times you make a copy of that particular thing can be
expressed as it’s Replication Factor. As we have seen in File blocks that the HDFS stores the
data in the form of various blocks at the same time Hadoop is also configured to make a copy
of those file blocks.
By default, the Replication Factor for Hadoop is set to 3 which can be configured means you
can change it manually as per your requirement like in above example we have made 4 file
blocks which means that 3 Replica or copy of each file block is made means total of 4×3 = 12
blocks are made for the backup purpose.
This is because for running Hadoop we are using commodity hardware (inexpensive system
hardware) which can be crashed at any time. We are not using the supercomputer for our
Hadoop setup. That is why we need such a feature in HDFS which can make copies of that file
blocks for backup purposes, this is known as fault tolerance.
Now one thing we also need to notice that after making so many replica’s of our file blocks we
are wasting so much of our storage but for the big brand organization the data is very much
important than the storage so nobody cares for this extra storage. You can configure the
Replication factor in your hdfs-site.xml file.
Rack Awareness The rack is nothing but just the physical collection of nodes in our Hadoop
cluster (maybe 30 to 40). A large Hadoop cluster is consists of so many Racks . with the help
of this Racks information Namenode chooses the closest Datanode to achieve the maximum
performance while performing the read/write information which reduces the Network Traffic.
HDFS Architecture
YARN is a Framework on which MapReduce works. YARN performs 2 operations that are Job
scheduling and Resource Management. The Purpose of Job schedular is to divide a big task
into small jobs so that each job can be assigned to various slaves in a Hadoop cluster and
Processing can be Maximized. Job Scheduler also keeps track of which job is important, which
job has more priority, dependencies between the jobs and all the other information like job
timing, etc. And the use of Resource Manager is to manage all the resources that are made
available for running a Hadoop cluster.
Features of YARN
Multi-Tenancy
Scalability
Cluster-Utilization
Compatibility
Hadoop common or Common utilities are nothing but our java library and java files or we can
say the java scripts that we need for all the other components present in a Hadoop cluster. these
utilities are used by HDFS, YARN, and MapReduce for running the cluster. Hadoop Common
verify that Hardware failure in a Hadoop cluster is common so it needs to be solved
automatically in software by Hadoop Framework.
UNIT 4
IOT SECURITY
Industrial IoT: Security and Fog Computing - Cloud Computing in IIoT, Fog Computing in IIoT,
Security in IIoT.
Fog computing or edge Computing, originated by cisco, refers to a computing structure located
between the cloud and devices that produce data.
However, this mobile structure allows placing resources including their applications and data in
logical locations to intensify performance.
Hence, the main focus of this mobile structure is to bring the basic analytic services to the
periphery of the network. Moreover, the periphery of the network is where data is produced from
IoT devices.
Fog computing compiles low-latency network connections between devices and analytics
endpoints.
Security
Fog computing security issues arise as there are many devices connected to fog nodes and
at different gateways.
Though authentication plays a major role in establishing the initial set of relations
between IoT devices and fog nodes in the network but this is not sufficient as devices
can always malfunction or are also susceptible to malicious attacks.
Privacy
In fog computing, privacy preservation is more challenging since fog nodes may collect
sensitive data. As a result, concerning the identity of end-users compared to the remote
cloud server that lies in the core network.
Moreover, since fog nodes are scattered in large areas, centralized control is difficult.
Although IoT is in the boom, it delivers a variety of services to the end-users. It still faces many
security and privacy issues.
Security issues
IoT devices are connected to desktops or laptops in our day-to-day life. Lack of security
increases the risk of your personal information leaking while the data is collected and
channeled to the IoT device.
IoT devices are connected to various networks. So, if the IoT device contains any security
issues, it can be harmful to the end-users network. These issues can attack other
systems and damage them.
Privacy issues
The main privacy issue is user information leakage in IoT devices such as data, location,
etc.
As relations between IoT devices and fog nodes in the network play a major role which
mitigates the impact of low latency, location awareness of many IoT applications.
The Six Applications and Benefits of IoT in Manufacturing
The Internet of Things (IoT) is a key component of global industrial transformation, and the
manufacturing sector leads in leveraging this technology. As per IDC Data, the sector had
invested $178 billion in IoT in 2016¹. Operations accounted for the main spending user case,
followed by production asset management, maintenance and field service.
Analysts expect manufacturing to retain that leading position until at least 2020 for multiple
reasons. IoT has the potential to bring another industrial revolution – Industry 4.0 – with
applications that bring rapid returns while enabling manufacturers to adopt digital transformation
in various perspectives: automation, visibility, customer-centricity and reduced time to market.
Let us look at a list of the top six uses, applications and advantages of IoT in the manufacturing
domain.
1. Quality Control
In a standard reactive quality control process, manufacturers produce an item, their quality
control unit tests it, and they hope to catch and rectify the flaws before the product reaches the
market.
IoT makes this process proactive with thermal and video sensors collecting complete product
data through different stages of a product cycle. The products can also be tested at each
manufacturing step to check if their attributes are within specifications. In addition,
instrumentation and monitoring of manufacturing equipment help quality control personnel to
check if and where equipment calibration diverges from standard settings – such inaccuracies
must be thwarted in time to avoid misalignment of products.
IoT’s support in monitoring both equipment settings and the outcomes of each production step
gives manufacturers a stronger assurance of detecting quality problems at the source. Measures
for improvement can, therefore, be taken in time.
RJ Corp, the largest bottler of Pepsi in India², uses IoT sensors to capture different data
parameters required to gauge quality on a real-time basis. As the material gets prepared,
deviations indicate at a quality concern, and the machine can be stopped for immediate
corrective action.
2. Inventory Management
Together with radio frequency identification (RFID), IoT makes inventory management an
efficient and seamless process. Every item in the inventory gets an RFID tag, and each tag has a
unique identification number (UID) comprising encoded digital information about the item.
RFID readers can scan the tags, and the data extracted gets transmitted to the cloud for
processing.
The role of industrial IoT here involves transforming the data acquired by RFID readers into
useful business insights. It creates a record of the location of inventory items, their statuses and
their movements in the supply chain and gives users comparable results. For instance, as per the
data on inventory quantity and location, IoT-based inventory management architecture can help
calculate the volume of raw materials required for an impending manufacturing cycle. The
outputs of IoT-based inventory management can be used in diverse ways. The system can send
an alert to the users if any individual inventory item is missing and can notify them when they
must replenish the materials.
IoT gives cross-channel visibility to supply chain managers with a realistic estimate of available
materials, the arrival of new materials and work-in-progress, optimising shared costs in the value
chain. By tracking the speed of movement and traffic flow of raw materials, manufacturers can
be better prepared to receive them. This improves handling times and enables more efficient
processing of those materials for production.
3. Predictive Maintenance
Traditionally, manufacturers have employed a time-based approach for planning the maintenance
schedules of their machinery and equipment. However, according to the ARC group study³, only
18% of equipment fail on account of age, whereas 82% of failures occur randomly. This proves
that a time-based approach is not efficient and may prove costly in the long run.
Manufacturers can avoid such ineffective maintenance routines by leveraging industrial IoT and
data science for predictive maintenance. By using IoT sensors (on the equipment), they can
monitor its operating environment and perform analytics using related data in the cloud to
evaluate the actual wear and tear. Prompt service and repair enable more efficiency in the
maintenance process, better allocation of work to field technicians and avoidance of downtime
along with more significant cost savings.
As an example, steel plants have several furnaces using water cooling panels for temperature
control. Any leakages in the panels can result in safety issues and production loss. With IoT-
based predictive maintenance, plant managers can strategically identify anomalies and conduct a
root cause analysis to prevent machine failures and delays in production.
If there are disparate suppliers and security protocols and no standardisation, IoT can cause
security concerns. To prevent their assets from getting targeted by malicious attackers,
manufacturers leveraging IoT must link their operations technologies and IT infrastructure. They
should also strategise their BYOD policies to regulate the interaction of personal devices with
manufacturing operations. Support from the cloud and IoT services vendor is critical in this
regard.
5. Smart Metering
IoT has also introduced the manufacturing sector, utilities and other industries to the world of
smart meters that can monitor the consumption of water, electric power and other fuels. IoT
sensors allow organisations to gauge the specific use and deploy practices for more effective
usage of resources.
With customisable end-user dashboards deployed by IoT services vendors, manufacturers can
comprehensively analyse the results of smart meter monitoring. They can also compare costs,
efficiencies and the carbon footprint of alternative resources to incorporate better options in their
manufacturing processes.
6. Smart Packaging
Smart packaging that directly uses materials with embedded interconnectedness provides
advanced benefits of IoT for manufacturers. A prime aspect of smart packaging is that it enables
consumers to engage with it, as well as generates data to handle a product more effectively.
Smart packaging may manifest itself in the form of recipe videos, beauty tutorials and other
demonstrations to explain the product usage.
IoT and packaging work together in different ways, including sensors, QR codes and augmented
reality/virtual reality/mixed reality options. The idea is to add value for consumers and also
collect data – through smart tracking – to further optimise operations and boost efficiency.
At Tata Tele Business Services, we design IoT solutions for a host of industrial operations. Our
team works seamlessly to deploy bespoke tools, connected devices and connected processes to
help clients leverage evolving technology for maximised ROI.
To know more about our IoT solutions for manufacturers, call us at 1800-266-1800.
Sources:
1-https://www.i-scoop.eu/internet-of-things-guide/iot-spending-2020/
2-https://www.dynamiccio.com/how-indias-largest-pepsi-bottler-is-using-iot-to-maintain-
stringent-quality/
3-https://www.arcweb.com/market-studies/enterprise-asset-management
Top of
Form
Cloud Computing
In simple terms, cloud computing can be referred to as data storage centers over the internet on a
pay-as-you-go basis.
With cloud computing, cost and complexity reduce the maintenance of IT infrastructure.
Instead of maintaining the data physically, cloud computing services can be used, such as
storage, databases, etc.
With fog computing, local data storage and scrutiny of time-sensitive data become easier.
With this the amount and the distance of data passed to the cloud are reduced, therefore
reducing the impact of security and privacy issues.
Fog computing upgrades time to action and reduces latency, consequently responding
faster to security and privacy threats.
There being a higher probability for security issues when large amounts of data are
transferred through networks, fog computing reduces the amount of data being
transferred back and forth to the cloud, hence, reducing latency as a result of local
computation while minimizing security risks.
One can be suggested for fog computing on the basis of the following norms:
Associated vehicles
To increase the potentiality of management of systems energy networks need to use real-
time data.
It is also important to execute the remote data near the place where it is produced.
Therefore, these issues can be addressed using fog computing.
Real-time analytics
Data can be transferred from place to place from where it is created using fog computing.
Fog computing is using real-time analytics that passes data from the production
organizations to the financial institutions.
Conclusion
Performance problem arises due to poor security and privacy systems. Therefore, these can be
avoided using fog computing which is a new computing paradigm to avail flexible resources at
the edge of the network to the end-users.
Therefore, fog computing processes data closer to where it builds and solves the challenges of
heightening the data volume.
Over the last decade, cloud service providers have improved technology, particularly data
security and integrity. These remain the top concern of many manufacturing businesses,
but as improvements are rolled out, more and more companies are adopting cloud
solutions.
The cloud is a collection of servers that enable access to software and databases over the internet.
The infrastructure that makes up the cloud comprises more than 8,000 data centers spread around
the world, of which 2,600 are located in the United States. Data centers are facilities that house
servers, network devices, and all the hardware that is needed to provide cloud services.
Cloud computing has substantial advantages over traditional computer and IT resource
management processes. The most notable is a significant reduction in investment needed in IT
infrastructure. Businesses do not need to make large purchases of servers and other hardware.
Also, the expertise required to maintain the infrastructure is handed over to the cloud provider.
Figure 1. The 'cloud' is a collective reference to the communication, storage, processing, and
accessing of data stored on remote data servers. Image used courtesy of Canva
Flexibility and scalability are two other reasons companies are opting for cloud services. For
example, when a business grows and requires more IT capacity, the cloud facilitates a seamless
expansion thanks to on-demand products, including storage increases.
Software programs (called applications, or ‘apps’) and databases stored in cloud repositories are
accessed over the internet. Another added flexibility is opening these programs from multiple
devices since local installation is no longer needed.
The cloud we know today had a different meaning in the early decades of the internet’s
existence. Then, the word was sometimes used as a synonym for the internet itself and referred to
nearly anything representing the internet's architecture that we did not understand.
The modern cloud was born with the first attempts at delivering software over the internet at the
end of the 1990s. Salesforce was one of the first companies that succeeded in doing this.
Figure 2. Amazon's AWS logo from the early period starting in 2006. Image used courtesy
of Amazon Web Services
Another breakthrough came several years later, in 2006, with Amazon Web Services (AWS)
launch. AWS was the first major provider of cloud-based solutions that included storage and
computing resources. AWS is still the largest provider in the cloud market, with a 33% share.
After that, Microsoft, Google, OpenStack, and others followed.
By 2010, there was a considerable offer of cloud services, but the industry started to resist
adopting the technology. Vendors discovered that many businesses, especially in the
manufacturing industry, had concerns over security and reliability. To address this, cloud
providers created the concept of a hybrid cloud.
In the hybrid cloud model, customers share computing resources between on-premise and in the
cloud. On-premise resources are accessible via what is known as a private cloud, while the rest
of the resources are stored in a public cloud. This model has proven successful as most
manufacturing businesses have expressed interest in it.
Figure 3. The 'hybrid' cloud model creates more security for sensitive data and transactions in a
private cloud, but also provides access to the capabilities of the public cloud for intensive
storage and calculations. Image used courtesy of Canva
In the last decade, cloud-service providers have considerably improved their technologies,
particularly data security and integrity. These remain the top concern of many manufacturing
businesses, but as improvements are rolled out, more and more companies are adopting cloud
solutions.
The Internet of Things is a somewhat intriguing term for a very popular category of modern
hardware. IoT is a concept that describes objects, many of them small battery-powered devices,
that can communicate over the internet and with other devices. The term became well-known
thanks to consumer electronics and everyday appliances offering this technology.
IoT devices are also increasingly gaining traction in the automation sector. A specific term was
coined to describe IoT in this area: the Industrial Internet of Things (IIoT). With the growing
interest in data analytics and machine learning, IoT devices play an essential role as data
sources.
IoT devices share raw data with higher-level devices and specialized IoT gateways which send
information over the internet using protocols such as File Transfer (FTP) or MQTT. The data
captured by IoT devices are then available to applications and databases that reside in the cloud.
In the downstream flow, the information captured by IoT devices can be pulled back for many
different purposes, including reporting and statistics, data analysis, and machine learning.
The Cloud, IIoT, and Automation
Thus far, we have touched on some definitions and context information about the Cloud and IoT.
Next, we would like to establish more connections between these concepts and the
manufacturing sector and industrial automation.
The manufacturing industry has expressed interest in cloud computing but has reservations due
to security and a sense of lack of ownership of data ownership. These fears have been easing
thanks to the hybrid cloud model. A big reason for preferring the hybrid model is that many
businesses in this sector already have the IT expertise required to support on-premise
infrastructure. Customers typically keep control of the private cloud portion and contract the
public cloud services.
Working with the hybrid cloud model requires constant collaboration between manufacturing
companies and cloud providers. This collaboration is critical in automated systems that require
high system availability.
On the other front, IIoT devices have also faced a lot of scrutiny for the same reasons of security.
Using the word "Internet" for things that have always been entirely isolated from the outside
world is a paradigm change. But, like with the cloud, as companies understand more about the
technology, they begin to open up and adopt it. This does not mean that the concerns become
less important; companies train themselves and prepare for any incidents.
All projections indicate tremendous growth for the Cloud and IIoT market this decade.
Moreover, as the early adoption phases of these technologies end, we will begin to see mass
adoption from middle to small automated systems.
UNIT 5
Precision Agriculture
Agriculture is one of the most promising fields where UAVs have made a significant impact.
UAVs are used in agriculture for crop monitoring, plant health analysis, and precision
agriculture. Farmers can use drones to collect data about their crops and detect problems such as
pests, nutrient deficiencies, and diseases. The data collected by the drones are used to create
maps and identify areas that need attention, allowing farmers to target their resources more
efficiently. UAVs also help farmers reduce the use of pesticides and fertilizers, resulting in cost
savings and environmental benefits.
Infrastructure
Infrastructure maintenance and inspection are essential for ensuring public safety. However, it
can be challenging and dangerous for human inspectors to access hard-to-reach areas. UAVs
equipped with cameras and sensors can help inspect bridges, tunnels, and buildings more
efficiently and safely. Drones can also be used for construction monitoring and to create 3D
models of structures, enabling engineers to identify potential problems before they become
severe.
UAVs can be used to monitor coastal erosion, sea level rise, and changes in beach morphology.
The data collected can be used to assess the impact of climate change on coastal ecosystems,
identify areas at risk of flooding, and develop strategies to mitigate the effects of sea level rise.
Find here the geospatial data and software that will take your project to another level!
From high to very high resolution datasets are available in cloudeo marketplace to match your
exact needs.
Show me products
Environmental Monitoring
UAVs equipped with sensors and cameras can be used to monitor water quality in lakes and
rivers, track changes in vegetation health, and assess the impact of natural disasters on
ecosystems. Additionally, they can also help detect wildlife poaching and aid in forest
management by mapping deforestation and tracking forest fires. UAVs have become a valuable
tool for environmental scientists, government agencies, and conservationists in the effort to
monitor and protect the environment.
Energy
The energy industry is another area where UAVs have found applications. Drones can be used
for inspecting power lines, wind turbines, and oil rigs, which can be difficult and dangerous for
human inspectors. UAVs can collect data on the condition of equipment and identify areas that
need maintenance or repair. Drones can also help companies reduce their environmental
footprint by monitoring emissions and identifying potential leaks.
Forestry Management
UAVs are useful in forestry for monitoring forest health, detecting wildfires, and mapping
deforestation. With advanced sensors and cameras, UAVs can collect high-resolution data on
tree cover, forest structure, and vegetation health. This data can be used to assess the impact of
climate change, identify areas at risk of deforestation, and plan reforestation efforts. UAVs can
also detect forest fires in their early stages, which can aid in early detection and prevention.