UIUX Module 5

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 37

MODULE 5 -UIUX

Information Search and Data Visualization:


Information visualization is the interactive visual illustrations of conceptual
data that strengthen human understanding. It has emerged from the research
in human-computer interaction and is applied as a critical component in varied
fields. It allows users to see, discover, and understand huge amounts of
information at once.
Information visualization is also an assumption structure, which is typically
followed by formal examination such as statistical hypothesis testing.

Advanced Filtering
Following are the advanced filtering procedures −

 Filtering with complex Boolean queries


 Automatic filtering
 Dynamic queries
 Faceted metadata search
 Query by example
 Implicit search
 Collaborative filtering
 Multilingual searches
 Visual field specification
Hypertext and Hypermedia
Hypertext can be defined as the text that has references to hyperlinks with
immediate access. Any text that provides a reference to another text can be
understood as two nodes of information with the reference forming the link. In
hypertext, all the links are active and when clicked, opens something new.
Hypermedia on the other hand, is an information medium that holds different
types of media, such as, video, CD, and so forth, as well as hyperlinks.
Hence, both hypertext and hypermedia refers to a system of linked
information. A text may refer to links, which may also have visuals or media.
So hypertext can be used as a generic term to denote a document, which may
in fact be distributed across several media.

Object Action Interface Model for Website


Design
Object Action Interface (OAI), can be considered as the next step of the
Graphical User Interface (GUI). This model focusses on the priority of the
object over the actions.
OAI Model
The OAI model allows the user to perform action on the object. First the object
is selected and then the action is performed on the object. Finally, the
outcome is shown to the user. In this model, the user does not have to worry
about the complexity of any syntactical actions.
The object–action model provides an advantage to the user as they gain a
sense of control due to the direct involvement in the design process. The
computer serves as a medium to signify different tools.

FIVE STAGE SEARCH FRAMEWORK:

The following points are the five-phase frameworks that clarifies user
interfaces for textual search −

Formulation − expressing the search

Initiation of action − launching the search

Review of results − reading messages and outcomes

Refinement − formulating the next step

Use − compiling or disseminating insight.

FORMULATION:-

What is the formulation stage of the information search process?

The fourth stage in the ISP, formulation, is the turning point of the ISP, when
feelings of uncertainty diminish and confidence increases. The task is to form
a focus from the information encountered.
Formulation, when a focused perspective is formed and uncertainty
diminishes as confidence begins to increase.

Initiation of Action:-
Consumer behaviour is the study of individuals, groups, or organisations and all the
activities associated with the purchase, use and disposal of goods and services.
Consumer behaviour consists of how the consumer's emotions, attitudes,
and preferences affect buying behaviour. Consumer behaviour emerged in the 1940–
1950s as a distinct sub-discipline of marketing, but has become an interdisciplinary social
science that blends elements from psychology, sociology, social
anthropology, anthropology, ethnography, ethnology, marketing,
and economics (especially behavioural economics).
The study of consumer behaviour formally investigates individual qualities such
as demographics, personality lifestyles, and behavioural variables (such as usage rates,
usage occasion, loyalty, brand advocacy, and willingness to provide referrals), in an
attempt to understand people's wants and consumption patterns. Consumer behaviour
also investigates on the influences on the consumer, from social groups such as family,
friends, sports, and reference groups, to society in general (brand-influencers, opinion
leaders).
Research has shown that consumer behaviour is difficult to predict, even for experts in
the field; however, new research methods, such as ethnography, consumer
neuroscience, and machine learning[1] are shedding new light on how consumers make
decisions. In addition, customer relationship management (CRM) databases have
become an asset for the analysis of customer behaviour. The extensive data produced
by these databases enables detailed examination of behavioural factors that contribute to
customer re-purchase intentions, consumer retention, loyalty, and other behavioural
intentions such as the willingness to provide positive referrals, become brand advocates,
or engage in customer citizenship activities. Databases also assist in market
segmentation, especially behavioural segmentation such as developing loyalty segments,
which can be used to develop tightly targeted customised marketing strategies on a one-
to-one basis..

Review of Results:-

Data visualization is the practice of translating information into a


visual context, such as a map or graph, to make data easier for the
human brain to understand and pull insights from. The main goal of
data visualization is to make it easier to identify patterns, trends and
outliers in large data sets. The term is often used interchangeably
with others, including information graphics, information visualization
and statistical graphics.

Data visualization is one of the steps of the data science process,


which states that after data has been collected, processed and
modeled, it must be visualized for conclusions to be made. Data
visualization is also an element of the broader data presentation
architecture (DPA) discipline, which aims to identify, locate,
manipulate, format and deliver data in the most efficient way
possible.

Data visualization is important for almost every career. It can be used


by teachers to display student test results, by computer scientists
exploring advancements in artificial intelligence (AI) or by executives
looking to share information with stakeholders. It also plays an
important role in big data projects. As businesses accumulated
massive collections of data during the early years of the big data
trend, they needed a way to get an overview of their data quickly and
easily. Visualization tools were a natural fit.

Visualization is central to advanced analytics for similar reasons.


When a data scientist is writing advanced predictive analytics or
machine learning (ML) algorithms, it becomes important to visualize
the outputs to monitor results and ensure that models are performing
as intended. This is because visualizations of complex algorithms are
generally easier to interpret than numerical outputs.

Refining data consists of cleansing and shaping it. When you cleanse data,
you fix or remove data that is incorrect, incomplete, improperly formatted, or
duplicated. And when you shape data, you customize it by filtering, sorting,
combining or removing columns, and performing operations.

As you manipulate your data, you build a customized Data Refinery flow that
you can modify in real time and save for future re-use. When you save the
refined data set, you typically load it to a different location than where you
read it from. In this way, your source data can remain untouched by the
refinement process.
 Prerequisites
 Refine your data
 Data set previews
 Data Refinery flows and steps

Prerequisites
Before you can refine data, you’ll need a project.

If you have data in cloud or on-premises data sources, you’ll


need connections to those sources and you’ll need to add data assets from
each connection. If you want to be able to save refined data to cloud or on-
premises data sources, create connections for this purpose as well. Source
connections can only be used to read data; target connections can only be
used to load (save) data. When you create a target connection, be sure to use
credentials that have Write permission or you won’t be able to save your Data
Refinery flow output to the target.

REFINEMENT:-
Refine your data

Access Data Refinery from within a project. Click Add to project, and
then choose DATA REFINERY FLOW. Then select the data you want
to work with.

Alternatively, from the Assets tab of a project page, you can perform
one of the following actions:

 Select Refine from the menu of an Avro, CSV, JSON, Parquet,


or text data asset
 Click an Avro, CSV, JSON, Parquet, or text data asset to
preview it first and then click the Refine link
 If you already have a Data Refinery flow, click New Data
Refinery flow from the Data Refinery flows section and
then select the data that you want to work with.

Tip: If your data doesn’t display in tabular form, specify the format of
your data source. Go to the Data tab. Scroll down to the SOURCE
FILE information at the bottom of the page. Click the Specify data
format icon.
On the Data tab, apply operations to cleanse, shape, and enrich your
data. You can enter R code in the command line and let autocomplete
assist you in getting the correct syntax.

Validate your data throughout the data refinement process.

Use visualizations to gain insights into your data and to uncover


patterns, trends, and correlations within your data.

When you’ve refined the sample data set to suit your needs, click the
Run Data Refinery flow icon in the toolbar to run the Data Refinery flow
on the entire data set.

By default, Data Refinery uses the name of the data source to name
the Data Refinery flow and the target data set. You can change these
names, but you can’t change the project that these data assets belong
to.

In the Data Refinery flow details pane, optionally edit the name
of the Data Refinery flow and enter a description for it. You can
also add a one-time or repeating schedule for the Data Refinery
flow.

In the Data Refinery flow output pane, optionally edit the name
of the target data set and enter a description for it. You can save
the target data set to the project, to a connection, or to a
connected data asset. If you save it to the project, you can save
it as a new data asset (by default) or you can replace an existing
data asset. To save the target data set to a connection or as a
connected data asset or to replace an existing data asset,
use Change Location.

If you select an existing relational database table or view or you


select a connected relational data asset as the target for your
Data Refinery flow output, you have a number of options for
impacting the existing data set.

 Overwrite - Overwrite the rows in the existing data set


with those in the Data Refinery flow output
 Recreate - Delete the rows in the existing data set and
replace them with the rows in the Data Refinery flow
output
 Insert - Append all rows of the Data Refinery flow output
to the existing data set
 Update - Update rows in the existing data set with the
Data Refinery flow output; don’t insert any new rows
 Upsert - Update rows in the existing data set and append
the rest of the Data Refinery flow output to it

For the Update and Upsert options, you’ll need to select the
columns in the output data set to compare to columns in the
existing data set. The output and target data sets must have the
same number of columns, and the columns must have the same
names and data types in both data sets.

Use - Dynamic Queries and Faceted Search :


The query cycle
1. Send the user's query to the search engine.
2. Execute the search and retrieve the records that match the query. In this step,
you'll derive the facets from the retrieved records .
3. Send back the results and facets.
4. Render the results and facets on screen.

Overview of dynamic faceting


Dynamic faceting displays to the user a different set of
facets depending on the user’s intent. To understand this,
consider an ecommerce music store that sells two
categories of items: CDs and audio equipment. This
business wants to display a set of relevant facets. What
good is it to propose “brand” when the user’s intent is to
find their favorite music? Likewise, what good is it to
propose “musical genre” when the user’s intent is to find
audio equipment? Dynamic faceting ensures that only the
most appropriate facets show up.

Ecommerce businesses with a diversity of products benefit


from displaying different facets depending on the items the
user is searching for:

 Pharmaceutical companies display different facets


for their medical vs. cosmetic products.
 Newspapers display different facets in their
Entertainment and Political sections.
 Online marketplaces, like the Amazon example
below, change facet lists as people navigate through
their vast diversity of offerings.

Example use case: ecommerce


marketplaces & dynamic faceting

Amazon uses dynamic faceting for many of its categories.


In the image below, you see two queries: “music” on the
left, “movies” on the right. As you can see, both sides
include the “price” facet, but the musical query includes
“customer reviews”, “artist”, and “musical format”, while the
movie query includes “director”, “video format” and “movie
genre”.
Amazon uses dynamic faceting to create an enhanced
search experience by guiding the user in a smart and
curated way depending on the products they are
searching for.

Let’s see how this is done.

The query cycle and the logic behind


facet search

First, some terminology:


 Facet keys are attributes like “color”, “price”,
“shoe_category”, and “sleeves”.
 Facet values are the key’s values. For example,
“color” contains “red” and “green”; “sleeves” contain
“short” and “long”.

The dataset

We’ll use a dataset with two kinds of products: shirts and


shoes. The example below contains two typical items. Both
items include the “price”, “color”, and “clothing_type” facets.
However, shirts contain a “sleeves” facet and shoes a
“shoe_category” facet.

The query cycle

A search query follows a 4-part cycle. Here’s an overview.


We’ll give more details and code examples in the section
that follows.

1. Send the user’s query to the search engine.


2. Execute the search and retrieve the records that
match the query. In this step, you’ll derive the facets
from the retrieved records .
3. Send back the results and facets.
4. Render the results and facets onscreen.

As you’ll see, instead of using a pre-defined list of facets,


the logic consists in dynamically generating a new list with
each query. This is possible by doing the following:

 On the back end, you’ll extract facets from every set


of query results.
 On the front end, you’ll use undefined container-
placeholders instead of pre-defined containers.

The query request: sending the query


with or without a filter

The starting point of the cycle is to send a query and any


facet value the user has selected to filter their results.
Filtering results creates a cohesive result set, which in turn
generates a list of facets relevant to all of the items that
appear in the results. On the other hand, if the user does
not select a facet, the items will be more diverse — and
therefore, the facets might not apply to all products.

However, this is perfectly fine. As you’ll see in the next


step, presenting the top 5 most common facets ensures
that most items will contain these facets.
Now for the code. Here’s how Algolia’s API implements the
query “Get all short-sleeved summer t-shirts”. (Since all
search tools allow filtering, the following code is only one
among many ways to do this).

The query execution: creating the list


of top 5 facets

The dataset we use in this article contains two kinds of


products, each with a set of unique facet attributes. After
executing the query, the search engine extracts every facet
key that shows up in every record, then selects 5 facets
that appear most often.

Why top 5? Because a screen with 5 facets is usually


enough. Ten is an outer limit – any more would be overkill
and create unused clutter.

There are two methods to create a list of facets:

1. A pre-defined list: save a list of facet keys either


directly in the code or in a separate dataset or local
storage.
2. A dynamically generated list: extract a distinct list of
facet keys from the results of the query and put them
in the query response as a separate record.
We’ve chosen to use method 2, but some implementations
use method 1.

Method 1 – Hardcoding the list of facet keys

For this method, you create a fixed set of facets based on


what you already know about your products. This list can
be either hardcoded, or placed in a file or table in a
database that can be manually updated whenever new
products are added or modified.

Whatever the manner of storing the fixed list, the list must
contain the following kind of example:

 For every “clothing_type=shirt”, send back the


following facets: “color, price, clothing_type, sleeves,
gender”.
 For every “clothing_type=shoe”, send back the
following facets: “color, price, clothing_type,
shoe_category, brand”.

This is not the preferred method because, as with all


hardcoding or semi-hardcoding, it has limited scalability. If
you want to add more relations, for example “shoe_styles“
= “high-top, leisure, and cross-fit”, you’ll need to manually
add a new line to the list. Manual maintenance is extra
work and prone to error and delay.
The approach we present in the rest of this article (method
2) removes manual maintenance from the process, making
the process deductive and therefore entirely dynamic.

Method 2 – Dynamically generate the list of facets

In this method, we’ll extract the list of facets from the


products themselves.

In a sense, the only difference between the manual and


dynamic approaches is that the generated list of top 5
facets is dynamic. The resulting list itself will be formatted
in the same way.

Here are the steps:

1. Get the query results:

o Execute the query, find X number of products.


o Save those records, to be sent back as the
search results.

2. Get the top 5 facet keys and all their values:


o From the results, collect all facet attributes
(we’ll show below how to identify an attribute
as a facet).
o Create a record that contains the full list of
extracted facet keys.
o Determine which facets appear the most often.
Sort the list by the highest number of records.
o Take only the top 5 from this list. These are the
most common facets.
o Add the product’s values to their respective
key.

Done.

The generated list would be the same as with method 1:

 For every “clothing_type=shirt”, send back the


following facets: “color, price, clothing_type, sleeves,
gender”.
 For every “clothing_type=shoe”, send back the
following facets: “color, price, clothing_type,
shoe_category, brand”.

Thus, if most products are shirts, we would display


“sleeves” and “gender” as the 4th and 5th facets to display.
If they were shoes, then the 4th and 5th facets would be
“shoe_category” and “brand”.

The query response: sending back the


response

Here we’ll simply send back the search results and the
generated list of 5 facet keys with their respective values.
The response should include whatever the front end needs
to build its search results page. In all facet search query
cycles, the front end needs:

 A list of products with “name”, “description”, “price”,


and “image_url”. (Don’t send image files, as their size
will slow down the overall response time of the
search).
 The list of facet key and their values.
 A response will also contain additional information
needed for display purposes or business logic. We
do not show those.

Results:

Here is how we would return the set of shirts. All attributes


will be used as information in the search results, except
“objectID”, which will be used to identify a product for
technical reasons (click analytics, detailed page view, or
other reasons).

Facets:

The facet response is a combination of facet keys with their


values. Here’s a small extract. The example does not
include all of the facets:
Two things to note:

 Only “sleeves” and “gender” show up, not


“shoe_category” or “brand”. As mentioned, this is
because there are more shirts than shoes.
 The number after each facet indicates the number of
records that have that value. We discuss this a bit
more in the section on adding number of facets.

The front-end display: dynamically


displaying the list of facets

The job here is to render the results and facets on the


screen. In terms of UX design, industry standard is to have
the results in the middle, and facets on the left.

There is one container (“facet-lists”) for the 5 facets. The


rendering code generate an unordered list to display the
facets in that container.

The results go in the “results-container”.

Next, render the data. As this can take many forms, and
this article is not strictly a tutorial, take a look at
our dynamic faceting GitHub repo for a complete front-end
implementation.
Making the solution more robust

Adding number of facets

You’ll want to let your users know how many records have
a given facet value. Getting the number of facets is useful
because they inform users about the search results. For
example, it is useful to know that there are more short-
sleeved shirts than long-sleeved shirts. These numbers of
facets are normally calculated in the back end during query
execution.

Adding facet metadata

Every record needs to contain information that helps the


process know which attributes are facets. To do this, you
need to add facet meta data to each record by using an
additional attribute that defines the record’s facets:

Going one step further, it’s also useful to include the type of
attribute:

With this information, the front-end code can apply a range


slider for price and a dropdown for the sleeves.
Grouping items using facets:

You’ll want to treat color differently from the other


facet attributes. This is because different colors appear
on the same shirt. For this, the logic question is: Do
you need 1 record per shirt that includes an attribute
with all available colors? or 1 record per color, which
requires multiple records for each shirt? Typical
database thinking would say, of course, only 1 record.
However, as discussed at length in our first
article, faceted search is different.

We put all searchable items in separate records. This


allows people to find “red shirts” using the search bar
without needing to click on a color facet.

Our main goal in this article was to add process to data in


our series on facets & data. We described a particular way
to execute a search and display a set of facets, following a
query cycle of request, execution, response, and
display. To stretch your understanding of facets, we did this
within the context of an advanced use case: we made our
facet search dynamic. Dynamic faceting creates a more
intuitive and useful facet search experience, particularly for
businesses that offer a diverse collection of products and
services.

Command languages

Typically think of OS command languages: DOS or Unix.


Interaction paradigm: user enters command, computer carries it out (with
feedback if required) then requests next command.

Cognitive requirements: relate to written languages

Not much new activity in command language development.

Command Language Design

Important design considerations:

 Provide only sufficient functionality for users to carry out tasks


 Macro or scripting capability
 Consistency in command language design
 Command abbreviations
 Help feature

 actions become commands


 objects become arguments

Consistency is very important in command language design


(very taxing on memory, so minimize the number of chunks)

 use of metaphor yields consistent terminology (command names)


(easier to learn, retain, recall -- e.g. up/down not forward/down)
 consistent argument ordering
(e.g. first argument is source and second is destination)
 consistent naming of command options
(e.g. v for verbose output regardless of command)

Command abbreviation useful for expert user

 important because expert benefits the most from CUI


 novice will require/prefer full names
 easier to learn if only one abbreviation strategy used.

Multimedia Document Search and Specialized Search:

Multimedia search enables information search using queries in multiple data types
including text and other multimedia formats. Multimedia search can be implemented
through multimodal search interfaces, i.e., interfaces that allow to submit search
queries not only as textual requests, but also through other media. We can distinguish
two methodologies in multimedia search:
 Metadata search: the search is made on the layers of metadata.
 Query by example: The interaction consists in submitting a piece of information
(e.g., a video, an image, or a piece of audio) for the purpose of finding similar
multimedia items.

Metadata search
Search is made using the layers in metadata which contain information of the content of
a multimedia file. Metadata search is easier, faster and effective because instead of
working with complex material, such as an audio, a video or an image, it searches using
text.
There are three processes which should be done in this method:

 Summarization of media content (feature extraction). The result of feature


extraction is a description.
 Filtering of media descriptions (for example, elimination of Redundancy)
 Categorization of media descriptions into classes.

Query by example[edit]
In query by example, the element used to search is a multimedia content (image, audio,
video). In other words, the query is a media. Often, it's used audiovisual indexing. It will
be necessary to choose the criteria we are going to use for creating metadata. The
process of search can be divided in three parts:

 Generate descriptors for the media which we are going to use as query and the
descriptors for the media in our database.
 Compare descriptors of the query and our database’s media.
 List the media sorted by maximum coincidence.

Multimedia search engine


There are two big search families, in function of the content:

 Visual search engine


 Audio search engine
Visual search engine
Inside this family we can distinguish two topics: image search and video search

 Image search: Although usually it's used simple metadata search, increasingly is
being used indexing methods for making the results of users queries more accurate
using query by example. For example, QR codes.
 Video search: Videos can be searched for simple metadata or by complex metadata
generated by indexing. The audio contained in the videos is usually scanned by audio
search engines.
Audio search engine
There are different methods of audio searching:

 Voice search engine: Allows the user to search using speech instead of text. It uses
algorithms of speech recognition. An example of this technology is Google Voice
Search.
 Music search engine: Although most of applications which searches music works on
simple metadata (artist, name of track, album…) . There are some programs of music
recognition, for example Shazam or SoundHound.
Image Search – Video Search – Audio Search:

Searching
To search for media with the Shutterstock API, you need:

 An account at https://www.shutterstock.com
 An application at https://www.shutterstock.com/account/developers/apps
 One of these types of authentication:
o Basic authentication: You provide the client ID and client secret with
each API request.
o OAuth authentication: You use the client ID and client secret to get
an authentication token and use this token in each request. In the
examples on this page, the access token is in
the SHUTTERSTOCK_API_TOKEN environment variable. For more
information about authentication, see Authentication.
Some API subscriptions return a limited set of results. See Subscriptions in the API
reference.

Keyword searches
To search for media with a keyword, pass the search keywords to the appropriate
endpoint:

 Images: GET https://api.shutterstock.com/v2/images/search


 Video: GET https://api.shutterstock.com/v2/videos/search
 Audio: GET https://api.shutterstock.com/v2/audio/search
The search keywords must be URL encoded and in the query query parameter. The
API searches for the keywords in all textual fields, including but not limited to the
description, keywords, and title. For examples of search results, see Results.
Searches do not support wildcards such as *.
Here are some examples of simple image, video, and audio search requests:

Simple image keyword search


cURL
CLI
PHP
JavaScript
curl -X GET "https://api.shutterstock.com/v2/images/search" \

-H "Authorization: Bearer $SHUTTERSTOCK_API_TOKEN" \

-G \

--data-urlencode "query=kites"

Simple video keyword search


cURL
CLI
PHP
JavaScript
curl -X GET "https://api.shutterstock.com/v2/videos/search" \

-H "Authorization: Bearer $SHUTTERSTOCK_API_TOKEN" \

-G \

--data-urlencode "query=hot air balloon"

Simple audio keyword search


cURL
CLI
PHP
JavaScript
curl -X GET "https://api.shutterstock.com/v2/audio/search" \

-H "Authorization: Bearer $SHUTTERSTOCK_API_TOKEN" \

-G \

--data-urlencode "query=bluegrass"

Conditional searches
Searches for images and video can use AND, OR, and NOT conditions, but searches
for audio do not support these keywords.
To use AND, OR, or NOT in searches, you include these operators in
the query query parameter. The operators must be in upper case and they must be
in English, regardless of the language the keywords are in.

 AND is added implicitly between each search keyword. Therefore, searching


for dog AND cat is equivalent to searching for dog cat.
 OR searches for results that include any of the specified keywords, such
as dog OR cat OR mouse.
 NOT searches exclude keywords from search results, such as dog NOT hot
dog. You can also use NOT in the contributor search field.

You can group conditional search terms with parentheses. For example, to search for
images with either dogs or cats, but not both, use (dog NOT cat) OR (cat NOT
dog).
Here are some examples of searching with conditions:
cURL
CLI
PHP
JavaScript
curl -X GET "https://api.shutterstock.com/v2/images/search" \

-H "Authorization: Bearer $SHUTTERSTOCK_API_TOKEN" \

-G \
--data-urlencode "query=dog AND cat"

curl -X GET "https://api.shutterstock.com/v2/videos/search" \

-H "Authorization: Bearer $SHUTTERSTOCK_API_TOKEN" \

-G \

--data-urlencode "query=mountain NOT camping"

Bulk searches
You can run up to 5 image searches in a single request with the POST
/v2/images/bulk_search endpoint. The API returns a maximum of 20 results for
each search. To use this endpoint, pass multiple searches in the queries array in
the request body. Each search in the queries array has the same parameters as an
individual image search.
You can also pass query parameters in the request. These parameters become
defaults for each search, but the parameters in the individual searches override
them.
Here is an example of sending 2 searches in a single request:
DATA='[

"query": "cat",

"license": ["editorial"],

"sort": "popular"

},

"query": "dog",

"orientation": "horizontal"

]'

curl -X POST "https://api.shutterstock.com/v2/images/bulk_search"


\

-H "Authorization: Bearer $SHUTTERSTOCK_API_TOKEN" \

-H "Content-Type: application/json" \
-d "$DATA"

GEOGRAPHIC INFORMATION
SYSTEMS (GIS) SEARCH:
What We Do
Geographic Information Systems (GIS) is used to enhance decision making and provide
maps and information to the public and County agencies including:
 Assessors Office
 Community Development Resource Agency
 County Executive Office
 Elections
 Facility Services
 Health and Human Services
 Public Works

Requests
 For zoning information, please email our Public Counter Technicians.
 For GIS/mapping requests (not zoning), please email the GIS Division.

Services & Information


 Placer County Open GIS Data: Web mapping applications and GIS data download. The place
to find GIS data for Placer County that you can download or incorporate into existing ArcGIS
projects. Here you can also find web apps that include active building permits, code
compliance cases, and planning projects, among others.
 Interactive GIS/Parcel Data Search: Search by Assessors Parcel Number or address for
zoning, parcel related information, aerial photography, etc.
 Assessment Inquiry Website: Data information from the County Assessment Roll (assessed
values, assessment maps, etc.)
 Current Permit Information: Check permit status, permit activity, project descriptions and other
permit-related information
 Permit Research Tools: Various tools available for conducting land-use-related permit
research.

Multilingual Searches :
Multilingual meta-search engines: These search engines aggregate results from
multiple search engines, each specialized in a particular language or region.
They provide users with a unified interface to search across different search
engines and retrieve multilingual results from various sources.

A Moroccan immigrant who speaks some French and German in addition to his
Arabic dialect is multilingual, as is the conference interpreter who confidently
uses her three native or first languages of English, German and French as
working languages..
The Main Benefits of Multilingualism
 Cognitive Benefits. Learning and knowing several languages sharpens the
mind and improves memory. ...
 Improvement of Ability to Multitask. Being able to speak in multiple languages
allows you to perform better in many ways. ...
 Improvement of Communication Skills.

DATA VISUALIZATION:

Data visualization is the graphical representation of information and data. By


using visual elements like charts, graphs, and maps, data visualization tools
provide an accessible way to see and understand trends, outliers, and patterns
in data. Additionally, it provides an excellent way for employees or business
owners to present data to non-technical audiences without confusion.

In the world of Big Data, data visualization tools and technologies are
essential to analyze massive amounts of information and make data-driven
decisions.

What are the advantages and disadvantages of data


visualization?
Something as simple as presenting data in graphic format may seem to have
no downsides. But sometimes data can be misrepresented or misinterpreted
when placed in the wrong style of data visualization. When choosing to create
a data visualization, it’s best to keep both the advantages and disadvantages
in mind.

Advantages

Our eyes are drawn to colors and patterns. We can quickly identify red from
blue, and squares from circles. Our culture is visual, including everything from
art and advertisements to TV and movies. Data visualization is another form of
visual art that grabs our interest and keeps our eyes on the message. When
we see a chart, we quickly see trends and outliers. If we can see something, we
internalize it quickly. It’s storytelling with a purpose. If you’ve ever stared at a
massive spreadsheet of data and couldn’t see a trend, you know how much
more effective a visualization can be.

Some other advantages of data visualization include:

 Easily sharing information.


 Interactively explore opportunities.
 Visualize patterns and relationships.

Disadvantages

While there are many advantages, some of the disadvantages may seem less
obvious. For example, when viewing a visualization with many different
datapoints, it’s easy to make an inaccurate assumption. Or sometimes the
visualization is just designed wrong so that it’s biased or confusing.

Some other disadvantages include:

 Biased or inaccurate information.


 Correlation doesn’t always mean causation.
 Core messages can get lost in translation.

Why data visualization is important


The importance of data visualization is simple: it helps people see, interact
with, and better understand data. Whether simple or complex, the right
visualization can bring everyone on the same page, regardless of their level of
expertise.

It’s hard to think of a professional industry that doesn’t benefit from making
data more understandable. Every STEM field benefits from understanding data
—and so do fields in government, finance, marketing, history, consumer
goods, service industries, education, sports, and so on.

While we’ll always wax poetically about data visualization (you’re on the
Tableau website, after all) there are practical, real-life applications that are
undeniable. And, since visualization is so prolific, it’s also one of the most
useful professional skills to develop. The better you can convey your points
visually, whether in a dashboard or a slide deck, the better you can leverage
that information. The concept of the citizen data scientist is on the rise. Skill
sets are changing to accommodate a data-driven world. It is increasingly
valuable for professionals to be able to use data to make decisions and use
visuals to tell stories of when data informs the who, what, when, where, and
how.

While traditional education typically draws a distinct line between creative


storytelling and technical analysis, the modern professional world also values
those who can cross between the two: data visualization sits right in the
middle of analysis and visual storytelling.

TASKS IN DATA VISUALIZATION:

Common roles for data visualization include:


 showing change over time.
 showing a part-to-whole composition.
 looking at how data is distributed.
 comparing values between groups.
 observing relationships between variables.
 looking at geographical data.

Task-based effectiveness of
basic visualizations
This is a summary of a recent paper on an age-old topic: what visualisation
should I use? No prizes for guessing “it depends!” Is this the paper to finally
settle the age-old debate surrounding pie-charts??

Task-based effectiveness of basic visualizations Saket et al., IEEE

Transactions on Visualization and Computer Graphics 2019

So far this week we’ve seen how to create all sorts of fantastic

interactive visualisations, and taken a look at what data analysts

actually do when they do ‘exploratory data analysis.’ To round off the


week today’s choice is a recent paper on an age-old topic: what

visualisation should I use?

…the effectiveness of a visualization depends on

several factors including task at the hand, and

data attributes and datasets visualized.

Saket et al. look at five of the most basic visualisations —bar charts,

line charts, pie charts, scatterplots, and tables— and study their

effectiveness when presenting modest amounts of data (less than 50

visual marks) across 10 different tasks. The task taxonomy comes

from the work of Amar et al., describing a set of ten low-level analysis

tasks that describe users’ activities while using visualization tools.

1. Finding anomalies

2. Finding clusters (counting the number of groups with similar data

attribute values)
3. Finding correlations (determining whether or not there is a

correlation between two data attributes)


4. Computing derived values, for example, computing an aggregate

value
5. Characterising distributions, for example, figuring out which

percentage of data points have a value over a certain threshold


6. Finding extremes (i.e., min and max)

7. Filtering (finding data points that satisfy a condition)

8. Ordering (ranking data points according to some metric)


9. Determining a range (finding the span of values – pretty

straightforward if you can find the extremes – #6)


10. Retrieving a value (identifying the values of attributes for given

points).
DATA & VIEW SPECIFICATION:

These controls enable users to selectively visualize the data,


to filter out unrelated information to focus on relevant items,
and to sort information to expose patterns.

Data and View Specification involve determining which data is to be shown and visualized with
programs such as Microsoft Excel. Then it involves the filtering of data which shifts the focus
among the different data subsets to isolate specific categories of values. Sorting the data can show
surface trends and clusters and organize data according to a unit of analysis. The following image
shows a more complex form of a matrix-based visualization of a social network.

The first matrix plot shows


a social network when people are sorted alphabetically. The second plot shows a reordering by
node degrees resulting in more structure and the third plot is permutated by network connectivity,
showing underlying clusters of communities. The final step is to derive new attributes from
existing values when input data is insufficient.

View Manipulation :

The second dynamic is View Manipulation which consists of highlighting patterns, investigation of
hypotheses and revealing additional details. Selection allows for pointing to an item of interest, for
example, dragging along axes to "create interactive selections that highlight automobiles with low
weight and high mileage." Navigating is determined by where the analyst begins, such as in a crime
map that depicts crime activity by time and region. Coordinating allows the analyst to see multiple
coordinated views at once which can facilitate comparison. This can be done in histograms, maps
or network diagrams. The following image shows a complex patchwork of interlinked tables, plots
and maps to analyze outcomes of elections in Michigan.
The image shows a combination of tables, plots and maps. The final step, organization, involves
arranging visualization views, legends and controls for more simplified viewing.

Process and Provenance:

The final dynamic is Process and Provenance which involves the actual interpretation of data.
Recording involves chronicling and visualizing analysts' interaction histories in both a
chronological and sequential fashion. Annotation includes recording, organizing and
communicating insights gained during analysis. Sharing involves the accumulation of multiple
analyses and interpretations derived from several people and the dissemination of results. Guiding
is the final step and includes developing new strategies to guide newcomers.

Visualization by Data Type :

Visualizing data is crucial to ensure that all the data you collect translates into
decisions that amplify your business growth. Here are five data visualizations
that are commonly used by companies across the world.

Common Types of Data Visualizations

1. Bar Chart
2. Doughnut Chart or Pie Chart
3. Line Graph or Line Chart
4. Pivot Table
5. Scatter Plot

1. Bar Chart
Bar charts or column charts have rectangle bars arranged on the X or Y-axis.
Comparing objects by aligning them with the same parameters is the most
popular visualization out there. Bar charts can be used to track changes over
time. However, bar graphs used for time series yield accurate results when
the changes are considerably large. There are different categories of bar
graphs like stacked bar graphs, 100% stacked bar graphs, grouped bar
graphs, box plots, and waterfall charts (advanced bar graphs).

Bar charts are most suitable for:

 Comparing a numerical value across categories


 Identifying the order within a category
 Representing a histogram (where the values on the X-axis are grouped into
buckets)

Use Case: You can use a bar chart for a visual representation of your overall
business revenue against your peers. If you want to compare the individual
split of product-wise revenue, your best choice would be a stacked/grouped
bar chart.

2. Doughnut Chart or Pie Chart


A Doughnut chart slices a doughnut into multiple parts based on the field
value. Doughnut charts or pie charts are suitable to depict parts of a whole
relationship, where all units together represent 100%. Drilled-down doughnut
charts are interactive and help users decipher complex data to get to the
source of the issue or the solution.

Use Case: You can analyze your budget using a doughnut chart. Splitting the
total budget across your expenses, investments, loan, savings, etc. would
give you an instant understanding of your budget plan.

3. Line Graph or Line Chart


Line graphs have values plotted as lines across the X and Y-axis. They are
used to track changes over a short/long time frame. Line graphs are better to
use than bar graphs particularly when the changes are minor. You also have
the multi-line graph option when you need to compare changes over the same
time period for more than one attribute.

Line graphs are best used to:

 Display trend over a time series


 Pinpoint outliers
 Visualize forecasted data

Use Case: Say, you want to analyze your business’ month-wise expenditure.
The line graph will give you the best rendition by plotting the values across
months on the X-axis and expenditure on the Y-axis.

4. Pivot Table
Pivot table as the name indicates has columns and rows with aggregated
values populated in the cells. The pivot table is the most straight-forward
visualization that can be used to convey a huge amount of data at a single
glance. It is easy to build and flexible to modify. However, unlike the other
infographic visualizations discussed here, tables are not graphical and hence
can be used only in specific cases —

 When you want to compare different unrelated metrics required


 When there are relatively lesser rows (display of data at the top level as
opposed to the granular level)

Use Case: Financial reports are generally depicted over tables. Bringing in
the years on rows and operating cash flow, investing cash flow, cash from
financing, and other metrics on the columns will help you understand your
business’s cash flow over the years.

5. Scatter Plot
Scatter plot shows the relationship of the common attribute between two
numerical variables plotted along both X and Y axes. If you are a data
scientist working with different sets of data, scatter plot would be something
that you commonly work with, but for a novice user, it could be a little
unfamiliar. Scatter plots are best suitable to compare two numerical values
simultaneously. Segmentation charts and bubble charts are the advanced
versions of the scatter plot. The segmentation chart demarcates the scatter
plot into four quadrants, making the choice of the users easier. Bubble sort
brings in an additional dimension to the chart by displaying varied sizes of
bubbles over the scatter plot.

Use Case: You can present data for product price revision using the scatter
plot by bringing in the number of units sold on the X-axis, current price on the
Y-axis, and the products on the quadrant. It will give you a clear outlook on
the products that have a low price yet have sold a good deal and can be
considered for price increment. Alternatively, you can bring in a price drop for
products that have a high price but are in low demand.
These types of charts, along with area charts, heat maps, and treemaps are
widely used visualization techniques by data analysts, marketers, and
financial analysts across the world. However, there are specific visualizations
that can be used to tackle the reporting needs of unique data sets – for
instance, it would be ideal to use choropleth (Map chart), tree diagram, and
radar chart for geospatial, conditional, and multivariable data points
respectively. Choosing the perfect visualization from different types of data
visualizations can be challenging, but with a basic understanding of these
fundamental charts, your choice will be easier.

Challenges for Data Visualization:

Data visualization is a quick and simple technique to depict complicated ideas for improved
comprehension and intuition graphically. It must find diverse relationships and patterns
concealed by the massive amounts of data. We can use traditional graphical representations to
organize and represent data. Still, it can be challenging to display Huge amounts of data that
is very diverse uncertain, and semi-structured or unstructured in real-time. Massive
parallelization is required to handle and visualize such dynamic data. In this article, we will
cover the major challenges faced by big data visualization and its solutions.

Content

 Data Quality
 Accuracy
 Completeness
 Consistency
 Format
 Integrity
 Timeliness
 Not Choosing the Right Data Visualization Tools
 Confusing Color Palate
 Analytical & Technical Challenges
Data Quality
The data quality plays a crucial role in determining its effectiveness. Not all data is created in
the same way, and each has a different origin, hence its heterogeneity.

No matter how powerful and comprehensive the big data tools at the organization’s disposal
are, insufficient or incomplete data can often lead data scientists to conclusions that may not
be entirely correct and, therefore, could negatively impact business.

The effectiveness of big data in analysis depends on the accuracy, consistency, relevance,
completeness, and updating of the data used. With these factors, the data analysis ceases to be
reliable.

Accuracy

The most important thing is: How accurate is the data? How much can you trust it? Is there
certainty about the collection of relevant data? The values in each field of the database must
be correct and accurately represent “real world” values.

Example: A registered address must be a real address. Names must be spelled correctly.

Completeness

The data must contain all the necessary information and be easily understandable by the user.

Example: If the first and last name are required in a form, but the middle name is optional,
the form can be considered complete even if the middle name is not entered.

Consistency

The data must be the same throughout the organization and in all systems.

Example: The data of a sale registered in the CRM of a company must match the data
registered in the accounting dashboard that you manage.

Format

The data must meet certain standards of type, format, size, etc.

Example: All dates must follow the format DD/MM/YY, or whichever format you prefer.
Names should only have letters, no numbers or symbols.

Integrity
The data must be valid, which means that there are registered relationships that connect all the
data. Keep in mind that unlinked records could introduce duplicate entries to your system.

Example: If you have an address registered in your database, but it is not linked to any
individual, business, or other entity, it is invalid.

Timeliness

Data must be available when the user expects and needs it.

Example: A hospital must track the time of patient care events in the ICU or emergency room
to assist doctors and nurses.

Not Choosing the Right Data Visualization Tools


The selection of big data tools often focuses on the technical plane, leaving aside everything
that is not directly related to the analysis. Acting this way means ending up implementing
solutions whose visualization potential is narrower.

When this happens, the consequences do not take long to appear:

 Causing difficulties in understanding the data.


 Subtracting agility from the process of extracting and sharing knowledge within the
organization.
 Increasing latencies in taking action.
 Being able even to divert decision-making, which would lose effectiveness.

Many data visualization tools, such as Tableau, Microsoft Power BI, Looker, Sisense, Qlik,
etc., offer data visualization integration capabilities. If your organization already uses one of
these tools, start there. If not, try one. Once you select a tool, you’ll need to do a series of
prototypes to validate capabilities, ease of use, and operational considerations.

Here is a detailed list of considerations:

 Do the chart types meet the business needs


 How easy is it to integrate?
 What are the flexibilities in device design and compatibility?
 Is the security configurable for the required end user/consumer rights?
 Is it performing fast enough to integrate into an application?
 Are the platform’s costs and pricing models aligned?

You might also like