Discover millions of ebooks, audiobooks, and so much more with a free trial

Only $9.99/month after trial. Cancel anytime.

PROC SQL: Beyond the Basics Using SAS, Third Edition
PROC SQL: Beyond the Basics Using SAS, Third Edition
PROC SQL: Beyond the Basics Using SAS, Third Edition
Ebook747 pages7 hours

PROC SQL: Beyond the Basics Using SAS, Third Edition

Rating: 0 out of 5 stars

()

Read preview

About this ebook

PROC SQL: Beyond the Basics Using SAS®, Third Edition, is a step-by-step, example-driven guide that helps readers master the language of PROC SQL. Packed with analysis and examples illustrating an assortment of PROC SQL options, statements, and clauses, this book not only covers all the basics, but it also offers extensive guidance on complex topics such as set operators and correlated subqueries. Programmers at all levels will appreciate Kirk Lafler’s easy-to-follow examples, clear explanations, and handy tips to extend their knowledge of PROC SQL.

This third edition explores new and powerful features in SAS® 9.4, including topics such as:

  • IFC and IFN functions
  • nearest neighbor processing
  • the HAVING clause
  • indexes

It also features two completely new chapters on fuzzy matching and data-driven programming. Delving into the workings of PROC SQL with greater analysis and discussion, PROC SQL: Beyond the Basics Using SAS®, Third Edition, explores this powerful database language using discussion and numerous real-world examples.

LanguageEnglish
PublisherSAS Institute
Release dateMar 20, 2019
ISBN9781635266818
PROC SQL: Beyond the Basics Using SAS, Third Edition
Author

Kirk Paul Lafler

Kirk Paul Lafler is founder and entrepreneur at Software Intelligence Corporation. He has worked with SAS software since 1979 as a consultant, application designer and developer, programmer, SAS solutions provider, data analyst, educator, and author. As a SAS Certified professional, mentor, and educator at Software Intelligence Corporation and as an advisor and SAS programming adjunct professor at the University of California San Diego Extension, Lafler has taught SAS courses, seminars, workshops, and webinars to thousands of users around the world. He received his BS and MS degrees from the University of Miami. Lafler is a frequent speaker at international, regional, special-interest, local, and in-house SAS users group conferences and meetings. He has also authored or co-authored several books, including Google® Search Complete and PROC SQL: Beyond the Basics Using SAS®; hundreds of papers and articles on a variety of SAS topics; and a popular SAS tips column, called “Kirk's Korner.” He has also served as an invited speaker, instructor, keynote speaker, and section leader at SAS users group conferences and meetings around the world and is the recipient of numerous “Best” contributed paper, hands-on workshop (HOW), and poster awards.

Related to PROC SQL

Related ebooks

Enterprise Applications For You

View More

Related articles

Reviews for PROC SQL

Rating: 0 out of 5 stars
0 ratings

0 ratings0 reviews

What did you think?

Tap to rate

Review must be at least 10 words

    Book preview

    PROC SQL - Kirk Paul Lafler

    Chapter 1: Designing Database Tables

    Introduction

    Database Design

    Conceptual View

    Table Definitions

    Redundant Information

    Normalization

    Normalization Strategies

    Column Names and Reserved Words

    ANSI SQL Reserved Words

    SQL Code

    Data Integrity

    Referential Integrity

    Database Tables Used in This Book

    CUSTOMERS Table

    INVENTORY Table

    INVOICE Table

    MANUFACTURERS Table

    PRODUCTS Table

    PURCHASES Table

    Table Contents

    The Database Structure

    Sample Database Tables

    Summary

    Introduction

    The area of database design is very important in relational processes. Much has been written on this subject, including entire textbooks and thousands of technical papers. No pretenses are made about the thoroughness of this very important subject in these pages. Rather, an attempt is made to provide a quick-start introduction for those readers who are unfamiliar with the issues and techniques of basic design principles. Readers needing more information are referred to the references listed in the back of this book. As you read this chapter, the following points should be kept in mind.

    Database Design

    Activities related to good database design require the identification of end-user requirements and involve defining the structure of data values on a physical level. Database design begins with a conceptual view of what is needed. The next step, called logical design, consists of developing a formal description of database entities and relationships to satisfy user requirements. Seldom does a database consist of a single table. Consequently, tables of interrelated information are created to enable more complex and powerful operations on data. The final step, referred to as physical design, represents the process of achieving optimal performance and storage requirements of the logical database.

    Conceptual View

    The health and well-being of a database depends on its database design. A database must be in balance with all of its components (or optimized) to avoid performance and operation bottlenecks. Database design doesn’t just happen and is not a process that occurs by chance. It involves planning, modeling, creating, monitoring, and adjusting to satisfy the endless assortment of user requirements without impeding resource requirements. Of central importance to database design is the process of planning. Planning is a valuable component that, when absent, causes a database to fall prey to a host of problems including poor performance and difficulty in operation. Database design consists of three distinct phases, as illustrated in Figure 1.1.

    Figure 1.1: Three Distinct Phases of Database Design

    Figure 1.1: Three Distinct Phases of Database Design

    Table Definitions

    PROC SQL uses a model of data that is conceptually stored as multisets rather than as physical files. A physical file consists of one or more records ordered sequentially or some other way. Programming languages such as COBOL and FORTRAN evolved to process files of this type by performing operations one record at a time. These languages were generally designed and used to mimic the way people process paper forms.

    PROC SQL was designed to work with multisets of data. Multisets have no order, and members of a multiset are of the same type using a data structure known as a table. For classification purposes, a table is a base table consisting of zero or more rows and one or more columns, or a table is a virtual table (called a view), which can be used the same way that a table can be used (see Chapter 8, Working with Views).

    Redundant Information

    One of the rules of good database design requires that data not be redundant or duplicated in the same database. The rationale for this conclusion originates from the belief that if data appears more than once in a database, then there is reason to believe that one of the pieces of data is likely to be in error. Furthermore, redundancy often leads to the following:

    ●          Inconsistencies, because errors are more likely to result when facts are repeated.

    ●          Update anomalies where the insertion, modification, or deletion of data may result in inconsistencies.

    Another thing to watch for is the appearance of too many columns containing NULL values. When this occurs, the database is probably not designed properly. To alleviate potential table design issues, a process referred to as normalizing is performed. When properly done, this ensures the complete absence of redundant information in a table.

    Normalization

    The development of an optimal database design is an important element in the life cycle of a database. Not only is it critical for achieving maximum performance and flexibility while working with tables and data, it is essential to the organization of data by reducing or minimizing redundancy in one or more database tables. The process of table design is frequently referred to by database developers and administrators as normalization.

    The normalization process is used for reducing redundancy in a database by converting complex data structures into simple data structures. It is carried out for the following reasons:

    ●          To organize the data to save space and to eliminate any duplication or repetition of data.

    ●          To enable simple retrieval of data to satisfy query and report requests.

    ●          To simplify data manipulation requests such as data insertions, updates, and deletions.

    ●          To reduce the impact associated with reorganizing or restructuring data as new application requirements arise.

    The normalization process attempts to simplify the relationship between columns in a database by splitting larger multicolumn tables into two or more smaller tables containing fewer columns. The rationale for doing this is contained in a set of data design guidelines called normal forms. The guidelines provide designers with a set of rules for converting one or two large database tables containing numerous columns into a normalized database consisting of multiple tables and only those columns that should be included in each table. The normalization process consists of multiple steps with each succeeding step subscribing to the rules of the previous steps.

    Normalization helps to ensure that a database does not contain redundant information in two or more of its tables. In an application, normalization prevents the destruction of data or the creation of incorrect data in a database. What this means is that information of fact is represented only once in a database, and any possibility of it appearing more than once is not, or should not be, allowed.

    As database designers and analysts proceed through the normalization process, many are not satisfied unless a database design is carried out to at least third normal form (3NF). Joe Celko in his popular book SQL for Smarties: Advanced SQL Programming (Morgan Kaufman, 2014), describes 3NF this way: Databases are considered to be in 3NF when a column is dependent on the key, the whole key, and nothing but the key.

    While the normalization guidelines are extremely useful, some database purists actually go to great lengths to remove any and all table redundancies even at the expense of performance. This is in direct contrast to other database experts who follow the guidelines less rigidly in an attempt to improve the performance of a database by only going as far as third normal form (or 3NF). Whatever your preference, you should keep this thought in mind as you normalize database tables. A fully normalized database often requires a greater number of joins and can adversely affect the speed of queries. Celko mentions that the process of joining multiple tables in a fully normalized database is costly, specifically affecting processing time and computer resources.

    Normalization Strategies

    After transforming entities and attributes from the conceptual design into a logical design, the tables and columns are created. This is when a process known as normalization occurs. Normalization refers to the process of making your database tables subscribe to certain rules. Many, if not most, database designers are satisfied when third normal form (3NF) is achieved and, for the objectives of this book, I will stop at 3NF, too. To help explain the various normalization steps, an example scenario follows.

    First Normal Form (1NF)

    First normal form (1NF) involves the elimination of data redundancy or repeating information from a table. A table is considered to be in first normal form when all of its columns describe the table completely and when each column in a row has only one value. A table satisfies 1NF when each column in a row has a single value and no repeating group information. Essentially, every table meets 1NF as long as an array, list, or other structure has not been defined. The following table illustrates a table satisfying the 1NF rule because it has only one value at each row‑and‑column intersection. The table is in ascending order by CUSTNUM and consists of customers and the purchases they made at an office supply store.

    Table 1.1: First Normal Form (1NF) Table

    CUSTNUM  CUSTNAME CUSTCITY       ITEM     UNITS  UNITCOST  MANUCITY

          1  Smith     San Diego      Chair       1   $179.00  San Diego

          1  Smith     San Diego      Pens       12     $0.89  Los Angeles

          1  Smith     San Diego      Paper       4     $6.95  Washington

          1  Smithe    San Diego      Stapler     1     $8.95  Los Angeles

          7  Lafler    Spring Valley  Mouse Pad   1    $11.79  San Diego

          7  Loffler   Spring Valley  Pens       24     $1.59  Los Angeles

         13  Thompson  Miami          Markers     .     $0.99  Los Angeles

    Second Normal Form (2NF)

    Second normal form (2NF) addresses the relationships between sets of data. A table is said to be in second normal form when all the requirements of 1NF are met and a foreign key is used to link any data in one table which has relevance to another table. The very nature of leaving a table in first normal form (1NF) may present problems due to the repetition of some information in the table. One noticeable problem is that Table 1.1 has repetitive information in it. Another problem is that there are misspellings in the customer name. Although repeating information may be permissible with hierarchical file structures and other legacy type file structures, it does pose a potential data consistency problem as it relates to relational data.

    To describe how data consistency problems can occur, let’s say that a customer takes a new job and moves to a new city. In changing the customer’s city to the new location, it would be very easy to miss one or more occurrences of the customer’s city resulting in a customer residing incorrectly in two different cities. Assuming that our table is only meant to track one unique customer per city, this would definitely be a data consistency problem. Essentially, second normal form (2NF) is important because it says that every non-key column must depend on the entire primary key.

    Tables that subscribe to 2NF prevent the need to make changes in more than one place. What this means in normalization terms is that tables in 2NF have no partial key dependencies. As a result, our database that consists of a single table that satisfies 1NF will need to be split into two separate tables in order to subscribe to the 2NF rule. Each table would contain the CUSTNUM column to connect the two tables. Unlike the single table in 1NF, the tables in 2NF allow a customer’s city to be easily changed whenever they move to another city because the CUSTCITY column only appears once. The tables in 2NF would be constructed as follows.

    Table 1.2: CUSTOMERS Table

    CUSTNUM   CUSTNAME    CUSTCITY

          1   Smith       San Diego

          1   Smithe      San Diego

          7   Lafler      Spring Valley

         13   Thompson    Miami

    Table 1.3: PURCHASES Table

    CUSTNUM   ITEM     UNITS     UNITCOST   MANUCITY

          1   Chair        1      $179.00   San Diego

          1   Pens        12        $0.89   Los Angeles

          1   Paper        4        $6.95   Washington

          1   Stapler      1        $8.95   Los Angeles

          7   Mouse Pad    1       $11.79   San Diego

          7   Pens        24        $1.59   Los Angeles

         13   Markers      .        $0.99   Los Angeles

    Third Normal Form (3NF)

    Referring to the two tables constructed according to the rules of 2NF, you may have noticed that the PURCHASES table contains a column called MANUCITY. The MANUCITY column stores the city where the product manufacturer is headquartered. Keeping this column in the PURCHASES table violates the third normal form (3NF) because MANUCITY does not provide factual information about the primary key column (CUSTNUM) in the PURCHASES table. Consequently, tables are considered to be in third normal form (3NF) when each column is dependent on the key, the whole key, and nothing but the key. The tables in 3NF are constructed so the MANUCITY column would be in a table of its own as follows.

    Table 1.4: CUSTOMERS Table

    CUSTNUM   CUSTNAME    CUSTCITY

          1   Smith       San Diego

          1   Smithe      San Diego

          7   Lafler      Spring Valley

         13   Thompson    Miami

    Table 1.5: PURCHASES Table

    CUSTNUM   ITEM     UNITS     UNITCOST

          1   Chair        1      $179.00

          1   Pens        12        $0.89

          1   Paper        4        $6.95

          1   Stapler      1        $8.95

          7   Mouse Pad    1       $11.79

          7   Pens        24        $1.59

         13   Markers      .        $0.99

    Table 1.6: MANUFACTURERS Table

    MANUNUM   MANUCITY      

        101   San Diego

        112   San Diego

        210   Los Angeles

        212   Los Angeles

        213   Los Angeles

        214   Los Angeles

        401   Washington

    Beyond Third Normal Form

    In general, database designers are satisfied when their database tables subscribe to the rules in 3NF. But, it is not uncommon for others to normalize their database tables to fourth normal form (4NF) where independent one-to-many relationships between primary key and non-key columns are forbidden. Some database purists will even normalize to fifth normal form (5NF) where tables are split into the smallest pieces of information in an attempt to eliminate any and all table redundancies. Although constructing tables in 5NF may provide the greatest level of database integrity, it is neither practical nor desired by most database practitioners.

    There is no absolute right or wrong reason for database designers to normalize beyond 3NF as long as they have considered all the performance issues that may arise by doing so. A common problem that occurs when database tables are normalized beyond 3NF is that a large number of small tables are generated. In these cases, an increase in time and computer resources frequently occurs because small tables must first be joined before a query, report, or statistic can be produced.

    Column Names and Reserved Words

    According to the American National Standards Institute (ANSI), SQL is the standard language used with relational database management systems. The ANSI Standard reserves a number of SQL keywords from being used as column names. The SAS SQL implementation is not as rigid, but users should be aware of what reserved words exist to prevent unexpected and unintended results during SQL processing. Column names should conform to proper SAS naming conventions (as described in the SAS Language Reference), and they should not conflict with certain reserved words found in the SQL language. The following list identifies the reserved words found in the ANSI SQL standard.

    ANSI SQL Reserved Words

    You probably will not encounter too many conflicts between a column name and an SQL reserved word, but when you do you will need to follow a few simple rules to prevent processing errors from occurring. As was stated earlier, although PROC SQL’s naming conventions are not as rigid as other vendor’s implementations, care should still be exercised, in particular when PROC SQL code is transferred to other database environments expecting it to run error free. If a column name in an existing table conflicts with a reserved word, you have three options at your disposal:

    1.       Physically rename the column name in the table, as well as any references to the column.

    2.       Use the RENAME= data set option to rename the desired column in the current query.

    3.       Specify the PROC SQL option DQUOTE=ANSI, and surround the column name (reserved word) in double quotes, as illustrated below.

    SQL Code

    PROC SQL DQUOTE=ANSI;

     SELECT *

      FROM RESERVED_WORDS

    WHERE "WHERE"=’EXAMPLE’;

    QUIT;

    Data Integrity

    Webster’s New World Dictionary defines integrity as the quality or state of being complete; perfect condition; reliable; soundness. Data integrity is a critical element that every organization must promote and strive for. It is imperative that the data tables in a database environment be reliable, free of errors, and sound in every conceivable way. The existence of data errors, missing information, broken links, and other related problems in one or more tables can impact decision-making and information reporting activities resulting in a loss of confidence among users.

    Applying a set of rules to the database structure and content can ensure the integrity of data resources. These rules consist of table and column constraints, and will be discussed in detail in Chapter 5, Creating, Populating, and Deleting Tables.

    Referential Integrity

    Referential integrity refers to the way in which database tables handle update and delete requests. Database tables frequently have a primary key where one or more columns have a unique value by which rows in a table can be identified and selected. Other tables may have one or more columns called a foreign key that are used to connect to some other table through its value. Database designers frequently apply rules to database tables to control what happens when a primary key value changes and its effect on one or more foreign key values in other tables. These referential integrity rules apply restrictions on the data that may be updated or deleted in tables.

    Referential integrity ensures that rows in one table have corresponding rows in another table. This prevents lost linkages between data elements in one table and those of another enabling the integrity of data to always be maintained. Using the 3NF tables defined earlier, a foreign key called CUSTNUM can be defined in the PURCHASES table that corresponds to the primary key CUSTNUM column in the CUSTOMERS table. Users are referred to Chapter 5, Creating, Populating, and Deleting Tables for more details on assigning referential integrity constraints.

    Database Tables Used in This Book

    This section describes a database or library of tables that is used by an imaginary computer hardware and software wholesaler. The library consists of six tables: Customers, Inventory, Invoice, Manufacturers, Products, and Purchases. The examples used throughout this book are based on this library (database) of tables and are described and displayed below. An alphabetical description of each table used throughout this book appears below.

    CUSTOMERS Table

    The CUSTOMERS table contains customers that have purchased computer hardware and software products from a manufacturer. Each customer is uniquely identified with a customer number. A description of each column in the Customers table follows.

    Table 1.7: Description of Columns in the Customers Table

    INVENTORY Table

    The INVENTORY table contains customer inventory information consisting of computer hardware and software products. The Inventory table contains no historical data. As inventories are replenished, the old quantity is overwritten with the new quantity. A description of each column in the Inventory table follows.

    Table 1.8: Description of Columns in the Inventory Table

    INVOICE Table

    The INVOICE table contains information about customers who purchased products. Each invoice is uniquely identified with an invoice number. A description of each column in the Invoice table follows.

    Table 1.9: Description of Columns in the Invoice Table

    MANUFACTURERS Table

    The MANUFACTURERS table contains companies who make computer hardware and software products. Two companies cannot have the same name. No historical data is kept in this table. If a company is sold or stops making computer hardware or software, then the manufacturer is dropped from the table. In the event that a manufacturer has an address change, the old address is overwritten with the new address. A description of each column in the Manufacturers table follows.

    Table 1.10: Description of Columns in the Manufacturers Table

    PRODUCTS Table

    The PRODUCTS table contains computer hardware and software products offered for sale by the manufacturer. Each product is uniquely identified with a product number. A description of each column in the Products table follows.

    Table 1.11: Description of Columns in the Products Table

    PURCHASES Table

    The PURCHASES table contains computer hardware and software products purchased by customers. Each product is uniquely identified with a product number. A description of each column in the Purchases table follows.

    Table 1.12: Description of Columns in the Purchases Table

    Table Contents

    An alphabetical list of tables, variables, and attributes for each table is displayed below.

    Output 1.1: Customers CONTENTS Output

    Output 1.1: Customers CONTENTS Output

    Output 1.2: Inventory CONTENTS Output

    Output 1.2: Inventory CONTENTS Output

    Output 1.3: Invoice CONTENTS Output

    Output 1.3: Invoice CONTENTS Output

    Output 1.4: Manufacturers CONTENTS Output

    Output 1.4: Manufacturers CONTENTS Output

    Output 1.5: Products CONTENTS Output

    Output 1.5: Products CONTENTS Output

    Output 1.6: Purchases CONTENTS Output

    Output 1.6: Purchases CONTENTS Output

    The Database Structure

    The logical relationship between each table, and the columns common to each, appear below.

    Figure 1.2. Logical Database Structure

    Figure 1.2. Logical Database Structure

    Sample Database Tables

    The following tables: Customers, Inventory, Manufacturers, Products, Invoice, and Purchases represent a relational database that will be illustrated in the examples in this book. These tables are small enough to follow easily, but complex enough to illustrate the power of SQL. The data contained in each table appears below.

    Table 1.13: CUSTOMERS Table

    Table 1.13: CUSTOMERS Table

    Table 1.14: INVENTORY Table

    Table 1.14: INVENTORY Table

    Table 1.15: INVOICE Table

    Table 1.15: INVOICE Table

    Table 1.16: MANUFACTURERS Table

    Table 1.16: MANUFACTURERS Table

    Table 1.17: PRODUCTS Table

    Table 1.17: PRODUCTS Table

    Table 1.18: PURCHASES Table

    Table 1.18: PURCHASES Tableimage shown hereimage shown here

    Summary

    1.       Good database design often improves the relative ease by which tables can be created and populated in a relational database and can be implemented into any database (see the Conceptual View section).

    2.       SQL was designed to work with sets of data and accesses a data structure known as a table or a virtual table, known as a view (see the Table Definitions section).

    3.       Achieving optimal design of a database means that the database contains little or no redundant information in two or more of its tables. This means that good database design calls for little or no replication of data (see the Redundant Information section).

    4.       Good database design avoids data redundancy, update anomalies, costly or inefficient processing, coding complexities, complex logical relationships, long application development times, and/or excessive storage requirements (see the Normalization section).

    5.       Design decisions made in one phase may involve making one or more tradeoffs in another phase (see the Normalization section).

    6.       A database in third normal form (3NF) is defined as a column that is dependent on the key, the whole key, and nothing but the key (see the Normalization section).

    Chapter 2: Working with Data in PROC SQL

    Introduction

    The SELECT Statement and Clauses

    Overview of Data Types

    Numeric Data

    Date and Time Column Definitions

    Character Data

    Missing Values and NULL

    Arithmetic and Missing Data

    SQL Keywords

    SQL Operators, Functions, and Keywords

    Comparison Operators

    Logical Operators

    Arithmetic Operators

    Character String Operators and Functions

    Summarizing Data

    Predicates

    CALCULATED Keyword

    Dictionary Tables

    Dictionary Tables and Metadata

    Displaying Dictionary Table Definitions

    Dictionary Table Column Names

    Accessing a Dictionary Table’s Contents

    Summary

    Introduction

    PROC SQL is essentially a database language as opposed to a procedural or computational language. This chapter’s focus is on working with data in PROC SQL using the SELECT statement. Often referred to as an SQL query, the SELECT statement is the most versatile statement in SQL and is used to read data from one or more database tables (or data sets). It also supports numerous extensions including keywords, operators, functions, and predicates, and returns the data in a table-like structure called a result-set.

    The SELECT Statement and Clauses

    The SELECT statement’s purpose is to retrieve (or read) data from the underlying tables (or views). Although it supports multiple clauses, the SELECT statement has only one clause that is required to be specified – the FROM clause. All the remaining clauses, described below, are optional and only used when needed. Note: Not every query needs to have all the clauses specified, but SQL provides developers and data analysts with a powerful and flexible language to access, manipulate, and display data without the need to write large amounts of code.

    During execution, SAS carries out the tasks associated with planning, optimizing, and performing the operations specified in the SELECT statement and its clauses to produce the desired results. To prevent syntax errors from occurring when using the SELECT statement, the clauses must be specified in the correct order. To help you remember the order of the SELECT statement’s clauses recite, SQL is fun when geeks help others. The first letter in each word corresponds to the name of the SELECT statement’s clause as shown in Figure 2.1.

    Figure 2.1: Order of the SELECT Statement Clauses

    Figure 2.1: Order of the SELECT Statement Clauses

    When constructed correctly, the SELECT statement and its clauses declares the database table (or data set) to find the data in, what data to retrieve, and whether any special transformations or processing is needed before the data is returned. The next example shows the correct syntax of a query’s SELECT statement and its clauses.

    SQL Code

    PROC SQL;

    SELECT PRODNAME

            ,PRODTYPE

            ,PRODCOST

    INTO :M_PRODNAME

            ,:M_PRODTYPE

            ,:M_PRODCOST

    FROM PRODUCTS

    WHERE PRODNAME CONTAINS Software

    GROUP BY PRODTYPE

    HAVING COUNT(PRODTYPE) > 3

    ORDER BY PRODNAME;

    QUIT;

    Results

    image shown here

    Now that we’ve explored the order that each clause is specified in an SQL query, let’s examine the order of execution of each clause in an SQL query. Table 2.1 illustrates and describes the execution order of each SELECT statement clause.

    Table 2.1: Clause Execution Order

    Enjoying the preview?
    Page 1 of 1