talk-data.com talk-data.com

Topic

data

5765

tagged

Activity Trend

3 peak/qtr
2020-Q1 2026-Q1

Activities

5765 activities · Newest first

Mastering the SAS DS2 Procedure

Enhance your SAS data-wrangling skills with high-precision and parallel data manipulation using the DS2 programming language. Now in its second edition, this book addresses the DS2 programming language from SAS, which combines the precise procedural power and control of the Base SAS DATA step language with the simplicity and flexibility of SQL. DS2 provides simple, safe syntax for performing complex data transformations in parallel and enables manipulation of native database data types at full precision. It also covers PROC FEDSQL, a modernized SQL language that blends perfectly with DS2. You will learn to harness the power of parallel processing to speed up CPU-intensive computing processes in Base SAS and how to achieve even more speed by processing DS2 programs on massively parallel database systems. Techniques for leveraging internet APIs to acquire data, avoiding large data movements when working with data from disparate sources, and leveraging DS2's new data types for full-precision numeric calculations are presented, with examples of why these techniques are essential for the modern data wrangler. Here's what's new in this edition: how to significantly improve performance by using the new SAS Viya architecture with its SAS Cloud Analytic Services (CAS) how to declare private variables and methods in a package the new PROC DSTODS2 the PCRXFIND and PCRXREPLACE packages While working though the code samples provided with this book, you will build a library of custom, reusable, and easily shareable DS2 program modules, execute parallelized DATA step programs to speed up a CPU-intensive process, and conduct advanced data transformations using hash objects and matrix math operations. This book is part of the SAS Press Series.

IBM Spectrum Archive Single Drive Edition and Library Edition: Installation and Configuration Guide

Abstract The IBM® Linear Tape File System™ (LTFS) is the first file system that works along with Linear Tape-Open (LTO) tape technology to set a new standard for ease of use and portability for open systems tape storage. In 2011, LTFS won an Engineering Emmy Award for Innovation from the Academy of Television Arts & Sciences. This IBM Redbooks® publication helps you install, tailor, and configure the IBM Spectrum™ Archive Single Drive Edition (SDE) and the IBM Spectrum Archive™ Library Edition (LE) products. LTFS is a file system that was originally implemented on dual-partition linear tape (IBM LTO Ultrium 5 tape drives (LTO-5) and IBM TS1140 tape drives). Now IBM Spectrum Archive SDE and LE support IBM LTO Ultrium 8, 7, 6, or 5 tape drives, and IBM TS1155, IBM TS1150, and IBM TS1140 tape drives. IBM Spectrum Archive LE supports the IBM TS4500 tape library, IBM TS3500 tape library, IBM TS3310 tape library, IBM TS3200 tape library express, IBM TS3100 tape library express, and IBM TS2900 tape autoloader express. IBM Spectrum Archive makes tape look and work like any removable media, such as a USB drive. Files and directories appear on the desktop as a directory listing. It is now simple to drag files to and from tape. Any application that is written to use disk files works with the same files on tape. IBM Spectrum Archive SDE supports stand-alone drives only. IBM Spectrum Archive LE supports tape libraries. IBM Spectrum Archive LE presents each cartridge in the library as a subdirectory in the LTFS file system. With IBM Spectrum Archive LE, you can list the contents and search all of the volumes in the library without mounting the volumes by using an in-memory index. This publication is intended for anyone who wants to understand more about IBM Linear Tape System products and their implementation. This book is suitable for IBM clients, IBM Business Partners, IBM specialist sales representatives, and technical specialists.

IBM TS4500 R4 Tape Library Guide

Abstract The IBM® TS4500 (TS4500) tape library is a next-generation tape solution that offers higher storage density and integrated management than previous solutions. This IBM Redbooks® publication gives you a close-up view of the new IBM TS4500 tape library. In the TS4500, IBM delivers the density that today's and tomorrow's data growth requires. It has the cost-effectiveness and the manageability to grow with business data needs, while you preserve existing investments in IBM tape library products. Now, you can achieve both a low cost per terabyte (TB) and a high TB density per square foot, because the TS4500 can store up to 8.25 petabytes (PB) of uncompressed data in a single frame library or scale up at 1.5 PB per square foot to over 263 PB, which is more than 4 times the capacity of the IBM TS3500 tape library. The TS4500 offers these benefits: High availability dual active accessors with integrated service bays to reduce inactive service space by 40%. The Elastic Capacity option can be used to completely eliminate inactive service space. Flexibility to grow: The TS4500 library can grow from both the right side and the left side of the first L frame because models can be placed in any active position. Increased capacity: The TS4500 can grow from a single L frame up to an additional 17 expansion frames with a capacity of over 23,000 cartridges. High-density (HD) generation 1 frames from the existing TS3500 library can be redeployed in a TS4500. Capacity on demand (CoD): CoD is supported through entry-level, intermediate, and base-capacity configurations. Advanced Library Management System (ALMS): ALMS supports dynamic storage management, which enables users to create and change logical libraries and configure any drive for any logical library. Support for the IBM TS1155 while also supporting TS1150 and TS1140 tape drive: The TS1155 gives organizations an easy way to deliver fast access to data, improve security, and provide long-term retention, all at a lower cost than disk solutions. The TS1155 offers high-performance, flexible data storage with support for data encryption. Also, this enhanced fifth-generation drive can help protect investments in tape automation by offering compatibility with existing automation. The new TS1155 Tape Drive Model 55E delivers a 10 Gb Ethernet host attachment interface optimized for cloud-based and hyperscale environments. The TS1155 Tape Drive Model 55F delivers a native data rate of 360 MBps, the same load/ready, locate speeds, and access times as the TS1150, and includes dual-port 8 Gb Fibre Channel support. Support of the IBM Linear Tape-Open (LTO) Ultrium 8 tape drive: The LTO Ultrium 8 offering represents significant improvements in capacity, performance, and reliability over the previous generation, LTO Ultrium 7, while they still protect your investment in the previous technology. Support of LTO 8 Type M cartridge (M8): The LTO Program is introducing a new capability with LTO-8 drives. The ability of the LTO-8 drive to write 9 TB on a brand new LTO-7 cartridge instead of 6 TB as specified by the LTO-7 format. Such a cartridge is called an LTO-7 initialized LTO-8 Type M cartridge. Integrated TS7700 back-end Fibre Channel (FC) switches are available. Up to four library-managed encryption (LME) key paths per logical library are available. This book describes the TS4500 components, feature codes, specifications, supported tape drives, encryption, new integrated management console (IMC), and command-line interface (CLI). You learn how to accomplish several specific tasks: Improve storage density with increased expansion frame capacity up to 2.4 times and support 33% more tape drives per frame. Manage storage by using the ALMS feature. Improve business continuity and disaster recovery with dual active accessor, automatic control path failover, and data path failover. Help ensure security and regulatory compliance with tape-drive encryption and Write Once Read Many (WORM) media. Support IBM LTO Ultrium 8, 7, 6, and 5, IBM TS1155, TS1150, and TS1140 tape drives. Provide a flexible upgrade path for users who want to expand their tape storage as their needs grow. Reduce the storage footprint and simplify cabling with 10 U of rack space on top of the library. This guide is for anyone who wants to understand more about the IBM TS4500 tape library. It is particularly suitable for IBM clients, IBM Business Partners, IBM specialist sales representatives, and technical specialists.

Implementing the IBM Storwize V5000 Gen2 (including the Storwize V5010, V5020, and V5030) with IBM Spectrum Virtualize V8.1

Abstract Organizations of all sizes face the challenge of managing massive volumes of increasingly valuable data. But storing this data can be costly, and extracting value from the data is becoming more difficult. IT organizations have limited resources but must stay responsive to dynamic environments and act quickly to consolidate, simplify, and optimize their IT infrastructures. The IBM® Storwize® V5000 Gen2 system provides a smarter solution that is affordable, easy to use, and self-optimizing, which enables organizations to overcome these storage challenges. The Storwize V5000 Gen2 delivers efficient, entry-level configurations that are designed to meet the needs of small and midsize businesses. Designed to provide organizations with the ability to consolidate and share data at an affordable price, the Storwize V5000 Gen2 offers advanced software capabilities that are found in more expensive systems. This IBM Redbooks® publication is intended for pre-sales and post-sales technical support professionals and storage administrators. It applies to the Storwize V5030, V5020, and V5010, and to IBM Spectrum Virtualize™ V8.1.

Agent-based Modeling of Tax Evasion

The only single-source guide to understanding, using, adapting, and designing state-of-the-art agent-based modelling of tax evasion A computational method for simulating the behavior of individuals or groups and their effects on an entire system, agent-based modeling has proven itself to be a powerful new tool for detecting tax fraud. While interdisciplinary groups and individuals working in the tax domain have published numerous articles in diverse peer-reviewed journals and have presented their findings at international conferences, until Agent-based Modelling of Tax Evasion there was no authoritative, single-source guide to state-of-the-art agent-based tax evasion modeling techniques and technologies. Featuring contributions from distinguished experts in the field from around the globe, Agent-Based Modelling of Tax Evasion provides in-depth coverage of an array of field tested agent-based tax evasion models. Models are presented in a unified format so as to enable readers to systematically work their way through the various modeling alternatives available to them. Three main components of each agent-based model are explored in accordance with the Overview, Design Concepts, and Details (ODD) protocol, each section of which contains several sub elements that help to illustrate the model clearly and that assist readers in replicating the modeling results described. Presents models in a unified and structured manner to provide a point of reference for readers interested in agent-based modelling of tax evasion Explores the theoretical aspects and diversity of agent-based modeling through the example of tax evasion Provides an overview of the characteristics of more than thirty agent-based tax evasion frameworks Functions as a solid foundation for lectures and seminars on agent-based modelling of tax evasion The only comprehensive treatment of agent-based tax evasion models and their applications, this book is an indispensable working resource for practitioners and tax evasion modelers both in the agent-based computational domain and using other methodologies. It is also an excellent pedagogical resource for teaching tax evasion modeling and/or agent-based modeling generally.

Beginning PostgreSQL on the Cloud: Simplifying Database as a Service on Cloud Platforms

Get started with PostgreSQL on the cloud and discover the advantages, disadvantages, and limitations of the cloud services from Amazon, Rackspace, Google, and Azure. Once you have chosen your cloud service, you will focus on securing it and developing a back-up strategy for your PostgreSQL instance as part of your long-term plan. Beginning PostgreSQL on the Cloud covers other essential topics such as setting up replication and high availability; encrypting your saved cloud data; creating a connection pooler for your database; and monitoring PostgreSQL on the cloud. The book concludes by showing you how to install and configure some of the tools that will help you get started with PostgreSQL on the cloud. This book shows you how database as a service enables you to spread your data across multiple data centers, ensuring that it is always accessible. You’ll discover that this model does not expect you to install and maintain databases yourself because the database service provider does it for you. You no longer have to worry about the scalability and high availability of your database. What You Will Learn Migrate PostgreSQL to the cloud Choose the best configuration and specifications of cloud instances Set up a backup strategy that enables point-in-time recovery Use connection pooling and load balancing on cloud environments Monitor database environments on the cloud Who This Book Is For Those who are looking to migrate to PostgreSQL on the Cloud. It will also help database administrators in setting up a cloud environment in an optimized way and help them with their day-to-day tasks.

Mathematical Statistics

Explores mathematical statistics in its entirety—from the fundamentals to modern methods This book introduces readers to point estimation, confidence intervals, and statistical tests. Based on the general theory of linear models, it provides an in-depth overview of the following: analysis of variance (ANOVA) for models with fixed, random, and mixed effects; regression analysis is also first presented for linear models with fixed, random, and mixed effects before being expanded to nonlinear models; statistical multi-decision problems like statistical selection procedures (Bechhofer and Gupta) and sequential tests; and design of experiments from a mathematical-statistical point of view. Most analysis methods have been supplemented by formulae for minimal sample sizes. The chapters also contain exercises with hints for solutions. Translated from the successful German text, Mathematical Statistics requires knowledge of probability theory (combinatorics, probability distributions, functions and sequences of random variables), which is typically taught in the earlier semesters of scientific and mathematical study courses. It teaches readers all about statistical analysis and covers the design of experiments. The book also describes optimal allocation in the chapters on regression analysis. Additionally, it features a chapter devoted solely to experimental designs. Classroom-tested with exercises included Practice-oriented (taken from day-to-day statistical work of the authors) Includes further studies including design of experiments and sample sizing Presents and uses IBM SPSS Statistics 24 for practical calculations of data Mathematical Statistics is a recommended text for advanced students and practitioners of math, probability, and statistics.

Relational Research and Organisation Studies

Relational Research and Organisation Studies does not only present and discuss guidelines for practice at a onto-epistemological level but also presents and discusses concrete cases of research projects building on relational constructionist ideas. Furthermore, excerpts of data are presented and analyzed in order to explain the co-constructed processes of the inquiries more in detail.

Mastering Qlik Sense

Mastering Qlik Sense is a comprehensive guide designed to empower you to utilize Qlik Sense for advanced data analytics and dynamic visualizations. This book provides detailed insights into creating seamless Business Intelligence solutions tailored to your needs. Whether you're building dashboards, optimizing data models, or exploring Qlik Cloud functionalities, this book has you covered. What this Book will help me do Build interactive and insightful dashboards using Qlik Sense's intuitive tools. Learn to model data efficiently and apply best practices for optimized performance. Master the Qlik Sense APIs and create advanced custom extensions. Understand enterprise security measures including role-based access controls. Gain expertise in migrating from QlikView to Qlik Sense effectively Author(s) Juan Ignacio Vitantonio is an experienced expert in Business Intelligence solutions and data analytics. With a profound understanding of Qlik technologies, Juan has developed and implemented impactful BI solutions across various industries. His writing reflects his practical knowledge and passion for empowering users with actionable insights into data. Who is it for? This book is perfect for BI professionals, data analysts, and organizations aiming to leverage Qlik Sense for advanced analytics. Ideal for those with a foundational grasp of Qlik Sense, it also provides comprehensive guidance for QlikView users transitioning to Qlik Sense. If you want to improve your BI solutions and data-driven decision-making skills, this book is for you.

SQL Server 2017 Developer???s Guide

"SQL Server 2017 Developer's Guide" provides a comprehensive approach to learning and utilizing the new features introduced in SQL Server 2017. From advanced Transact-SQL to integrating R and Python into your database projects, this book equips you with the knowledge to design and develop efficient database applications tailored to modern requirements. What this Book will help me do Master new features in SQL Server 2017 to enhance database application development. Implement In-Memory OLTP and columnstore indexes for optimal performance. Utilize JSON support in SQL Server to integrate modern data formats. Leverage R and Python integration to apply advanced data analytics and machine learning. Learn Linux and container deployment options to expand SQL Server usage scenarios. Author(s) The authors of "SQL Server 2017 Developer's Guide" are industry veterans with extensive experience in database design, business intelligence, and advanced analytics. They bring a practical, hands-on writing style that helps developers apply theoretical concepts effectively. Their commitment to teaching is evident in the clear and detailed guidance provided throughout the book. Who is it for? This book is ideal for database developers and solution architects aiming to build robust database applications with SQL Server 2017. It's a valuable resource for business intelligence developers or analysts seeking to harness SQL Server 2017's advanced features. Some familiarity with SQL Server and T-SQL is recommended to fully leverage the insights provided by this book.

The SAS Programmer's PROC REPORT Handbook

Generate reports with style! The SAS Programmer's PROC REPORT Handbook: ODS Companion explains how to use style elements within a style template to customize reports generated by PROC REPORT, leading to more appealing and effective business reports. Many programmers are faced with generating reports that are easy to read and comprehend for a wide variety of audiences, which is where the ODS destinations and style changes come into play. This book teaches you how to use style elements in PROC REPORT, a versatile reporting procedure, to customize your output. Mastering style elements allows you to change visual aspects of reports, such as borders, column widths, fonts, backgrounds, and more. This companion to The SAS Programmer’s PROC REPORT Handbook: Basic to Advanced Reporting Techniques explores how the style elements within a style template affect the output generated by PROC REPORT. It provides examples of altering the style elements and the effect on the main ODS destinations, while also discussing common pitfalls that programmers can avoid while working with tables, Microsoft Excel, Microsoft Power Point, and PDF output.

Cleaning Up the Data Lake with an Operational Data Hub

The data lake was once heralded as the answer to the flood of big data that arrived in a variety of structured and unstructured formats. But, due to the ease of integration and the lack of governance, data lakes in many companies have devolved into unusable data swamps. This short ebook shows you how to solve this problem using an Operational Data Hub (ODH) to collect, store, index, cleanse, harmonize, and master data of all shapes and formats. Gerhard Ungerer—CTO and co-founder of Random Bit LLC—explains how the ODH supports transactional integrity so that the hub can serve as integration point for enterprise applications. You’ll also learn how the ODH helps you leverage the investment in your data lake (or swamp), so that the data trapped there can finally be ingested, processed, and provisioned. With this ebook, you’ll learn how an ODH: Allows you to focus on categorizing data for easy and fast retrieval Provides flexible storage models, indexing support, query capabilities, security, and a governance framework Delivers flexible storage models; support for indexing, scripting, and automation; query capabilities; transactional integrity; and security Includes a governance model to help you access, ingest, harmonize, materialize, provision, and consume data

Data as a Feature

Business applications are evolving and user expectations for quality, easy-to-use software is at an all-time high. The consumerization of applications is making the role of product managers more difficult than ever. How do you build products or services that meet demands for both power and simplicity? Companies are now gaining competitive advantage by providing intuitive application experiences that help users achieve goals. The best applications—the ones that stick—are those that empower users to realize the full value of their data. In this book, we explore how treating data as a feature can help product managers create indisposable applications that help users solve their most critical goals. Understand your users’ goals, the data they’ll need to achieve them, where the data comes from, and how to visualize data effectively Use personas to help you keep users in mind when making critical development and design decisions Provide an interwoven data experience by immersing reports, dashboards, and visualizations into your applications Make your data “over-the-counter” so that you and your users can accurately and easily interpret it Learn how to manage your data roadmap and handle requests for additional features

Gaining Data Agility with Multi-Model Databases

Most organizations realize that their future depends on the ability to quickly adapt to constant changes brought on by variable and complex environments. It's become increasingly clear that the core source behind these innovative solutions is data. Polyglot persistence refers to systems that provide many different types of data storage technologies to deal with this vast variability of data. Applications that need to access data from more than one store have to navigate an array of databases in a complex—and ultimately unsustainable—maze. One solution to this problem is readily available. In this ebook, consultant Joel Ruisi explains how a multi-model database enables you to take advantage of many different types of data models (and multiple schemas) in a single backend. With a multi-model database, companies can easily centralize, manage, and search all the data the IT system collects. The result is data agility: the ability to adapt to changing environments and serve users what they need when they need it. Through several detailed use cases, this ebook explains how multi-model databases enable you to: Store and manage multiple heterogeneous data sources Consolidate your data by bringing everything in "as is" Invisibly extend model features from one model to another Take a hybrid approach to analytical and operational data Enhance user search experience, including big data search Conduct queries across data models Offer SQL without relational constraints

MarkLogic Cookbook

Learn how to get the most out of MarkLogic with recipes from people who understand this powerful multi-model database platform from the inside out. MarkLogic comes with a broad set of capabilities to help you quickly integrate data from silos, but it takes time to learn how to harness that power. In this three-part series, key members of the MarkLogic team—including engineers who built the database—provide targeted recipes to get you up to speed. In Part 1, you’ll learn how to solve real-world problems with XQuery, the functional language for working with hierarchical data structures such as XML. Part 2 helps you solve common search-related problems with recipes that work with MarkLogic 9 as well as with older versions. With recipes in Part 3, you’ll explore the multiple ways MarkLogic represents data. XQuery: Gain XQuery peak performance, and explore its use in maps, documents, document security, the task server, and administration Search-related problems: Conduct document searches, score search results, understand how data is used, and search with the Optic API MarkLogic and data: Work with input transformations, tokenization, template-driven extraction, and redaction

Neural networks and deep learning

Neural networks are at the very core of deep learning. They are versatile, powerful, and scalable, making them ideal to tackle large and highly complex Machine Learning tasks, such as classifying billions of images (e.g., Google Images), powering speech recognition services (e.g., Apple’s Siri), recommending the best videos to watch to hundreds of millions of users every day (e.g., YouTube), or learning to beat the world champion at the game of Go by examining millions of past games and then playing against itself (DeepMind’s AlphaGo). This lesson introduces artificial neural networks, starting with a quick tour of the very first ANN architectures, then covering topics such as training neural nets, recurrent neural networks, and reinforcement learning. This lesson will clarify what neural networks are and why you may want to use them.

SAS for Forecasting Time Series, Third Edition, 3rd Edition

To use statistical methods and SAS applications to forecast the future values of data taken over time, you need only follow this thoroughly updated classic on the subject. With this third edition of SAS for Forecasting Time Series, intermediate-to-advanced SAS users—such as statisticians, economists, and data scientists—can now match the most sophisticated forecasting methods to the most current SAS applications. Starting with fundamentals, this new edition presents methods for modeling both univariate and multivariate data taken over time. From the well-known ARIMA models to unobserved components, methods that span the range from simple to complex are discussed and illustrated. Many of the newer methods are variations on the basic ARIMA structures. Completely updated, this new edition includes fresh, interesting business situations and data sets, and new sections on these up-to-date statistical methods: ARIMA models Vector autoregressive models Exponential smoothing models Unobserved component and state-space models Seasonal adjustment Spectral analysis Focusing on application, this guide teaches a wide range of forecasting techniques by example. The examples provide the statistical underpinnings necessary to put the methods into practice. The following up-to-date SAS applications are covered in this edition: The ARIMA procedure The AUTOREG procedure The VARMAX procedure The ESM procedure The UCM and SSM procedures The X13 procedure The SPECTRA procedure SAS Forecast Studio Each SAS application is presented with explanation of its strengths, weaknesses, and best uses. Even users of automated forecasting systems will benefit from this knowledge of what is done and why. Moreover, the accompanying examples can serve as templates that you easily adjust to fit your specific forecasting needs. This book is part of the SAS Press program.

HBR Guide to Data Analytics Basics for Managers (HBR Guide Series)

Don't let a fear of numbers hold you back. Today's business environment brings with it an onslaught of data. Now more than ever, managers must know how to tease insight from data--to understand where the numbers come from, make sense of them, and use them to inform tough decisions. How do you get started? Whether you're working with data experts or running your own tests, you'll find answers in the HBR Guide to Data Analytics Basics for Managers. This book describes three key steps in the data analysis process, so you can get the information you need, study the data, and communicate your findings to others. You'll learn how to: Identify the metrics you need to measure Run experiments and A/B tests Ask the right questions of your data experts Understand statistical terms and concepts Create effective charts and visualizations Avoid common mistakes

IBM Storage Networking SAN512B-6 and SAN256B-6 Directors

This IBM® Redbooks® product guide describes the IBM Storage Networking SAN512B-6 (8961-F08) and SAN256B-6 (8961-F04) directors and the IBM b-type Gen 6 Extension Blade (FC 3892, 3893). Digital transformation is pushing mission-critical storage environments to the limit, with users expecting data to be accessible from anywhere, at any time, on any device. Faced with exponential data growth, the network must evolve to enable businesses to thrive in this new era. A new approach to storage networking is needed to enable databases, virtual servers, desktops, and critical applications, and to unlock the full capabilities of flash. By treating the network as a strategic part of a storage environment, organizations can maximize their productivity and efficiency even as they rapidly scale their environments. IBM Storage Networking SAN512B-6 and SAN256B-6 directors with Fabric Vision technology are modular building blocks that combine innovative hardware, software, and built-in instrumentation to ensure high levels of operational stability and redefine application performance. Fabric Vision technology enhances visibility into the health of storage environments, delivering greater control and insight to quickly identify problems and achieve critical service level agreements (SLAs). Breakthrough 32 Gbps performance shatters application performance barriers and provides support for more than 1 billion input/output operations per second (IOPS) for flash-based storage workloads while 128 Gbps UltraScale inter-chassis links enable simplified, high-bandwidth scalability between directors.

The Routledge Companion to Trust

Comprising six thematic sections, The Routledge Companion to Trust provides a comprehensive survey of trust research. With contributions from international experts, this volume examines the major topics and emerging areas within the field, including essays on the foundations, levels and theories of trust.