talk-data.com talk-data.com

Topic

data-engineering

3395

tagged

Activity Trend

1 peak/qtr
2020-Q1 2026-Q1

Activities

3395 activities · Newest first

Foundations of SQL Server 2008 R2 Business Intelligence, Second Edition

Foundations of SQL Server 2008 R2 Business Intelligence introduces the exciting gamut of business intelligence tools included with SQL Server 2008. Microsoft has designed SQL Server 2008 to be more than just a database. It's a complete business intelligence (BI) platform. The database is at the core, and surrounding the database are tools for data mining, modeling, reporting, analyzing, charting, and integration with other, enterprise-level software packages. SQL Server 2008 puts an incredible amount of BI functionality at your disposal. But how do you take advantage of it? That's what this book is all about. Authors Guy Fouché and Lynn Langit show how to implement end-to-end BI solutions using SQL Server Analysis Services (SSAS), SQL Server Integration Services (SSIS), SQL Server Reporting Services (SSRS), and other tools in the Microsoft BI toolkit. You'll learn about all-new features such as PowerPivot and Report Builder 3.0. Also provided are clear examples of predictive analysis made possible through powerful data mining features in SQL Server. If you're an analyst or developer working with SQL Server 2008 who is charged with delivering results that drive business success, you can't afford to be without this book; you can't afford to ignore the powerful BI suite that Microsoft has placed at your disposal. Provides the "big picture" of Microsoft's BI tool suite Covers PowerPivot and other game-changing technologies introduced alongside SQL Server 2008 Release 2 Gives a practical analysis of features based on real-world practices What you'll learn Apply dimensional- and fact-based modeling to create OLAP schemas Optimize data extraction and transformation with SQL Server Integration Services Develop OLAP and data-mining solutions using SQL Server Analysis Services Create compelling reports using the new Report Builder 3.0 in SQL Server Reporting Services Deliver self-service business-intelligence via Microsoft PowerPivot Manage and maintain control over your business intelligence environment Who this book is for Foundations of SQL Server 2008 R2 Business Intelligence is for business intelligence analysts and developers wanting to employ the full stack of business intelligence tools that are part of SQL Server. The book also appeals to database administrators managing databases in support of business intelligence efforts.

ActiveMQ in Action

ActiveMQ in Action is all you'll need to master ActiveMQ. It starts from the anatomy of a JMS message and moves quickly through connectors, message persistence, authentication, and authorization. By following a running example (a stock portfolio app), you'll pick up the best practices distilled by the authors from their long and deep involvement with this technology. About the Technology The Apache ActiveMQ message broker is an open source implementation of the Java Message Service spec. It makes for a reliable hub in any message-oriented enterprise application and integrates beautifully with Java EE containers, ESBs, and other JMS providers. About the Book What's Inside How to design message-based apps How to implement EI patterns using Camel How to administer ActiveMQ How to integrate with Geronimo, JBoss, Spring, and more About the Reader This book requires a working knowledge of Java, but no previous experience with ActiveMQ or other message brokers is needed. About the Authors Bruce Snyder is a co-founder of Apache Geronimo, a committer for ActiveMQ, Camel, and ServiceMix, and a member of various JCP expert groups. Dejan Bosanac is an ActiveMQ committer. Rob Davies is a co-founder of ActiveMQ, ServiceMix, and Camel. Quotes Covers everything you need to know about ActiveMQ. - Pratik Patel, AT&T A vital resource. - John Merryman, Yodle Complete and comprehensive, a must-have resource. - Rod Biresch, Chariot Solutions Authors have in-depth knowledge of ActiveMQ. - Roberto J. Rojas, Chariot Solutions Covers the basics, and then goes way beyond. - Jeff Davis, Author of Open Source SOA

Beginning Oracle Application Express 4

Beginning Oracle Application Express 4 introduces one of the most talked about development platforms to come out of Oracle Corporation in years. Oracle Application Express, called APEX for short, enables rapid and easy development of web-based applications that make full use of Oracle Database. The release of APEX 4 brings a huge leap forward in terms of functionality and usability for both the developer and the end user. Power users and programmers alike can quickly put together robust and scalable applications for use by one person, by a department, by an entire company. Whether you're new to Oracle or an old hand who's yet to test the waters of APEX, Beginning Oracle Application Express 4 introduces the processes and best practices you'll need to become proficient with APEX. The book shows off the programming environment, the utilities and tools available, and then continues by walking through the process of building a working system from the ground up. While the book assumes a foundational knowledge of SQL and PL/SQL, all code used is documented and explained so that those new to the languages will not be lost. Covers brand-new functionality in APEX 4 Provides fully documented and explained example code Guides you through creating a working and fully deployable application What you'll learn Provides fully documented and explained example code Guides you through creating a working and fully deployable application Design, code, and troubleshoot APEX applications Create forms and generate reports Apply best-practices to database and application design Create intuitive navigation that inspires confidence amongst users Secure your applications from outside attack Deploy applications for one, a dozen, or hundreds of users Who this book is for Beginning Oracle Application Express 4 is for programmers and power users wanting to create rich, data-centric applications that are web-based, and that rest on Oracle's industry-leading database technology.

Implementing an IBM b-type SAN with 8 Gbps Directors and Switches

"Do everything that is necessary and absolutely nothing that is not." This IBM® Redbooks® publication, written at a Data Center Fabric Manager v10.1.4 and Fabric Operating System v6.4 level, consolidates critical information while also covering procedures and tasks that you are likely to encounter on a daily basis when implementing an IBM b-type SAN. The products that we describe in this book have more functionality than we can possibly cover in a single book. A storage area network (SAN) is a powerful infrastructure for consolidation, distance solutions, and data sharing. The quality applications that the IBM SAN portfolio provides can help you take full advantage of the benefits of the SAN. In this book, we cover the latest additions to the IBM b-type SAN family and show how you can implement them in an open systems environment. In particular, we focus on the Fibre Channel Protocol (FCP) environment. We address the key concepts that these products bring to the market and, in each case, we provide an overview of the functions that are essential to building a robust SAN environment.

Scaling CouchDB

This practical guide offers a short course on scaling CouchDB to meet the capacity needs of your distributed application. Through a series of scenario-based examples, this book lets you explore several methods for creating a system that can accommodate growth and meet expected demand. In the process, you learn about several tools that can help you with replication, load balancing, clustering, and load testing and monitoring. Apply performance tips for tuning your database Replicate data, using Futon and CouchDB’s RESTful interface Distribute CouchDB’s workload through load balancing Learn options for creating a cluster of CouchDB nodes, including BigCouch, Lounge, and Pillow Conduct distributed load testing with Tsung

Microsoft SQL Server 2008 R2 Master Data Services

Best Practices for Deploying and Managing Master Data Services (MDS) Effectively manage master data and drive better decision making across your enterprise with detailed instruction from two MDS experts. Microsoft SQL Server 2008 R2 Master Data Services Implementation & Administration shows you how to use MDS to centralize the management of key data within your organization. Find out how to build an MDS model, establish hierarchies, govern data access, and enforce business rules. Legacy system integration and security are also covered. Real-world programming examples illustrate the material presented in this comprehensive guide. Create a process-agnostic solution for managing your business domains Learn how to take advantage of the data modeling capabilities of MDS Manage hierarchies and consolidations across your organization Import data by using SQL Server Integration Services and T-SQL statements Ensure data accuracy and completeness by using business rules and versioning Employ role-based security at functional, object, and attribute levels Design export views and publish data to subscribing systems Use Web services to progrmmatically interact with your implementation

IBM System Storage b-type Multiprotocol Routing: An Introduction and Implementation

The rapid spread and adoption of production storage area networks (SANs) has fueled the need for multiprotocol routers. The routers provide improved scalability, security, and manageability by enabling devices in separate SAN fabrics to communicate without merging fabrics into a single, large SAN fabric. This capability enables clients to initially deploy separate SAN solutions at the departmental and data center levels. Then, clients can consolidate these separate solutions into large enterprise SAN solutions as their experience and requirements grow and change. Alternatively, multiprotocol routers can help to connect existing enterprise SANs for a variety of reasons. For instance, the introduction of Small Computer System Interface over IP (iSCSI) provides for the connection of low-end, low-cost hosts to enterprise SANs. The use of an Internet Protocol (IP) in the Fibre Channel (FC) environment provides for resource consolidation and disaster recovery planning over long distances. And the use of FC-FC routing services provides connectivity between two or more fabrics without having to merge them into a single SAN. This IBM® Redbooks® publication targets storage network administrators, system designers, architects, and IT professionals who sell, design, or administer SANs. It introduces you to products, concepts, and technology in the IBM System Storage™ SAN Routing portfolio, which is based on Brocade products and technology. This book shows the features of these products and examples of how you can deploy and use them.

Achieving Extreme Performance with Oracle Exadata

Maximize Oracle Exadata Capabilities Leverage all of the powerful features available in the Oracle Exadata Database Machine using the proven techniques inside this Oracle Press guide. Written by Oracle experts, Achieving Extreme Performance with Oracle Exadata shows you how to take full advantage of this complete, optimized package of software, servers, and storage. Best practices for enterprise deployments, high availability, administration, backup and recovery, data warehousing, online transaction processing, consolidation, and migration are included in this authoritative resource. Take advantage of the tightly integrated hardware and software in the Oracle Exadata Database Machine Work with Oracle Exadata software features, including Smart Scans, Oracle Exadata Hybrid Columnar Compression, storage indexes, Oracle Exadata Smart Flash Cache, and I/O Resource Manager Understand Oracle Exadata Database Machine balanced hardware architecture Architect, administer, and monitor Oracle Exadata Storage Servers Deploy data warehouses on the Oracle Exadata Database Machine Run online transaction processing (OLTP) workloads on the Oracle Exadata Database Machine Consolidate databases with and migrate databases to the Oracle Exadata Database Machine

The Microsoft® Data Warehouse Toolkit: With SQL Server 2008 R2 and the Microsoft® Business Intelligence Toolset, Second Edition

Best practices and invaluable advice from world-renowned data warehouse experts In this book, leading data warehouse experts from the Kimball Group share best practices for using the upcoming "Business Intelligence release" of SQL Server, referred to as SQL Server 2008 R2. In this new edition, the authors explain how SQL Server 2008 R2 provides a collection of powerful new tools that extend the power of its BI toolset to Excel and SharePoint users and they show how to use SQL Server to build a successful data warehouse that supports the business intelligence requirements that are common to most organizations. Covering the complete suite of data warehousing and BI tools that are part of SQL Server 2008 R2, as well as Microsoft Office, the authors walk you through a full project lifecycle, including design, development, deployment and maintenance. Features more than 50 percent new and revised material that covers the rich new feature set of the SQL Server 2008 R2 release, as well as the Office 2010 release Includes brand new content that focuses on PowerPivot for Excel and SharePoint, Master Data Services, and discusses updated capabilities of SQL Server Analysis, Integration, and Reporting Services Shares detailed case examples that clearly illustrate how to best apply the techniques described in the book The accompanying Web site contains all code samples as well as the sample database used throughout the case studies The Microsoft Data Warehouse Toolkit, Second Edition provides you with the knowledge of how and when to use BI tools such as Analysis Services and Integration Services to accomplish your most essential data warehousing tasks.

Advanced Kalman Filtering, Least-Squares and Modeling: A Practical Handbook

This book provides a complete explanation of estimation theory and application, modeling approaches, and model evaluation. Each topic starts with a clear explanation of the theory (often including historical context), followed by application issues that should be considered in the design. Different implementations designed to address specific problems are presented, and numerous examples of varying complexity are used to demonstrate the concepts. This book is intended primarily as a handbook for engineers who must design practical systems. Its primary goal is to explain all important aspects of Kalman filtering and least-squares theory and application. Discussion of estimator design and model development is emphasized so that the reader may develop an estimator that meets all application requirements and is robust to modeling assumptions. Since it is sometimes difficult to a priori determine the best model structure, use of exploratory data analysis to define model structure is discussed. Methods for deciding on the "best" model are also presented. A second goal is to present little known extensions of least squares estimation or Kalman filtering that provide guidance on model structure and parameters, or make the estimator more robust to changes in real-world behavior. A third goal is discussion of implementation issues that make the estimator more accurate or efficient, or that make it flexible so that model alternatives can be easily compared. The fourth goal is to provide the designer/analyst with guidance in evaluating estimator performance and in determining/correcting problems. The final goal is to provide a subroutine library that simplifies implementation, and flexible general purpose high-level drivers that allow both easy analysis of alternative models and access to extensions of the basic filtering.

Data Source Handbook

If you're a developer looking to supplement your own data tools and services, this concise ebook covers the most useful sources of public data available today. You'll find useful information on APIs that offer broad coverage, tie their data to the outside world, and are either accessible online or feature downloadable bulk data. You'll also find code and helpful links. This guide organizes APIs by the subjects they cover—such as websites, people, or places—so you can quickly locate the best resources for augmenting the data you handle in your own service. Categories include: Website tools such as WHOIS, bit.ly, and Compete Services that use email addresses as search terms, including Github Finding information from just a name, with APIs such as WhitePages Services, such as Klout, for locating people with Facebook and Twitter accounts Search APIs, including BOSS and Wikipedia Geographical data sources, including SimpleGeo and U.S. Census Company information APIs, such as CrunchBase and ZoomInfo APIs that list IP addresses, such as MaxMind Services that list books, films, music, and products

IBM System Storage DS Storage Manager Copy Services Guide

The purpose of this IBM® Redbooks® publication is to provide customers with guidance and recommendations for how and when to use the IBM System Storage® Copy Services premium features. The topics discussed in this publication apply to the IBM System Storage DS® models DS3000, DS4000®, and DS5000 running the firmware v7.70, and IBM System Storage DS Storage Manager v10.70. Customers in today’s IT world are finding a major need to ensure a good archive of their data and a requirement to create these archives with minimal interruptions. The IBM Midrange System Storage helps to fulfill these requirements by offering three copy services premium features: IBM FlashCopy® VolumeCopy Enhanced Remote Mirroring (ERM) This publication specifically addresses the copy services premium features and can be used in conjunction with the following IBM DS System Storage books: IBM System Storage DS4000 and Storage Manager V10.30, SG24-7010 IBM System Storage DS3000: Introduction and Implementation Guide, SG24-7065 IBM System Storage DS3500: Introduction and Implementation Guide, SG24-7914 IBM Midrange System Storage Hardware Guide, SG24-7676 IBM Midrange System Storage Implementation and Best Practices Guide, SG24-6363

Scaling MongoDB

Create a MongoDB cluster that will to grow to meet the needs of your application. With this short and concise book, you'll get guidelines for setting up and using clusters to store a large volume of data, and learn how to access the data efficiently. In the process, you'll understand how to make your application work with a distributed database system. Scaling MongoDB will help you: Set up a MongoDB cluster through sharding Work with a cluster to query and update data Operate, monitor, and backup your cluster Plan your application to deal with outages By following the advice in this book, you'll be well on your way to building and running an efficient, predictable distributed system using MongoDB.

Database Modeling and Design, 5th Edition

Database Modeling and Design, Fifth Edition, focuses on techniques for database design in relational database systems. This extensively revised fifth edition features clear explanations, lots of terrific examples and an illustrative case, and practical advice, with design rules that are applicable to any SQL-based system. The common examples are based on real-life experiences and have been thoroughly class-tested. This book is immediately useful to anyone tasked with the creation of data models for the integration of large-scale enterprise data. It is ideal for a stand-alone data management course focused on logical database design, or a supplement to an introductory text for introductory database management. In-depth detail and plenty of real-world, practical examples throughout Loaded with design rules and illustrative case studies that are applicable to any SQL, UML, or XML-based system Immediately useful to anyone tasked with the creation of data models for the integration of large-scale enterprise data

Automated Physical Database Design and Tuning

Due to the increasing complexity in application workloads and query engines, database administrators are turning to automated tuning tools that systematically explore the space of physical design alternatives. A critical element of such tuning is physical database design since the choice of physical structures has a significant impact on the performance of the database system. Automated Physical Database Design and Tuning presents a detailed overview of the fundamental ideas and algorithms for automatically recommending changes to the physical design of a database system. The first part of the book introduces the necessary technical background. The author explains SQL, the space of execution plans for answering SQL queries, query optimization, how the choice of access paths (e.g., indexes) is crucial to performance, and the complexity of the physical design problem. The second part extensively discusses automated physical design techniques, covering fundamental research ideas in the last 15 years that have resulted in a new generation of tuning tools. The text focuses on the search space of alternatives, the necessity of a cost model to compare such alternatives, different mechanisms to traverse and enumerate the search space, and practical aspects in real-world tuning tools. In the third part, the author explores new advances in automated physical design. He applies previous approaches to other physical structures, such as materialized views, partitioning, and multidimensional clustering. He also analyzes workload models for new types of applications, generalizes the optimizing function of current physical design tools to cope with other application scenarios, and examines open-ended challenges in physical database design. This book offers valuable insights on well-established principles and cutting-edge research results in automated physical design. It helps readers gain a deeper understanding of how automated tuning tools work in database installations as well as the challenges and opportunities involved in designing next-generation tuning tools.

Developing High Quality Data Models

Developing High Quality Data Models provides an introduction to the key principles of data modeling. It explains the purpose of data models in both developing an Enterprise Architecture and in supporting Information Quality; common problems in data model development; and how to develop high quality data models, in particular conceptual, integration, and enterprise data models. The book is organized into four parts. Part 1 provides an overview of data models and data modeling including the basics of data model notation; types and uses of data models; and the place of data models in enterprise architecture. Part 2 introduces some general principles for data models, including principles for developing ontologically based data models; and applications of the principles for attributes, relationship types, and entity types. Part 3 presents an ontological framework for developing consistent data models. Part 4 provides the full data model that has been in development throughout the book. The model was created using Jotne EPM Technologys EDMVisualExpress data modeling tool. This book was designed for all types of modelers: from those who understand data modeling basics but are just starting to learn about data modeling in practice, through to experienced data modelers seeking to expand their knowledge and skills and solve some of the more challenging problems of data modeling. Uses a number of common data model patterns to explain how to develop data models over a wide scope in a way that is consistent and of high quality Offers generic data model templates that are reusable in many applications and are fundamental for developing more specific templates Develops ideas for creating consistent approaches to high quality data models

Extremely pureXML in DB2 10 for z/OS

The DB2® pureXML® feature offers sophisticated capabilities to store, process and manage XML data in its native hierarchical format. By integrating XML data intact into a relational database structure, users can take full advantage of DB2's relational data management features. In this IBM® Redbooks® publication, we document the steps for the implementation of a simple but meaningful XML application scenario. We have chosen to provide samples in COBOL and Java™ language. The purpose is to provide an easy path to follow to integrate the XML data type for the traditional DB2 for z/OS® user. We also add considerations for the data administrator and suggest best practices for ease of use and better performance.

Mastering XPages: A Step-by-Step Guide to XPages Application Development and the XSP Language

The first complete, practical guide to XPages development - direct from members of the XPages development team at IBM Lotus Martin Donnelly, Mark Wallace, and Tony McGuckin have written the definitive programmer's guide to utilizing this breakthrough technology. Packed with tips, tricks, and best practices from IBM's own XPages developers, Mastering XPages brings together all the information developers need to become experts - whether you’re experienced with Notes/Domino development or not. The authors start from the very beginning, helping developers steadily build your expertise through practical code examples and clear, complete explanations. Readers will work through scores of real-world XPages examples, learning cutting-edge XPages and XSP language skills and gaining deep insight into the entire development process. Drawing on their own experience working directly with XPages users and customers, the authors illuminate both the technology and how it can be applied to solving real business problems. Martin Donnelly previously led a software startup that developed and distributed small business accounting software. Donnelly holds a Commerce degree from University College Cork and an M.S. in Computer Science from Boston University. Mark Wallace has worked at IBM for 15 years on many projects as a technical architect and application developer. Tony McGuckin participates in the Lotus OneUI Web Application and iWidget Adoption Workgroup. He holds a bachelor's degree in Software Engineering from the University of Ulster.

Tivoli Integration Scenarios

This IBM® Redbooks® publication provides a broad view of how Tivoli® system management products work together in several common scenarios. You must achieve seamless integration for operations personnel to work with the solution. This integration is necessary to ensure that the product can be used easily by the users. Product integration contains multiple dimensions, such as security, navigation, data and task integrations. Within the context of the scenarios in this book, you see examples of these integrations. The scenarios implemented in this book are largely based on the input from the integration team, and several clients using IBM products. We based these scenarios on common real-life examples that IT operations often have to deal with. Of course, these scenarios are only a small subset of the possible integration scenarios that can be accomplished by the Tivoli products, but they were chosen to be representative of the integration possibilities using the Tivoli products. We discuss these implementations and benefits that are realized by these integrations, and also provide sample scenarios of how these integrations work. This book is a reference guide for IT architects and IT specialists working on integrating Tivoli products in real-life environments.