talk-data.com talk-data.com

Event

O'Reilly Data Science Books

2013-08-09 – 2026-02-25 Oreilly Visit website ↗

Activities tracked

2118

Collection of O'Reilly books on Data Science.

Sessions & talks

Showing 2076–2100 of 2118 · Newest first

Search within this event →
Even You Can Learn Statistics A Guide for Everyone Who Has Ever Been Afraid of Statistics

Even You Can Learn Statistics A Guide for Everyone Who Has Ever Been Afraid Of Statistics One easy step at a time, this book will teach you the key statistical techniques you'll need for finance, quality, marketing, the social sciences, or just about any other field. Each technique is introduced with a simple, jargon-free explanation, practical examples, and hands-on guidance for solving real problems with Excel or a TI-83/84 series calculator, including Plus models. Hate math? No sweat. You'll be amazed how little you need! For those who do have an interest in mathematics, optional "Equation Blackboard" sections review the equations that provide the foundations for important concepts. David M. Levine is a much-honored innovator in statistics education. He is Professor Emeritus of Statistics and Computer Information Systems at Bernard M. Baruch College (CUNY), and co-author of several best-selling books, including Statistics for Managers using Microsoft Excel, Basic Business Statistics, Quality Management, and Six Sigma for Green Belts and Champions. Instructional designer David F. Stephan pioneered the classroom use of personal computers, and is a leader in making Excel more accessible to statistics students. He has co-authored several textbooks with David M. Levine. Here's just some of what you'll learn how to do... Use statistics in your everyday work or study Perform common statistical tasks using a Texas Instruments statistical calculator or Microsoft Excel Build and interpret statistical charts and tables "Test Yourself" at the end of each chapter to review the concepts and methods that you learned in the chapter Work with mean, median, mode, standard deviation, Z scores, skewness, and other descriptive statistics Use probability and probability distributions Work with sampling distributions and confidence intervals Test hypotheses and decision-making risks with Z, t, Chi-Square, ANOVA, and other techniques Perform regression analysis and modeling The easy, practical introduction to statistics–for everyone! Thought you couldn't learn statistics? Think again. You can–and you will! Complementary Web site Downloadable practice files at http://www.ftpress.com/youcanlearnstatistics

PROC SQL: Beyond the Basics Using SAS®

Master the language of PROC SQL! Get faster and better results when you use the SQL procedure with this step-by-step, example-driven guide in your hands. You will find information about implementing both American National Standards Institute (ANSI) guidelines and non-ANSI features. A broad range of SAS users, including applications programmers, database designers and administrators, statisticians, and systems analysts, will find techniques in Kirk Lafler's PROC SQL: Beyond the Basics Using SAS to enhance their data access, manipulation, and presentation prowess. Topics examined include database design; an introduction to SQL, dictionary tables, and tables; coding PROC SQL logic as well as complex queries; working with views; troubleshooting and debugging SQL code; and tuning for performance. With the growing popularity of databases, PROC SQL is becoming the language of choice for SAS users around the world. Intermediate and advanced users will appreciate the straightforward explanations and numerous examples.

Say It With Charts Workbook

Hands-on tips for powerful presentations in this all-new companion to the bestselling Say It with Charts Through four editions, Gene Zelazny's classic how-to Say It with Charts has generated more than $1.5 million in revenues. Now, in the companion Say It with Charts Workbook, Zelazny shows you how to make even more of your visual communication skills, working "one-on-one" with you on how to masterfully use the latest techniques and tools to enliven every presentation. More than just a rote listing of techniques, Say It with Charts Workbook features performance-improving strategies and suggestions that will help keep both you--and, even more important, the audience--comfortable and at ease. Part refresher course, part workbook, part self-test, it arms you with: • Step-by-step instructions and guidelines • Performance-improving strategies and suggestions • Tactics for customizing graphics to specific audiences

Elasticity

Although there are several books in print dealing with elasticity, many focus on specialized topics such as mathematical foundations, anisotropic materials, two-dimensional problems, thermoelasticity, non-linear theory, etc. As such they are not appropriate candidates for a general textbook. This book provides a concise and organized presentation and development of general theory of elasticity. Complemented by a Solutions Manual and including MatLab codes and coding, this text is an excellent book teaching guide. - Contains exercises for student engagement as well as the integration and use of MATLAB Software - Provides development of common solution methodologies and a systematic review of analytical solutions useful in applications of engineering interest - Presents applications of contemporary interest

Preparing for DB2 Near-Realtime Business Intelligence

In this IBM Redbooks publication we discuss primary processes and various alternatives that prepare you in implementing a DB2 near-realtime business intelligence environment. We discuss architectural alternatives and include overviews of software products that you can use in an implementation. As a primary focus, we tested the capabilities for supporting continuous update of a DB2 data warehouse while running a continuous concurrent query workload against that data warehouse. We tested several implementation scenarios and the variables that impact them. The results of our testing and the issues we encountered are provided.

Algorithmic Graph Theory and Perfect Graphs, 2nd Edition

Algorithmic Graph Theory and Perfect Graphs, first published in 1980, has become the classic introduction to the field. This new Annals edition continues to convey the message that intersection graph models are a necessary and important tool for solving real-world problems. It remains a stepping stone from which the reader may embark on one of many fascinating research trails. The past twenty years have been an amazingly fruitful period of research in algorithmic graph theory and structured families of graphs. Especially important have been the theory and applications of new intersection graph models such as generalizations of permutation graphs and interval graphs. These have lead to new families of perfect graphs and many algorithmic results. These are surveyed in the new Epilogue chapter in this second edition. New edition of the "Classic" book on the topic Wonderful introduction to a rich research area Leading author in the field of algorithmic graph theory Beautifully written for the new mathematician or computer scientist Comprehensive treatment

Spidering Hacks

The Internet, with its profusion of information, has made us hungry for ever more, ever better data. Out of necessity, many of us have become pretty adept with search engine queries, but there are times when even the most powerful search engines aren't enough. If you've ever wanted your data in a different form than it's presented, or wanted to collect data from several sites and see it side-by-side without the constraints of a browser, then Spidering Hacks is for you. Spidering Hacks takes you to the next level in Internet data retrieval--beyond search engines--by showing you how to create spiders and bots to retrieve information from your favorite sites and data sources. You'll no longer feel constrained by the way host sites think you want to see their data presented--you'll learn how to scrape and repurpose raw data so you can view in a way that's meaningful to you.Written for developers, researchers, technical assistants, librarians, and power users, Spidering Hacks provides expert tips on spidering and scraping methodologies. You'll begin with a crash course in spidering concepts, tools (Perl, LWP, out-of-the-box utilities), and ethics (how to know when you've gone too far: what's acceptable and unacceptable). Next, you'll collect media files and data from databases. Then you'll learn how to interpret and understand the data, repurpose it for use in other applications, and even build authorized interfaces to integrate the data into your own content. By the time you finish Spidering Hacks, you'll be able to: Aggregate and associate data from disparate locations, then store and manipulate the data as you like Gain a competitive edge in business by knowing when competitors' products are on sale, and comparing sales ranks and product placement on e-commerce sites Integrate third-party data into your own applications or web sites Make your own site easier to scrape and more usable to others Keep up-to-date with your favorite comics strips, news stories, stock tips, and more without visiting the site every dayLike the other books in O'Reilly's popular Hacks series, Spidering Hacks brings you 100 industrial-strength tips and tools from the experts to help you master this technology. If you're interested in data retrieval of any type, this book provides a wealth of data for finding a wealth of data.

The Little SAS® Book: A Primer, Third Edition

Getting started with SAS is easy with The Little SAS Book. This friendly, easy-to-read guide gently introduces you to the most commonly used features of SAS software plus a whole lot more! The book has been revised to include concepts of the Output Delivery System; the STYLE= option in the PRINT, REPORT, and TABULATE procedures; ODS HTML, RTF, PRINTER, and OUTPUT destinations; PROC REPORT; more on PROC TABULATE; exporting data; and the colon modifier for informats. You'll find clear and concise explanations of basic SAS concepts (such as DATA and PROC steps), inputting data, modifying and combining data sets, summarizing and presenting data, basic statistical procedures, and debugging SAS programs. Each topic is presented in a self-contained, two-page layout complete with examples and graphics. This format enables new users to get up and running quickly, while the examples allow you to type in the program and see it work!

Supply Chains: A Manager's Guide

“An excellent summary of the state of supply chain management going into the twenty-first century. Explains the essential concepts clearly and offers practical, down-to-earth advice for making supply chains more efficient and adaptive. Truly a survival guide for executives as they struggle to cope with the increasing competition between supply chains.” — Christian Knoll, Vice President of Global Supply Chain Management, SAP AG “Through real-world case studies and graphic illustrations, David Taylor clearly demonstrates the bottom-line benefits of managing the supply chain effectively. Although the book is written for managers, I recommend it for everyone from the executive suite to the shipping floor because they all have to work together to master the supply chain. But beware—you can expect many passionate employees demanding improvements in your company’s supply chain after reading this book!” — David Myers, President, WinfoSoft Inc., Former Board Member of Supply Chain Council “A comprehensive, thoroughly researched, and well-designed book that gives managers the information they need in a highly readable form. I am already starting to use the techniques in this book to improve our international distribution system.” — Jim Muller, Vice President of Produce Sales, SoFresh Produce “Supply chain management is a deceptively deep subject. Simple business practices combine to form complex systems that seem to defy rational analysis: Companies that form trading partnerships continue to compete despite their best efforts to cooperate; small variations in consumer buying create devastating swings in upstream demand, and so on. In his trademark fashion, Taylor clearly reveals the hidden logic at work in your supply chain and gives you the practical tools you need to make better management decisions. A must-read for every manager who affects a supply chain, and in today's marketplace there are few managers who are exempt from this requirement.” — Adrian J. Bowles, Ph.D., President, CoSource.net “David Taylor has done it again. With his new book, David makes supply chain management easy to grasp for the working manager, just as he did with his earlier guides to business technology. If you work for a company that is part of a supply chain, you need this book.” — Dirk Riehle, Ph.D. “David Taylor has done a masterful job of defining the core issues in supply chain management without getting trapped in the quicksand of jargon. This concise book is well written, highly informative, and easy to read.” — Marcia Robinson, President, E-Business Strategies, author of Services Blueprint: Roadmap “Taylor has done a tremendous job of giving readers an intuitive grasp of a complicated subject. If you’re new to supply chains, this book will give you an invaluable map of the territory. If you're already among the initiated, it will crystallize your insights and help you make better decisions. In either case, you can only come out ahead by reading this book.” — Kevin Dick, Founder of Kevin Dick Associates, author of XML: A Manager’s Guide “My motto for compressing data is ‘squeeze it til it gags.’ In the current business climate, that’s what you have to do to costs, and Taylor shows you many ways to squeeze costs out of your supply chain. He also writes with the same economy: This book contains exactly what you need to manage your supply chain effectively. Nothing is missing, and nothing is extra.” — Charles Ashbacher, President, Charles Ashbacher Technologies Today's fiercest business battles are taking place between competitors' supply chains, with victory dependent on finding a way to deliver products to customers more quickly and efficiently than the competition. For proof, just look to Dell and Amazon.com, both of which revolutionized their industries by changing how companies produce, distribute, and sell physical goods. But they're hardly alone. By revamping their supply chains, Siemens CT improved lead time from six months to two weeks, Gillette slashed $400 million of inventory, and Chrysler saved $1.7 billion a year. It's a high-stakes game, and you don't have a lot of choice about playing: If your company touches a physical product, it's part of a supply chain--and your success ultimately hangs on the weakest link in that chain. In , best-selling author David Taylor explains how to assemble a killer supply chain using the knowledge, technology, and tools employed in supply-chain success stories. Using his signature fast-track summaries and informative graphics, Taylor offers a clear roadmap to understanding and solving the complex problems of supply-chain management. Supply Chains: A Manager's Guide Modern manufacturing has driven down the time and cost of the production process, leaving supply chains as the final frontier for cost reduction and competitive advantage. will quickly give managers the foundation they need to contribute effectively to their company's supply-chain success. Supply Chains: A Manager's Guide

Database Modeling with Microsoft® Visio for Enterprise Architects

This book is for database designers and database administrators using Visio, which is the database component of Microsoft's Visual Studio .NET for Enterprise Architects suite, also included in MSDN subscriptions. This is the only guide to this product that tells DBAs how to get their job done. Although primarily focused on tool features, the book also provides an introduction to data modeling, and includes practical advice on managing database projects. The principal author was the program manager of VEA's database modeling solutions. · Explains how to model databases with Microsoft® Visio for Enterprise Architects (VEA), focusing on tool features.· Provides a platform-independent introduction to data modeling using both Object Role Modeling (ORM) and Entity Relationship Modeling (ERM), and includes practical advice on managing database projects.· Additional ORM models, course notes, and add-ins available online.

Bioinformatics

Life science data integration and interoperability is one of the most challenging problems facing bioinformatics today. In the current age of the life sciences, investigators have to interpret many types of information from a variety of sources: lab instruments, public databases, gene expression profiles, raw sequence traces, single nucleotide polymorphisms, chemical screening data, proteomic data, putative metabolic pathway models, and many others. Unfortunately, scientists are not currently able to easily identify and access this information because of the variety of semantics, interfaces, and data formats used by the underlying data sources. Bioinformatics: Managing Scientific Data tackles this challenge head-on by discussing the current approaches and variety of systems available to help bioinformaticians with this increasingly complex issue. The heart of the book lies in the collaboration efforts of eight distinct bioinformatics teams that describe their own unique approaches to data integration and interoperability. Each system receives its own chapter where the lead contributors provide precious insight into the specific problems being addressed by the system, why the particular architecture was chosen, and details on the system's strengths and weaknesses. In closing, the editors provide important criteria for evaluating these systems that bioinformatics professionals will find valuable. * Provides a clear overview of the state-of-the-art in data integration and interoperability in genomics, highlighting a variety of systems and giving insight into the strengths and weaknesses of their different approaches. * Discusses shared vocabulary, design issues, complexity of use cases, and the difficulties of transferring existing data management approaches to bioinformatics systems, which serves to connect computer and life scientists. * Written by the primary contributors of eight reputable bioinformatics systems in academia and industry including: BioKris, TAMBIS, K2, GeneExpress, P/FDM, MBM, SDSC, SRS, and DiscoveryLink.

BLAST

Sequence similarity is a powerful tool for discovering biological function. Just as the ancient Greeks used comparative anatomy to understand the human body and linguists used the Rosetta stone to decipher Egyptian hieroglyphs, today we can use comparative sequence analysis to understand genomes. BLAST (Basic Local Alignment Search Tool), is a sophisticated software package for rapid searching of nucleotide and protein databases. It is one of the most important software packages used in sequence analysis and bioinformatics. Most users of BLAST, however, seldom move beyond the program's default parameters, and never take advantage of its full power. BLAST is the only book completely devoted to this popular suite of tools. It offers biologists, computational biology students, and bioinformatics professionals a clear understanding of BLAST as well as the science it supports. This book shows you how to move beyond the default parameters, get specific answers using BLAST, and how to interpret your results. The book also contains tutorial and reference sections covering NCBI-BLAST and WU-BLAST, background material to help you understand the statistics behind BLAST, Perl scripts to help you prepare your data and analyze your results, and a wealth of tips and tricks for configuring BLAST to meet your own research needs. Some of the topics covered include: BLAST basics and the NCBI web interface How to select appropriate search parameters BLAST programs: BLASTN, BLASTP, BLASTX, TBLASTN, TBLASTX, PHI-BLAST, and PSI BLAST Detailed BLAST references, including NCBI-BLAST and WU-BLAST Understanding biological sequences Sequence similarity, homology, scoring matrices, scores, and evolution Sequence Alignment Calculating BLAST statistics Industrial-strength BLAST, including developing applications with Perl and BLAST BLAST is the only comprehensive reference with detailed, accurate information on optimizing BLAST searches for high-throughput sequence analysis. This is a book that any biologist should own.

Practical RDF

The Resource Description Framework (RDF) is a structure for describing and interchanging metadata on the Web--anything from library catalogs and worldwide directories to bioinformatics, Mozilla internal data structures, and knowledge bases for artificial intelligence projects. RDF provides a consistent framework and syntax for describing and querying data, making it possible to share website descriptions more easily. RDF's capabilities, however, have long been shrouded by its reputation for complexity and a difficult family of specifications. Practical RDF breaks through this reputation with immediate and solvable problems to help you understand, master, and implement RDF solutions. Practical RDF explains RDF from the ground up, providing real-world examples and descriptions of how the technology is being used in applications like Mozilla, FOAF, and Chandler, as well as infrastructure you can use to build your own applications. This book cuts to the heart of the W3C's often obscure specifications, giving you tools to apply RDF successfully in your own projects.The first part of the book focuses on the RDF specifications. After an introduction to RDF, the book covers the RDF specification documents themselves, including RDF Semantics and Concepts and Abstract Model specifications, RDF constructs, and the RDF Schema. The second section focuses on programming language support, and the tools and utilities that allow developers to review, edit, parse, store, and manipulate RDF/XML. Subsequent sections focus on RDF's data roots, programming and framework support, and practical implementation and use of RDF and RDF/XML.If you want to know how to apply RDF to information processing, Practical RDF is for you. Whether your interests lie in large-scale information aggregation and analysis or in smaller-scale projects like weblog syndication, this book will provide you with a solid foundation for working with RDF.

Up and Running with DB2 UDB ESE: Partitioning for Performance in an e-Business Intelligence World

Data warehouses in the 1990s were for the privileged few business analysts. Business Intelligence is now being democratized by being shared with the rank and file employee demanding higher levels of RDBMS scalability and ease of use, being delivered through Web portals. To support this emerging e-Business Intelligence world, the challenges that face the enterprises for their centralized data warehouse RDBMS technology are scalability, performance, availability and smart manageability. This IBM Redbooks publication focuses on the innovative technical functionalities of DB2 UDB ESE V8.1 and discusses: This book positions the new functionalities, so you can understand and evaluate their applicability in your own enterprise data warehouse environment, and get started prioritizing and implementing them. Please note that the additional material referenced in the text is not available from IBM.

Service- and Component-based Development: Using Select Perspective™ and UML

This book presents the approaches and practices for the construction of software systems using Select Perspective. It details the key workflows for a contemporary approach to supplying, managing, and consuming software assets to deliver business IT systems. This book provides a comprehensive development lifecycle (Select Perspective) for component and web service based IT solutions, that supports parallel development activities, to reduce time-to-market. It introduces the Select Perspective as a Supply, Manage, Consume software development process. It provides a real world project experience and examples. Technology: Unlike other development processes, Select Perspective is focused on a small number of key deliverables within an organizational framework of suppliers and consumers, of solution driven components. Audience: For CIOs, IT Directors, Project Managers, and solution developers. Level Intermediate. Hedley Apperly is Vice President, Product Marketing and Development with Aonix Hedley has graduate and post-graduate qualifications in production engineering, business computing and strategic marketing. His 19 years experience in IT, have focused on the design and development of relational, object-oriented and component-bases systems. He is also a committee member of the British Computer Societies (BCS), Object-Oriented Programming and Systems (OOPS) specialist group. As well as his involvement Component Based Development for Enterprise Systems, published by Cambridge University Press, Hedley co-authored Component Based Software Engineering; Putting the Pieces Together, published by Addison Wesley. Ralph Hofman works in Manager Services (Benelux) at Aonix. Ralph studied computer science at the University of Twente in the Netherlands. He started as a freelance consultant for different companies and specialized in methods and tools for system development. Ralph initiated component-based development as a way of working within a major international Bank. He joined Aonix in 2000, where he is responsible for the consultancy and services in the Benelux. Steve Latchem is Director of International Services with Aonix. Steve has been within the IT industry for over 18 years, holding positions in large consultancy groups and IT Departments ranging from business analyst to object oriented consultant, architect and project manager. Steve now directs the global professional services group at Aonix. Steve collaborated on AntiPatterns: Refactoring Software & Projects in Crisis and co-authored Component Based Software Engineering; Putting the Pieces Together, published by Addison Wesley. Barry Maybank is Principal Consultant with Aonix . Barry has been within the IT industry for over 17 years, holding positions in consultancy groups, IT Product Companies and Engineering companies with roles ranging from software engineer to architect. Barry McGibbon is Associate Consultant with Aonix. Barry has worked in the IT industry for over 35 years, holding very senior management positions with leading computing services providers. He has been involved in component-based development initiatives for significant enterprises in the UK and Europe. As well as a frequent contributor to major journals, he is author of Managing Your Move To Object Technology: Guidelines & Strategies for a Smooth Transition published by SIGS Books Inc. He is also Technical Chair for Europe's largest CBD/OO conference and a series editor for Cambridge University Press. David Piper is a Principal Consultant with Aonix. David has been working in the IT industry for over 20 years holding positions in manufacturing, financial services and IT consultancy with roles ranging from analyst to quality assurance manager and project manager. Chris Simons is a Senior Consultant with Aonix. Christopher has been within the IT industry for over 12 years, holding positions in real-time, defense, retail, public sector and finance domains, with roles ranging from software engineer, lead analyst to technical architect. He has also taught object-orientation and development process at various universities as a visiting lecturer.

Real R & D Options

Real R&D options are among the earliest modelled real options, with now ten primary practical uses: general R&D planning, planning R&D in stages, evaluating test information, new product development timing, operations, abandonment, risk sharing, market funding, industry strategy and regulation. This book was partly motivated by requests to identify and develop real option models for R&D in telecommunications, petroleum technology and biotechnology. Nine new models cover information and implementation costs, analytical solutions for mean reverting, or fat tailed revenues, endogenous learning and exogenous and experiential shocks, American sequential options, and innovator advantages. Four new applications include forward start development options, exploration options, innovation with information costs, and innovator's real values with changing market share. R&D directors and researchers will find several uses for these models: · general R&D planning · evaluating test information · new product development timing · risk sharing · industry strategy and regulation A practical guide to how organizations can use Real Option techniques to effectively value research and development by companies Provides a rigorous theoretical underpinning of the use of Real Option techniques *Real Options applications are orientated around the economies of North America, Europe and Asia, for an international perspective

Random Processes: Filtering, Estimation, and Detection

An understanding of random processes is crucial to many engineering fields-including communication theory, computer vision, and digital signal processing in electrical and computer engineering, and vibrational theory and stress analysis in mechanical engineering. The filtering, estimation, and detection of random processes in noisy environments are critical tasks necessary in the analysis and design of new communications systems and useful signal processing algorithms. Random Processes: Filtering, Estimation, and Detection clearly explains the basics of probability and random processes and details modern detection and estimation theory to accomplish these tasks. In this book, Lonnie Ludeman, an award-winning authority in digital signal processing, joins the fundamentals of random processes with the standard techniques of linear and nonlinear systems analysis and hypothesis testing to give signal estimation techniques, specify optimum estimation procedures, provide optimum decision rules for classification purposes, and describe performance evaluation definitions and procedures for the resulting methods. The text covers four main, interrelated topics: Probability and characterizations of random variables and random processes Linear and nonlinear systems with random excitations Optimum estimation theory including both the Wiener and Kalman Filters Detection theory for both discrete and continuous time measurements Lucid, thorough, and well-stocked with numerous examples and practice problems that emphasize the concepts discussed, Random Processes: Filtering, Estimation, and Detection is an understandable and useful text ideal as both a self-study guide for professionals in the field and as a core text for graduate students.

Process Control: Modeling, Design, and Simulation

Master process control hands on, through practical examples and MATLAB® simulations This is the first complete introduction to process control that fully integrates software tools—enabling professionals and students to master critical techniques hands on, through computer simulations based on the popular MATLAB environment. Process Control: Modeling, Design, and Simulation teaches the field's most important techniques, behaviors, and control problems through practical examples, supplemented by extensive exercises—with detailed derivations, relevant software files, and additional techniques available on a companion Web site. Coverage includes: Fundamentals of process control and instrumentation, including objectives, variables, and block diagrams Methodologies for developing dynamic models of chemical processes Dynamic behavior of linear systems: state space models, transfer function-based models, and more Feedback control; proportional, integral, and derivative (PID) controllers; and closed-loop stability analysis Frequency response analysis techniques for evaluating the robustness of control systems Improving control loop performance: internal model control (IMC), automatic tuning, gain scheduling, and enhancements to improve disturbance rejection Split-range, selective, and override strategies for switching among inputs or outputs Control loop interactions and multivariable controllers An introduction to model predictive control (MPC) Bequette walks step by step through the development of control instrumentation diagrams for an entire chemical process, reviewing common control strategies for individual unit operations, then discussing strategies for integrated systems. The book also includes 16 learning modules demonstrating how to use MATLAB and SIMULINK to solve several key control problems, ranging from robustness analyses to biochemical reactors, biomedical problems to multivariable control.

Enhance Your Business Applications: Simple Integration of Advanced Data Mining Functions

Today data mining is no longer thought of as a set of stand-alone techniques, far from the business applications, and used only by data mining specialists or statisticians. Integrating data mining with mainstream applications is becoming an important issue for e-business applications. To support this move to applications, data mining is now an extension of the relational databases that database administrators or IT developers use. They use data mining as they would use any other standard relational function that they manipulate. This IBM Redbooks publication positions the new DB2 data mining functions: Part 1 of this book helps business analysts and implementers to understand and position these new DB2 data mining functions. Part 2 provides examples for implementers on how to easily and quickly integrate the data mining functions in business applications to enhance them. And part 3 helps database administrators and IT developers to configure these functions once to prepare them for use and integration in any application. Please note that the additional material referenced in the text is not available from IBM.

Mining the Web

Mining the Web: Discovering Knowledge from Hypertext Data is the first book devoted entirely to techniques for producing knowledge from the vast body of unstructured Web data. Building on an initial survey of infrastructural issues—including Web crawling and indexing—Chakrabarti examines low-level machine learning techniques as they relate specifically to the challenges of Web mining. He then devotes the final part of the book to applications that unite infrastructure and analysis to bring machine learning to bear on systematically acquired and stored data. Here the focus is on results: the strengths and weaknesses of these applications, along with their potential as foundations for further progress. From Chakrabarti's work—painstaking, critical, and forward-looking—readers will gain the theoretical and practical understanding they need to contribute to the Web mining effort. * A comprehensive, critical exploration of statistics-based attempts to make sense of Web Mining. * Details the special challenges associated with analyzing unstructured and semi-structured data. * Looks at how classical Information Retrieval techniques have been modified for use with Web data. * Focuses on today's dominant learning methods: clustering and classification, hyperlink analysis, and supervised and semi-supervised learning. * Analyzes current applications for resource discovery and social network analysis. * An excellent way to introduce students to especially vital applications of data mining and machine learning technology.

Inclusion Breakthrough

Constant, continuing, and cataclysmic change is causing a major crisis within business organizations today. Faced with constantly advancing technology, unpredictable market shifts, intense global competition, and an increasingly independent "free agent" workforce, the only way for an organization to adapt and succeed is to build a "culture of inclusion" that nurtures and draws on the talents of a diverse workforce. Easy to say but hard to do; most organizations are mired in industrial revolution, static-world business models administered by monocultural, bordering-on-oppressive, "command and control" hierarchies. Organizations at risk include Fortune 500 giants, entrepreneurial start-ups, manufacturing and retail operations, government agencies, not-for-profits, educational institutions, and others. Most organizational change efforts-whether labeled as diversity efforts, re-engineering, right-sizing, or total-quality-management-are a waste of time, money, and human effort. Most produce more cynicism than results, and they can poison the waters for future change efforts. The Inclusion Breakthrough cuts a path through this potential minefield, offering a proven methodology for strategic organizational change, including models for diagnosing, planning, and implementing inclusion-focused, culture-change strategies tailored to each organization's individual needs. It also describes the key competencies for leading and sustaining a culture of inclusion. Offering real-world results of "before and after" surveys, including anecdotal and statistical reports of organizational change achieved using the methodologies described, The Inclusion Breakthrough presents an overview of current workplace conditions, attitudes, and policies based on interviews, surveys, and focus groups encompassing thousands of people in major organizations. The Inclusion Breakthrough demonstrates why the bottom line must be the central focus of any change strategy-and more importantly, how to carry that strategy out successfully.

SAS for Linear Models, Fourth Edition

This clear and comprehensive guide provides everything you need for powerful linear model analysis. Using a tutorial approach and plenty of examples, authors Ramon Littell, Walter Stroup, and Rudolf Freund lead you through methods related to analysis of variance with fixed and random effects. You will learn to use the appropriate SAS procedure for most experiment designs (including completely random, randomized blocks, and split plot) as well as factorial treatment designs and repeated measures. SAS for Linear Models, Fourth Edition, also includes analysis of covariance, multivariate linear models, and generalized linear models for non-normal data. Find inside: regression models; balanced ANOVA with both fixed- and random-effects models; unbalanced data with both fixed- and random-effects models; covariance models; generalized linear models; multivariate models; and repeated measures. New in this edition: MIXED and GENMOD procedures, updated examples, new software-related features, and other new material.

This book is part of the SAS Press program.

The Boost Graph Library: User Guide and Reference Manual

The Boost Graph Library (BGL) is the first C++ library to apply the principles of generic programming to the construction of the advanced data structures and algorithms used in graph computations. Problems in such diverse areas as Internet packet routing, molecular biology, scientific computing, and telephone network design can be solved by using graph theory. This book presents an in-depth description of the BGL and provides working examples designed to illustrate the application of BGL to these real-world problems. Written by the BGL developers, gives you all the information you need to take advantage of this powerful new library. Part I is a complete user guide that begins by introducing graph concepts, terminology, and generic graph algorithms. This guide also takes the reader on a tour through the major features of the BGL; all motivated with example problems. Part II is a comprehensive reference manual that provides complete documentation of all BGL concepts, algorithms, and classes. The Boost Graph Library: User Guide and Reference Manual Readers will find coverage of: Graph terminology and concepts Generic programming techniques in C++ Shortest-path algorithms for Internet routing Network planning problems using the minimum-spanning tree algorithms BGL algorithms with implicitly defined graphs BGL Interfaces to other graph libraries BGL concepts and algorithms BGL classes–graph, auxiliary, and adaptor Groundbreaking in its scope, this book offers the key to unlocking the power of the BGL for the C++ programmer looking to extend the reach of generic programming beyond the Standard Template Library.