talk-data.com talk-data.com

Topic

Analytics

data_analysis insights metrics

4552

tagged

Activity Trend

398 peak/qtr
2020-Q1 2026-Q1

Activities

4552 activities · Newest first

IBM z13 Technical Guide

Digital business has been driving the transformation of underlying IT infrastructure to be more efficient, secure, adaptive, and integrated. Information Technology (IT) must be able to handle the explosive growth of mobile clients and employees. IT also must be able to use enormous amounts of data to provide deep and real-time insights to help achieve the greatest business impact. This IBM® Redbooks® publication addresses the new IBM Mainframe, the IBM z13. The IBM z13 is the trusted enterprise platform for integrating data, transactions, and insight. A data-centric infrastructure must always be available with a 99.999% or better availability, have flawless data integrity, and be secured from misuse. It needs to be an integrated infrastructure that can support new applications. It needs to have integrated capabilities that can provide new mobile capabilities with real-time analytics delivered by a secure cloud infrastructure. IBM z13 is designed with improved scalability, performance, security, resiliency, availability, and virtualization. The superscalar design allows the z13 to deliver a record level of capacity over the prior z Systems. In its maximum configuration, z13 is powered by up to 141 client characterizable microprocessors (cores) running at 5 GHz. This configuration can run more than 110,000 millions of instructions per second (MIPS) and up to 10 TB of client memory. The IBM z13 Model NE1 is estimated to provide up to 40% more total system capacity than the IBM zEnterprise® EC12 (zEC1) Model HA1. This book provides information about the IBM z13 and its functions, features, and associated software support. Greater detail is offered in areas relevant to technical planning. It is intended for systems engineers, consultants, planners, and anyone who wants to understand the IBM z Systems functions and plan for their usage. It is not intended as an introduction to mainframes. Readers are expected to be generally familiar with existing IBM z Systems technology and terminology.

The Security Data Lake

Companies of all sizes are considering data lakes as a way to deal with terabytes of security data that can help them conduct forensic investigations and serve as an early indicator to identify bad or relevant behavior. Many think about replacing their existing SIEM (security information and event management) systems with Hadoop running on commodity hardware. Before your company jumps into the deep end, you first need to weigh several critical factors. This O'Reilly report takes you through technological and design options for implementing a data lake. Each option not only supports your data analytics use cases, but is also accessible by processes, workflows, third-party tools, and teams across your organization. Within this report, you'll explore: Five questions to ask before choosing architecture for your backend data store How data lakes can overcome scalability and data duplication issues Different options for storing context and unstructured log data Data access use cases covering both search and analytical queries via SQL Processes necessary for ingesting data into a data lake, including parsing, enrichment, and aggregation Four methods for embedding your SIEM into a data lake

podcast_episode
by Val Kroll , Julie Hoyer , Tim Wilson (Analytics Power Hour - Columbus (OH) , Moe Kiss (Canva) , Michael Helbling (Search Discovery)

Building your chops as an analyst is hard enough. Building an analyst team is even harder. In this episode, the three amigos of measurment share best practices in HR, Training, Management and Planning to help you in growing your own team. This episode will take 5 hours you say? Nope, we got it done in 40ish minutes.

Advanced Analytics with Spark

In this practical book, four Cloudera data scientists present a set of self-contained patterns for performing large-scale data analysis with Spark. The authors bring Spark, statistical methods, and real-world data sets together to teach you how to approach analytics problems by example. You’ll start with an introduction to Spark and its ecosystem, and then dive into patterns that apply common techniques—classification, collaborative filtering, and anomaly detection among others—to fields such as genomics, security, and finance. If you have an entry-level understanding of machine learning and statistics, and you program in Java, Python, or Scala, you’ll find these patterns useful for working on your own data applications.

IBM z13 Technical Introduction

This IBM® Redbooks® publication introduces the IBM z13™. IBM z13 delivers a data and transaction system reinvented as a system of insight for digital business. IBM z Systems™ leadership is extended with these features: Improved ability to meet service level agreements with new processor chip technology that includes simultaneous multithreading, analytical vector processing, redesigned and larger cache, and enhanced accelerators for hardware compression and cryptography Better availability and more efficient use of critical data with up to 10 TB available redundant array of independent memory (RAIM) Validation of transactions, management, and assignment of business priority for SAN devices through updates to the I/O subsystem Continued management of heterogeneous workloads with IBM z BladeCenter Extension (zBX) Model 004 and IBM z Unified Resource Manager This Redbooks publication can help you become familiar with the z Systems platform, and understand how the platform can help integrate data, transactions, and insight for faster and more accurate business decisions. This book explains how, with innovations and traditional strengths, IBM z13 can play an essential role in today's IT environments, and satisfy the demands for cloud deployments, analytics, mobile, and social applications in a trustful, reliable, and secure environment with operations that lessen business risk.

Fifteen years ago, digital analytics tooling was pretty straightforward (something that looks at log files). In 2015, there are literally hundreds of tools that can be used to measure every aspect of a digital sales and marketing ecosystem. Most companies still think “Google or Adobe?” when making a digital analytics tool purchase. Are they missing out? With very special guest Hiten Shah from KISSmetrics, Michael, Tim and Jim talk a little tooling and a lot of trash - in almost 60 minutes.

Data Mining and Predictive Analytics, 2nd Edition

Learn methods of data analysis and their application to real-world data sets This updated second edition serves as an introduction to data mining methods and models, including association rules, clustering, neural networks, logistic regression, and multivariate analysis. The authors apply a unified "white box" approach to data mining methods and models. This approach is designed to walk readers through the operations and nuances of the various methods, using small data sets, so readers can gain an insight into the inner workings of the method under review. Chapters provide readers with hands-on analysis problems, representing an opportunity for readers to apply their newly-acquired data mining expertise to solving real problems using large, real-world data sets. Data Mining and Predictive Analytics, Second Edition: Offers comprehensive coverage of association rules, clustering, neural networks, logistic regression, multivariate analysis, and R statistical programming language Features over 750 chapter exercises, allowing readers to assess their understanding of the new material Provides a detailed case study that brings together the lessons learned in the book Includes access to the companion website, www.dataminingconsultant, with exclusive password-protected instructor content Data Mining and Predictive Analytics, Second Edition will appeal to computer science and statistic students, as well as students in MBA programs, and chief executives.

Knowledge Discovery Process and Methods to Enhance Organizational Performance

Although the terms "data mining" and "knowledge discovery and data mining" (KDDM) are sometimes used interchangeably, data mining is actually just one step in the KDDM process. Data mining is the process of extracting useful information from data, while KDDM is the coordinated process of understanding the business and mining the data in order to identify previously unknown patterns. Knowledge Discovery Process and Methods to Enhance Organizational Performance explains the knowledge discovery and data mining (KDDM) process in a manner that makes it easy for readers to implement. Sharing the insights of international KDDM experts, it details powerful strategies, models, and techniques for managing the full cycle of knowledge discovery projects. The book supplies a process-centric view of how to implement successful data mining projects through the use of the KDDM process. It discusses the implications of data mining including security, privacy, ethical and legal considerations. Provides an introduction to KDDM, including the various models adopted in academia and industry Details critical success factors for KDDM projects as well as the impact of poor quality data or inaccessibility to data on KDDM projects Proposes the use of hybrid approaches that couple data mining with other analytic techniques (e.g., data envelopment analysis, cluster analysis, and neural networks) to derive greater value and utility Demonstrates the applicability of the KDDM process beyond analytics Shares experiences of implementing and applying various stages of the KDDM process in organizations The book includes case study examples of KDDM applications in business and government. After reading this book, you will understand the critical success factors required to develop robust data mining objectives that are in alignment with your organization’s strategic business objectives.

Mastering R for Quantitative Finance

Dive deeply into the quantitative finance domain using R with 'Mastering R for Quantitative Finance.' Through this book, you'll explore advanced R programming techniques tailored to financial modeling, risk assessment, and trading strategy optimization. This comprehensive guide aims to equip you with the tools to build practical quantitative finance solutions. What this Book will help me do Analyze detailed financial data using R and quantitative techniques. Develop predictive models for time series and risk management. Implement advanced trading strategies tailored to current market conditions. Master simulation techniques for scenarios without analytical solutions. Evaluate portfolio risks and potential returns with advanced methods. Author(s) None Gabler is a seasoned expert in quantitative finance and R programming, bringing years of practical experience to this book. Her approach combines theoretical depth with practical examples to ensure readers can apply the learned concepts in real-world financial contexts. Her passion for teaching and clear writing style make complex topics accessible to both practitioners and learners. Who is it for? This book is for financial professionals and data scientists seeking to delve into quantitative finance using R. Ideal readers are familiar with the basics of economics and statistics and are looking to apply advanced analytics in finance. If you are aiming to refine your modeling skills or develop precise strategies, this book is tailored for you. It's perfect for those eager to bridge the gap between theory and practical application.

Big Data

Convert the promise of big data into real world results There is so much buzz around big data. We all need to know what it is and how it works - that much is obvious. But is a basic understanding of the theory enough to hold your own in strategy meetings? Probably. But what will set you apart from the rest is actually knowing how to USE big data to get solid, real-world business results - and putting that in place to improve performance. Big Data will give you a clear understanding, blueprint, and step-by-step approach to building your own big data strategy. This is a well-needed practical introduction to actually putting the topic into practice. Illustrated with numerous real-world examples from a cross section of companies and organisations, Big Data will take you through the five steps of the SMART model: Start with Strategy, Measure Metrics and Data, Apply Analytics, Report Results, Transform. Discusses how companies need to clearly define what it is they need to know Outlines how companies can collect relevant data and measure the metrics that will help them answer their most important business questions Addresses how the results of big data analytics can be visualised and communicated to ensure key decisions-makers understand them Includes many high-profile case studies from the author's work with some of the world's best known brands

Apache Hive Essentials

Apache Hive Essentials is the perfect guide for understanding and mastering Hive, the SQL-like big data query language built on top of Hadoop. With this book, you will gain the skills to effectively use Hive to analyze and manage large data sets. Whether you're a developer, data analyst, or just curious about big data, this hands-on guide will enhance your capabilities. What this Book will help me do Understand the core concepts of Hive and its relation to big data and Hadoop. Learn how to set up a Hive environment and integrate it with Hadoop. Master the SQL-like query functionalities of Hive to select, manipulate, and analyze data. Develop custom functions in Hive to extend its functionality for your own specific use cases. Discover best practices for optimizing Hive performance and ensuring data security. Author(s) Dayong Du is an expert in big data analytics with extensive experience in implementing and using tools like Hive in professional settings. Having worked on practical big data solutions, Dayong brings a wealth of knowledge and insights to his writing. His clear, approachable style makes complex topics accessible to readers. Who is it for? This book is ideal for developers, data analysts, and data engineers looking to leverage Hive for big data analysis. If you are familiar with SQL and Hadoop basics and aim to enhance your understanding of Hive, this book is for you. Beginners with some programming background eager to dive into big data technologies will also benefit. It's tailored for learners wanting actionable knowledge to advance their data processing skills.

Apache Flume: Distributed Log Collection for Hadoop - Second Edition

"Apache Flume: Distributed Log Collection for Hadoop - Second Edition" is your hands-on guide to learning how to use Apache Flume to reliably collect and move logs and data streams into your Hadoop ecosystem. Through practical examples and real-world scenarios, this book will help you master the setup, configuration, and optimization of Flume for various data ingestion use cases. What this Book will help me do Understand the key concepts and architecture behind Apache Flume to build reliable and scalable data ingestion systems. Set up Flume agents to collect and transfer data into the Hadoop File System (HDFS) or other storage solutions effectively. Learn stream data processing techniques, such as filtering, transforming, and enriching data during transit to improve data usability. Integrate Flume with other tools like Elasticsearch and Solr to enhance analytics and search capabilities. Implement monitoring and troubleshooting workflows to maintain healthy and optimized Flume data pipelines. Author(s) Steven Hoffman, a seasoned software developer and data engineer, brings years of practical experience working with big data technologies to this book. He has a strong background in distributed systems and big data solutions, having implemented enterprise-scale analytics projects. Through clear and approachable writing, he aims to empower readers to successfully deploy reliable data pipelines using Apache Flume. Who is it for? This book is written for Hadoop developers, data engineers, and IT professionals who seek to build robust pipelines for streaming data into Hadoop environments. It is ideal for readers who have a basic understanding of Hadoop and HDFS but are new to Apache Flume. If you are looking to enhance your analytics capabilities by efficiently ingesting, routing, and processing streaming data, this book is for you. Beginners as well as experienced engineers looking to dive deeper into Flume will find it insightful.

Hadoop MapReduce v2 Cookbook - Second Edition

Explore insights from vast datasets with "Hadoop MapReduce v2 Cookbook - Second Edition." This book serves as a practical guide for developers and system administrators who aim to master big data processing using Hadoop v2. By engaging with its step-by-step recipes, you will learn to harness the Hadoop MapReduce ecosystem for scalable and efficient data solutions. What this Book will help me do Master the configuration and management of Hadoop YARN, MapReduce v2, and HDFS clusters. Integrate big data tools such as Hive, HBase, Pig, Mahout, and Nutch with Hadoop v2. Develop analytics solutions for large-scale datasets using MapReduce-based applications. Address specific challenges like data classification, recommendations, and text analytics leveraging Hadoop MapReduce. Deploy and manage big data clusters effectively, including options for cloud environments. Author(s) The authors behind "Hadoop MapReduce v2 Cookbook - Second Edition" combine their deep expertise in big data technology and years of experience working directly with Hadoop. They have helped numerous organizations implement scalable data processing solutions and are passionate about teaching others. Their approach ensures readers gain both foundational knowledge and practical skills. Who is it for? This book is perfect for developers and system administrators who want to learn Hadoop MapReduce v2, including configuring and managing big data clusters. Beginners with basic Java knowledge can follow along to advance their skills in big data processing. Ideal for those transitioning to Hadoop v2 or requiring practical recipes for immediate application. Great for professionals aiming to deepen their expertise in scalable data technologies.

NoSQL For Dummies

Get up to speed on the nuances of NoSQL databases and what they mean for your organization This easy to read guide to NoSQL databases provides the type of no-nonsense overview and analysis that you need to learn, including what NoSQL is and which database is right for you. Featuring specific evaluation criteria for NoSQL databases, along with a look into the pros and cons of the most popular options, NoSQL For Dummies provides the fastest and easiest way to dive into the details of this incredible technology. You'll gain an understanding of how to use NoSQL databases for mission-critical enterprise architectures and projects, and real-world examples reinforce the primary points to create an action-oriented resource for IT pros. If you're planning a big data project or platform, you probably already know you need to select a NoSQL database to complete your architecture. But with options flooding the market and updates and add-ons coming at a rapid pace, determining what you require now, and in the future, can be a tall task. This is where NoSQL For Dummies comes in! Learn the basic tenets of NoSQL databases and why they have come to the forefront as data has outpaced the capabilities of relational databases Discover major players among NoSQL databases, including Cassandra, MongoDB, MarkLogic, Neo4J, and others Get an in-depth look at the benefits and disadvantages of the wide variety of NoSQL database options Explore the needs of your organization as they relate to the capabilities of specific NoSQL databases Big data and Hadoop get all the attention, but when it comes down to it, NoSQL databases are the engines that power many big data analytics initiatives. With NoSQL For Dummies, you'll go beyond relational databases to ramp up your enterprise's data architecture in no time.

TIBCO Spotfire: A Comprehensive Primer

TIBCO Spotfire: A Comprehensive Primer is the go-to guide for mastering TIBCO Spotfire, a leading data visualization and analytics tool. Whether you are new to Spotfire or data visualization in general, this book will provide you with a solid foundation to create impactful and actionable visual insights. What this Book will help me do Understand the fundamentals of TIBCO Spotfire and its application in data analytics. Learn how to design compelling visualizations and dashboards that convey meaningful insights. Master advanced data transformations and analysis techniques in TIBCO Spotfire. Integrate Spotfire with external data sources and scripting languages, enhancing its functionality. Optimize Spotfire's performance and usability for enterprise-level implementations. Author(s) None Phillips, an experienced analytics professional and educator, specializes in creating accessible learning materials for data science tools. With a decade of experience in the field, None has helped many organizations unlock their data potential through tools like TIBCO Spotfire. Their approach emphasizes practical understanding, making complex concepts approachable for learners of all levels. Who is it for? The book is perfect for business analysts, data scientists, and other professionals involved in data-driven decision making who want to master TIBCO Spotfire. It's designed for beginners without prior exposure to data visualization or TIBCO Spotfire, offering an accessible entry into the field. Individuals aiming to gain hands-on experience and create enterprise-grade solutions will find this book invaluable. Additionally, it serves as a reference for experienced Spotfire users looking to refine their skills.

Learning Spark

Data in all domains is getting bigger. How can you work with it efficiently? Recently updated for Spark 1.3, this book introduces Apache Spark, the open source cluster computing system that makes data analytics fast to write and fast to run. With Spark, you can tackle big datasets quickly through simple APIs in Python, Java, and Scala. This edition includes new information on Spark SQL, Spark Streaming, setup, and Maven coordinates.

Data: Emerging Trends and Technologies

What are the emerging trends and technologies that will transform the data landscape in coming months? In this report from Strata + Hadoop World co-chair Alistair Croll, you'll learn how the ubiquity of cheap sensors, fast networks, and distributed computing have given rise to several developments that will soon have a profound effect on individuals and society as a whole. Machine learning, for example, has quickly moved from lab tool to hosted, pay-as-you-go services in the cloud. Those services, in turn, are leading to predictive apps that will provide individuals with the right functionality and content at the right time by continuously learning about them and predicting what they'll need. Computational power can produce cognitive augmentation. Report topics include: The swing between centralized and distributed computing Machine learning as a service Personal digital assistants and cognitive augmentation Graph databases and analytics Regulating complex algorithms The pace of real-time data and automation Solving dire problems with big data Implications of having sensors everywhere This report contains many more examples of how big data is starting to reshape business and change behavior, and it's just a small sample of the in-depth information Strata + Hadoop World provides. Pick up this report and make plans to attend one of several Strata + Hadoop World conferences in the San Francisco Bay Area, London, and New York.

Big Data Analytics

With this book, managers and decision makers are given the tools to make more informed decisions about big data purchasing initiatives. Big Data Analytics: A Practical Guide for Managers not only supplies descriptions of common tools, but also surveys the various products and vendors that supply the big data market. Comparing and contrasting the different types of analysis commonly conducted with big data, this accessible reference presents clear-cut explanations of the general workings of big data tools. Instead of spending time on HOW to install specific packages, it focuses on the reasons WHY readers would install a given package. The book provides authoritative guidance on a range of tools, including open source and proprietary systems. It details the strengths and weaknesses of incorporating big data analysis into decision-making and explains how to leverage the strengths while mitigating the weaknesses. Describes the benefits of distributed computing in simple terms Includes substantial vendor/tool material, especially for open source decisions Covers prominent software packages, including Hadoop and Oracle Endeca Examines GIS and machine learning applications Considers privacy and surveillance issues The book further explores basic statistical concepts that, when misapplied, can be the source of errors. Time and again, big data is treated as an oracle that discovers results nobody would have imagined. While big data can serve this valuable function, all too often these results are incorrect, yet are still reported unquestioningly. The probability of having erroneous results increases as a larger number of variables are compared unless preventative measures are taken. The approach taken by the authors is to explain these concepts so managers can ask better questions of their analysts and vendors as to the appropriateness of the methods used to arrive at a conclusion. Because the world of science and medicine has been grappling with similar issues in the publication of studies, the authors draw on their efforts and apply them to big data.

Practical Business Analytics Using SAS: A Hands-on Guide

Practical Business Analytics Using SAS: A Hands-on Guide shows SAS users and businesspeople how to analyze data effectively in real-life business scenarios. The book begins with an introduction to analytics, analytical tools, and SAS programming. The authors—both SAS, statistics, analytics, and big data experts—first show how SAS is used in business, and then how to get started programming in SAS by importing data and learning how to manipulate it. Besides illustrating SAS basic functions, you will see how each function can be used to get the information you need to improve business performance. Each chapter offers hands-on exercises drawn from real business situations. The book then provides an overview of statistics, as well as instruction on exploring data, preparing it for analysis, and testing hypotheses. You will learn how to use SAS to perform analytics and model using both basic and advanced techniques like multiple regression, logistic regression, and time series analysis, among other topics. The book concludes with a chapter on analyzing big data. Illustrations from banking and other industries make the principles and methods come to life. Readers will find just enough theory to understand the practical examples and case studies, which cover all industries. Written for a corporate IT and programming audience that wants to upgrade skills or enter the analytics field, this book includes: More than 200 examples and exercises, including code and datasets for practice. Relevant examples for all industries. Case studies that show how to use SAS analytics to identify opportunities, solve complicated problems, and chart a course. Practical Business Analytics Using SAS: A Hands-on Guide gives you the tools you need to gain insight into the data at your fingertips, predict business conditions for better planning, and make excellent decisions. Whether you are in retail, finance, healthcare, manufacturing, government, or any other industry, this book will help your organization increase revenue, drive down costs, improve marketing, and satisfy customers better than ever before.

To win at digital measurement in 2015, you need more data capture tools than just your web analytics tool of record. In episode 3, the 3 musketeers of measurement try to outline what they feel the core tools are that any digital analyst should be familiar with and thinking about. What do you need to have? What should you want to have? What should you be careful about? Get ready to have the hype separated from the important, and done so efficiently it will make an hour feel like 43 minutes.