talk-data.com talk-data.com

Topic

data-science-tasks

849

tagged

Activity Trend

1 peak/qtr
2020-Q1 2026-Q1

Activities

849 activities · Newest first

Advanced Audio Visualization Using ThMAD: Creating Amazing Graphics with Open Source Software

Learn advanced techniques and improve your audio visualization skills with Thinking Machine Audio Dreams (ThMAD). With this book, you can concentrate on advanced examples and usage patterns, including using shaders in a more profound way, and how to incorporate ThMAD into a tool chain using the professional sound server JACK. Advanced Audio Visualization Using ThMAD provides advanced techniques for generating graphics, improving performance, and providing readers with the skills needed to create more interesting visualizations. You will also learn professional setups with highly developed visual and aural art tool chains. What You'll Learn Use the ThMAD software for advanced setups in their personal and professional projects Gain a pragmatic introduction to using shaders Use JACK sound servers with ThMAD Control the timing ThMAD Work withadvanced configurations Who This Book Is For Artists and developers already familiar with ThMAD and looking to enhance their projects. In addition, readers primarily interested in using shaders or the Jack audio server for graphics generation can benefit from the book as well.

Mathematical Statistics

Explores mathematical statistics in its entirety—from the fundamentals to modern methods This book introduces readers to point estimation, confidence intervals, and statistical tests. Based on the general theory of linear models, it provides an in-depth overview of the following: analysis of variance (ANOVA) for models with fixed, random, and mixed effects; regression analysis is also first presented for linear models with fixed, random, and mixed effects before being expanded to nonlinear models; statistical multi-decision problems like statistical selection procedures (Bechhofer and Gupta) and sequential tests; and design of experiments from a mathematical-statistical point of view. Most analysis methods have been supplemented by formulae for minimal sample sizes. The chapters also contain exercises with hints for solutions. Translated from the successful German text, Mathematical Statistics requires knowledge of probability theory (combinatorics, probability distributions, functions and sequences of random variables), which is typically taught in the earlier semesters of scientific and mathematical study courses. It teaches readers all about statistical analysis and covers the design of experiments. The book also describes optimal allocation in the chapters on regression analysis. Additionally, it features a chapter devoted solely to experimental designs. Classroom-tested with exercises included Practice-oriented (taken from day-to-day statistical work of the authors) Includes further studies including design of experiments and sample sizing Presents and uses IBM SPSS Statistics 24 for practical calculations of data Mathematical Statistics is a recommended text for advanced students and practitioners of math, probability, and statistics.

Relational Research and Organisation Studies

Relational Research and Organisation Studies does not only present and discuss guidelines for practice at a onto-epistemological level but also presents and discusses concrete cases of research projects building on relational constructionist ideas. Furthermore, excerpts of data are presented and analyzed in order to explain the co-constructed processes of the inquiries more in detail.

SAS for Forecasting Time Series, Third Edition, 3rd Edition

To use statistical methods and SAS applications to forecast the future values of data taken over time, you need only follow this thoroughly updated classic on the subject. With this third edition of SAS for Forecasting Time Series, intermediate-to-advanced SAS users—such as statisticians, economists, and data scientists—can now match the most sophisticated forecasting methods to the most current SAS applications. Starting with fundamentals, this new edition presents methods for modeling both univariate and multivariate data taken over time. From the well-known ARIMA models to unobserved components, methods that span the range from simple to complex are discussed and illustrated. Many of the newer methods are variations on the basic ARIMA structures. Completely updated, this new edition includes fresh, interesting business situations and data sets, and new sections on these up-to-date statistical methods: ARIMA models Vector autoregressive models Exponential smoothing models Unobserved component and state-space models Seasonal adjustment Spectral analysis Focusing on application, this guide teaches a wide range of forecasting techniques by example. The examples provide the statistical underpinnings necessary to put the methods into practice. The following up-to-date SAS applications are covered in this edition: The ARIMA procedure The AUTOREG procedure The VARMAX procedure The ESM procedure The UCM and SSM procedures The X13 procedure The SPECTRA procedure SAS Forecast Studio Each SAS application is presented with explanation of its strengths, weaknesses, and best uses. Even users of automated forecasting systems will benefit from this knowledge of what is done and why. Moreover, the accompanying examples can serve as templates that you easily adjust to fit your specific forecasting needs. This book is part of the SAS Press program.

The Routledge Companion to Trust

Comprising six thematic sections, The Routledge Companion to Trust provides a comprehensive survey of trust research. With contributions from international experts, this volume examines the major topics and emerging areas within the field, including essays on the foundations, levels and theories of trust.

Python Web Scraping Cookbook

Python Web Scraping Cookbook is your comprehensive guide to building efficient and functional web scraping tools using Python. With practical recipes, you'll learn to overcome the challenges of dynamic content, captcha, and irregular web structures while deploying scalable solutions. What this Book will help me do Master the use of Python libraries like BeautifulSoup and Scrapy for scraping data. Perfect techniques for handling JavaScript-heavy sites using Selenium. Learn to overcome web scraping challenges, such as captchas and rate-limiting. Design scalable scraping pipelines with cloud deployment in AWS. Understand web data extraction techniques with XPath, CSS selectors, and more. Author(s) Michael Heydt is a seasoned software engineer and technical author with a focus on data engineering and cloud solutions. Having worked with Python extensively, he brings real-world insights into web scraping. His practical approach simplifies complex concepts. Who is it for? This book is perfect for Python developers and data enthusiasts keen to master web scraping techniques. If you're a programmer with insights into Python scripting and wish to scrape, analyze, and utilize web data efficiently, this book is for you.

An Introduction to Discrete-Valued Time Series

A much-needed introduction to the field of discrete-valued time series, with a focus on count-data time series Time series analysis is an essential tool in a wide array of fields, including business, economics, computer science, epidemiology, finance, manufacturing and meteorology, to name just a few. Despite growing interest in discrete-valued time series—especially those arising from counting specific objects or events at specified times—most books on time series give short shrift to that increasingly important subject area. This book seeks to rectify that state of affairs by providing a much needed introduction to discrete-valued time series, with particular focus on count-data time series. The main focus of this book is on modeling. Throughout numerous examples are provided illustrating models currently used in discrete-valued time series applications. Statistical process control, including various control charts (such as cumulative sum control charts), and performance evaluation are treated at length. Classic approaches like ARMA models and the Box-Jenkins program are also featured with the basics of these approaches summarized in an Appendix. In addition, data examples, with all relevant R code, are available on a companion website. Provides a balanced presentation of theory and practice, exploring both categorical and integer-valued series Covers common models for time series of counts as well as for categorical time series, and works out their most important stochastic properties Addresses statistical approaches for analyzing discrete-valued time series and illustrates their implementation with numerous data examples Covers classical approaches such as ARMA models, Box-Jenkins program and how to generate functions Includes dataset examples with all necessary R code provided on a companion website An Introduction to Discrete-Valued Time Series is a valuable working resource for researchers and practitioners in a broad range of fields, including statistics, data science, machine learning, and engineering. It will also be of interest to postgraduate students in statistics, mathematics and economics.

Loss Data Analysis

This volume deals with two complementary topics. On one hand the book deals with the problem of determining the the probability distribution of a positive compound random variable, a problem which appears in the banking and insurance industries, in many areas of operational research and in reliability problems in the engineering sciences. On the other hand, the methodology proposed to solve such problems, which is based on an application of the maximum entropy method to invert the Laplace transform of the distributions, can be applied to many other problems. The book contains applications to a large variety of problems, including the problem of dependence of the sample data used to estimate empirically the Laplace transform of the random variable. Contents Introduction Frequency models Individual severity models Some detailed examples Some traditional approaches to the aggregation problem Laplace transforms and fractional moment problems The standard maximum entropy method Extensions of the method of maximum entropy Superresolution in maxentropic Laplace transform inversion Sample data dependence Disentangling frequencies and decompounding losses Computations using the maxentropic density Review of statistical procedures

Regression Analysis with R

Dive into the world of regression analysis with this hands-on guide that covers everything you need to know about building effective regression models in R. You'll learn both the theoretical foundations and how to apply them using practical examples and R code. By the end, you'll be equipped to interpret regression results and use them to make meaningful predictions. What this Book will help me do Master the fundamentals of regression analysis, from simple linear to logistic regression. Gain expertise in R programming for implementing regression models and analyzing results. Develop skills in handling missing data, feature engineering, and exploratory data analysis. Understand how to identify, prevent, and address overfitting and underfitting issues in modeling. Apply regression techniques in real-world applications, including classification problems and advanced methods like Bagging and Boosting. Author(s) Giuseppe Ciaburro is an experienced data scientist and author with a passion for making complex technical topics accessible. With expertise in R programming and regression analysis, he has worked extensively in statistical modeling and data exploration. Giuseppe's writing combines clear explanations of theory with hands-on examples, ideal for learners and practitioners alike. Who is it for? This book is perfect for aspiring data scientists and analysts eager to understand and apply regression analysis using R. It's suited for readers with a foundational knowledge of statistics and basic R programming experience. Whether you're delving into data science or aiming to strengthen existing skills, this book offers practical insights to reach your goals.

Complex Network Analysis in Python

Construct, analyze, and visualize networks with networkx, a Python language module. Network analysis is a powerful tool you can apply to a multitude of datasets and situations. Discover how to work with all kinds of networks, including social, product, temporal, spatial, and semantic networks. Convert almost any real-world data into a complex network--such as recommendations on co-using cosmetic products, muddy hedge fund connections, and online friendships. Analyze and visualize the network, and make business decisions based on your analysis. If you're a curious Python programmer, a data scientist, or a CNA specialist interested in mechanizing mundane tasks, you'll increase your productivity exponentially. Complex network analysis used to be done by hand or with non-programmable network analysis tools, but not anymore! You can now automate and program these tasks in Python. Complex networks are collections of connected items, words, concepts, or people. By exploring their structure and individual elements, we can learn about their meaning, evolution, and resilience. Starting with simple networks, convert real-life and synthetic network graphs into networkx data structures. Look at more sophisticated networks and learn more powerful machinery to handle centrality calculation, blockmodeling, and clique and community detection. Get familiar with presentation-quality network visualization tools, both programmable and interactive--such as Gephi, a CNA explorer. Adapt the patterns from the case studies to your problems. Explore big networks with NetworKit, a high-performance networkx substitute. Each part in the book gives you an overview of a class of networks, includes a practical study of networkx functions and techniques, and concludes with case studies from various fields, including social networking, anthropology, marketing, and sports analytics. Combine your CNA and Python programming skills to become a better network analyst, a more accomplished data scientist, and a more versatile programmer. What You Need: You will need a Python 3.x installation with the following additional modules: Pandas (>=0.18), NumPy (>=1.10), matplotlib (>=1.5), networkx (>=1.11), python-louvain (>=0.5), NetworKit (>=3.6), and generalizesimilarity. We recommend using the Anaconda distribution that comes with all these modules, except for python-louvain, NetworKit, and generalizedsimilarity, and works on all major modern operating systems.

Statistical Rethinking

Statistical Rethinking: A Bayesian Course with Examples in R and Stan builds readers’ knowledge of and confidence in statistical modeling. Reflecting the need for even minor programming in today’s model-based statistics, the book pushes readers to perform step-by-step calculations that are usually automated. This unique computational approach ensures that readers understand enough of the details to make reasonable choices and interpretations in their own modeling work. The text presents generalized linear multilevel models from a Bayesian perspective, relying on a simple logical interpretation of Bayesian probability and maximum entropy. It covers from the basics of regression to multilevel models. The author also discusses measurement error, missing data, and Gaussian process models for spatial and network autocorrelation. By using complete R code examples throughout, this book provides a practical foundation for performing statistical inference. Designed for both PhD students and seasoned professionals in the natural and social sciences, it prepares them for more advanced or specialized statistical modeling. Web Resource The book is accompanied by an R package (rethinking) that is available on the author’s website and GitHub. The two core functions (map and map2stan) of this package allow a variety of statistical models to be constructed from standard model formulas.

IBM SPSS Modeler Essentials

Learn how to leverage IBM SPSS Modeler for your data mining and predictive analytics needs in this comprehensive guide. With step-by-step instructions, you'll acquire the skills to import, clean, analyze, and model your data using this robust platform. By the end, you'll be equipped to uncover patterns and trends, enabling data-driven decision-making confidently. What this Book will help me do Understand the fundamentals of data mining and the visual programming interface of IBM SPSS Modeler. Prepare, clean, and preprocess data effectively for analysis and modeling. Build robust predictive models such as decision trees using best practices. Evaluate the performance of your analytical models to ensure accuracy and reliability. Export resulting analyses to apply insights to real-world data projects. Author(s) Keith McCormick and Jesus Salcedo are accomplished professionals in data analytics and statistical modeling. With extensive experience in consulting and teaching, they have guided many in mastering IBM SPSS Modeler through both hands-on workshops and written material. Their approachable teaching style and commitment to clarity ensure accessibility for learners. Who is it for? This book is designed for beginner users of IBM SPSS Modeler who wish to gain practical and actionable skills in data analytics. If you're a data enthusiast looking to explore predictive analytics or a professional eager to discover the insights hidden in your organizational data, this book is for you. A basic understanding of data mining concepts is advantageous but not required. This resource will set any novice on the path toward expert-level comprehension and application.

Making Data Visual

You have a mound of data front of you and a suite of computation tools at your disposal. Which parts of the data actually matter? Where is the insight hiding? If you’re a data scientist trying to navigate the murky space between data and insight, this practical book shows you how to make sense of your data through high-level questions, well-defined data analysis tasks, and visualizations to clarify understanding and gain insights along the way. When incorporated into the process early and often, iterative visualization can help you refine the questions you ask of your data. Authors Danyel Fisher and Miriah Meyer provide detailed case studies that demonstrate how this process can evolve in the real world. You’ll learn: The data counseling process for moving from general to more precise questions about your data, and arriving at a working visualization The role that visual representations play in data discovery Common visualization types by the tasks they fulfill and the data they use Visualization techniques that use multiple views and interaction to support analysis of large, complex data sets

Data Mining Algorithms in C++: Data Patterns and Algorithms for Modern Applications

Discover hidden relationships among the variables in your data, and learn how to exploit these relationships. This book presents a collection of data-mining algorithms that are effective in a wide variety of prediction and classification applications. All algorithms include an intuitive explanation of operation, essential equations, references to more rigorous theory, and commented C++ source code. Many of these techniques are recent developments, still not in widespread use. Others are standard algorithms given a fresh look. In every case, the focus is on practical applicability, with all code written in such a way that it can easily be included into any program. The Windows-based DATAMINE program lets you experiment with the techniques before incorporating them into your own work. What You'll Learn Use Monte-Carlo permutation tests to provide statistically sound assessments of relationships present in your data Discover how combinatorially symmetric cross validation reveals whether your model has true power or has just learned noise by overfitting the data Work with feature weighting as regularized energy-based learning to rank variables according to their predictive power when there is too little data for traditional methods See how the eigenstructure of a dataset enables clustering of variables into groups that exist only within meaningful subspaces of the data Plot regions of the variable space where there is disagreement between marginal and actual densities, or where contribution to mutual information is high Who This Book Is For Anyone interested in discovering and exploiting relationships among variables. Although all code examples are written in C++, the algorithms are described in sufficient detail that they can easily be programmed in any language.

D3.js in Action, Second Edition

D3.js in Action, Second Edition is completely revised and updated for D3 v4 and ES6. It's a practical tutorial for creating interactive graphics and data-driven applications using D3. About the Technology Visualizing complex data is hard. Visualizing complex data on the web is darn near impossible without D3.js. D3 is a JavaScript library that provides a simple but powerful data visualization API over HTML, CSS, and SVG. Start with a structure, dataset, or algorithm; mix in D3; and you can programmatically generate static, animated, or interactive images that scale to any screen or browser. It's easy, and after a little practice, you'll be blown away by how beautiful your results can be! About the Book D3.js in Action, Second Edition is a completely updated revision of Manning's bestselling guide to data visualization with D3. You'll explore dozens of real-world examples in full-color, including force and network diagrams, workflow illustrations, geospatial constructions, and more! Along the way, you'll pick up best practices for building interactive graphics, animations, and live data representations. You'll also step through a fully interactive application created with D3 and React. What's Inside Rich full-color diagrams and illustrations Updated for D3 v4 and ES6 Reusable layouts and components Geospatial data visualizations Mixed-mode rendering About the Reader Suitable for web developers with HTML, CSS, and JavaScript skills. No specialized data science skills required. About the Author Elijah Meeks is a senior data visualization engineer at Netflix. Quotes From basic to complex, this book gives you the tools to create beautiful data visualizations. - Claudio Rodriguez, Cox Media Group The best reference for one of the most useful DataViz tools. - Jonathan Rioux, TD Insurance From toy examples to techniques for real projects. Shows how all the pieces fit together. - Scott McKissock, USAID A clever way to immerse yourself in the D3.js world. - Felipe Vildoso Castillo, University of Chile

Learning D3.js 5 Mapping - Second Edition

This book, "Learning D3.js 5 Mapping", guides developers through the process of creating dynamic and interactive data visualizations. With a focus on D3.js, you'll learn to harness the power of JavaScript to create maps and graphical objects that inform and engage. What this Book will help me do Gain expertise in working with SVG geometric shapes to design compelling graphics. Learn techniques to manage, process, and use geographic data effectively. Master adding interactivity to visual maps to provide an immersive user experience. Understand how to optimize and manipulate geoJSON files using topoJSON. Learn to create varied map types, such as hexbins and globes, using D3.js and Canvas. Author(s) None Newton and Oscar Villarreal, among others, collaborated to author this guide. They are experienced in front-end development and data visualization, bringing a practical and hands-on approach to learning through this book. Their backgrounds ensure the book addresses common challenges faced during implementation, offering thoughtful solutions. Who is it for? "Learning D3.js 5 Mapping" is perfect for web developers familiar with HTML, CSS, and JavaScript who want to expand their expertise in data visualization and mapping. If you're looking to incorporate interactive charts or maps into your web applications, this book will provide practical guidance and solid fundamentals. No prior experience with D3.js is necessary.

Analyzing Multidimensional Well-Being

“An indispensable reference for all researchers interested in the measurement of social welfare. . .” —François Bourguignon, Emeritus Professor at Paris School of Economics, Former Chief Economist of the World Bank. “. . .a detailed, insightful, and pedagogical presentation of the theoretical grounds of multidimensional well-being, inequality, and poverty measurement. Any student, researcher, and practitioner interested in the multidimensional approach should begin their journey into such a fascinating theme with this wonderful book.” —François Maniquet, Professor, Catholic University of Louvain, Belgium. A Review of the Multidimensional Approaches to the Measurement of Welfare, Inequality, and Poverty Analyzing Multidimensional Well-Being: A Quantitative Approach offers a comprehensive approach to the measurement of well-being that includes characteristics such as income, health, literacy, and housing. The author presents a systematic comparison of the alternative approaches to the measurement of multidimensional welfare, inequality, poverty, and vulnerability. The text contains real-life applications of some multidimensional aggregations (most of which have been designed by international organizations such as the United Nations Development Program and the Organization for Economic Co-operation and Development) that help to judge the performance of a country in the various dimensions of well-being. The text offers an evaluation of how well a society is doing with respect to achievements of all the individuals in the dimensions considered and clearly investigates how achievements in the dimensions can be evaluated from different perspectives. The author includes a detailed scrutiny of alternative techniques for setting weights to individual dimensional metrics and offers an extensive analysis into both the descriptive and welfare theoretical approaches to the concerned multi-attribute measurement and related issues. This important resource: • Contains a synthesis of multidimensional welfare, inequality, poverty, and vulnerability analysis • Examines aggregations of achievement levels in the concerned dimensions of well-being from various standpoints • Shows how to measure poverty using panel data instead of restricting attention to a single period and when we have imprecise information on dimensional achievements • Argues that multidimensional analysis is intrinsically different from marginal distributions-based analysis Written for students, teachers, researchers, and scholars, Analyzing Multidimensional Well-Being: A Quantitative Approach puts the focus on various approaches to the measurementof the many aspects of well-being and quality of life. Satya R. Chakravarty is a Professor of Economics at the Indian Statistical Institute, Kolkata, India. He is an Editor of Social Choice and Welfare and a member of the Editorial Board of Journal of Economic Inequality.

Statistics for Data Science

Dive into the world of statistics specifically tailored for the needs of data science with 'Statistics for Data Science'. This book guides you from the fundamentals of statistical concepts to their practical application in data analysis, machine learning, and neural networks. Learn with clear explanations and practical R examples to fully grasp statistical methods for data-driven challenges. What this Book will help me do Understand foundational statistical concepts such as variance, standard deviation, and probability. Gain proficiency in using R for programmatically performing statistical computations for data science. Learn techniques for applying statistics in data cleaning, mining, and analysis tasks. Master methods for executing linear regression, regularization, and model assessment. Explore advanced techniques like boosting, SVMs, and neural network applications. Author(s) James D. Miller brings years of experience as a data scientist and educator. He has a deep understanding of how statistics foundationally supports data science and has worked across multiple industries applying these principles. Dedicated to teaching, James simplifies complex statistical concepts into approachable and actionable knowledge for developers aspiring to master data science applications. Who is it for? This book is intended for developers aiming to transition into the field of data science. If you have some basic programming knowledge and a desire to understand statistics essentials for data science applications, this book is designed for you. It's perfect for those who wish to apply statistical methods to practical tasks like data mining and analysis. A prior hands-on experience with R is helpful but not mandatory, as the book explains R methodologies comprehensively.

The State of Data Analytics and Visualization Adoption

Businesses regardless of industry or company size increasingly rely on data analytics and visualization to gain competitive advantage. That’s why organizations today are racing to gather, store, and analyze data from many sources in a wide range of formats. In the spring of 2017, Zoomdata commissioned an O’Reilly survey to assess the state of data analytics and visualization technology adoption across several industries, including manufacturing, financial services, and healthcare. Roughly 875 respondents answered questions online about their industry, job role, company size, and reasons for using analytics, as well as technologies they use in analytics programs, the perceived value of analytics programs, and many other topics. This report reveals: The industries furthest along in adopting big data analytics and visualization technologies The most commonly analyzed sources of big data The most commonly used technologies for analyzing streaming data Which analytics skills are in most demand The most valued characteristic of big data across all industries The types of users big data analytics and visualization projects typically target If you’re a technology decision maker, a product manager looking to embed analytics, a business user relying on analytics, or a developer pursuing the most marketable skills, this report provides valuable details on today’s data analytics trends.