Back to Courses

Data Analysis Courses - Page 24

Showing results 231-240 of 998
Data Mining Methods
This course covers the core techniques used in data mining, including frequent pattern analysis, classification, clustering, outlier analysis, as well as mining complex data and research frontiers in the data mining field. Data Mining Methods can be taken for academic credit as part of CU Boulder’s Master of Science in Data Science (MS-DS) degree offered on the Coursera platform. The MS-DS is an interdisciplinary degree that brings together faculty from CU Boulder’s departments of Applied Mathematics, Computer Science, Information Science, and others. With performance-based admissions and no application process, the MS-DS is ideal for individuals with a broad range of undergraduate education and/or professional experience in computer science, information science, mathematics, and statistics. Learn more about the MS-DS program at https://www.coursera.org/degrees/master-of-science-data-science-boulder. Course logo image courtesy of Lachlan Cormie, available here on Unsplash: https://unsplash.com/photos/jbJp18srifE
Data Warehousing with Microsoft Azure Synapse Analytics
In this course, you will explore the tools and techniques that can be used to work with Modern Data Warehouses productively and securely within Azure Synapse Analytics. You will learn how Azure Synapse Analytics enables you to build Data Warehouses using modern architecture patterns and how the common schema is implemented in a data warehouse. You'll learn the best practices you need to adopt to load data into a data warehouse and the techniques that you can use to optimize query performance within Azure Synapse Analytics. This course is part of a Specialization intended for Data engineers and developers who want to demonstrate their expertise in designing and implementing data solutions that use Microsoft Azure data services for anyone interested in preparing for the Exam DP-203: Data Engineering on Microsoft Azure (beta). This is the fifth course in a program of 10 courses to help prepare you to take the exam so that you can have expertise in designing and implementing data solutions that use Microsoft Azure data services. The Data Engineering on Microsoft Azure exam is an opportunity to prove knowledge expertise in integrating, transforming, and consolidating data from various structured and unstructured data systems into structures that are suitable for building analytics solutions that use Microsoft Azure data services. Each course teaches you the concepts and skills that are measured by the exam. By the end of this Specialization, you will be ready to take and sign-up for the Exam DP-203: Data Engineering on Microsoft Azure (beta).
Doing Clinical Research: Biostatistics with the Wolfram Language
This course has a singular and clear aim, to empower you to do statistical tests, ready for incorporation into your dissertations, research papers, and presentations. The ability to summarize data, create plots and charts, and to do the tests that you commonly see in the literature is a powerful skill indeed. Not only will it further your career, but it will put you in the position to contribute to the advancement of humanity through scientific research. We live in a wonderful age with great tools at our disposal, ready to achieve this goal. None are quite as easy to learn, yet as powerful to use, as the Wolfram Language. Knowledge is literally built into the language. With its well-structured and consistent approach to creating code, you will become an expert in no time. This course follows the modern trend of learning statistical analysis through the use of a computer language. It requires no prior knowledge of coding. An exciting journey awaits. If you wanting even more, there are optional Honors lessons on machine learning that cover the support in the Wolfram Language for deep learning.
Bayesian Statistics: Time Series Analysis
This course for practicing and aspiring data scientists and statisticians. It is the fourth of a four-course sequence introducing the fundamentals of Bayesian statistics. It builds on the course Bayesian Statistics: From Concept to Data Analysis, Techniques and Models, and Mixture models. Time series analysis is concerned with modeling the dependency among elements of a sequence of temporally related variables. To succeed in this course, you should be familiar with calculus-based probability, the principles of maximum likelihood estimation, and Bayesian inference. You will learn how to build models that can describe temporal dependencies and how to perform Bayesian inference and forecasting for the models. You will apply what you've learned with the open-source, freely available software R with sample databases. Your instructor Raquel Prado will take you from basic concepts for modeling temporally dependent data to implementation of specific classes of models
Understanding China, 1700-2000: A Data Analytic Approach, Part 1
The purpose of this course is to summarize new directions in Chinese history and social science produced by the creation and analysis of big historical datasets based on newly opened Chinese archival holdings, and to organize this knowledge in a framework that encourages learning about China in comparative perspective. Our course demonstrates how a new scholarship of discovery is redefining what is singular about modern China and modern Chinese history. Current understandings of human history and social theory are based largely on Western experience or on non-Western experience seen through a Western lens. This course offers alternative perspectives derived from Chinese experience over the last three centuries. We present specific case studies of this new scholarship of discovery divided into two stand-alone parts, which means that students can take any part without prior or subsequent attendance of the other part. Part 1 (this course) focuses on comparative inequality and opportunity and addresses two related questions ‘Who rises to the top?’ and ‘Who gets what?’. Part 2 (https://www.coursera.org/learn/understanding-china-history-part-2) turns to an arguably even more important question ‘Who are we?’ as seen through the framework of comparative population behavior - mortality, marriage, and reproduction – and their interaction with economic conditions and human values. We do so because mortality and reproduction are fundamental and universal, because they differ historically just as radically between China and the West as patterns of inequality and opportunity, and because these differences demonstrate the mutability of human behavior and values. Course Overview video: https://youtu.be/dzUPRyJ4ETk
Data Manipulation with dplyr in R
Welcome to this project-based course Data Manipulation with dplyr in R. In this project, you will learn how to manipulate data with the dplyr package in R. By the end of this 2-hour long project, you will understand how to use different dplyr verbs such as the select verb, filter verb, arrange verb, mutate verb, summarize verb, and the group_by verb to manipulate the gapminder dataset. Also, you will learn how to combine different dplyr verbs to manipulate the gapminder dataset to get the desired result. Note, you do not need to be an expert data analyst, data scientist or statistical analyst to be successful in this guided project, just a familiarity with the R language will suffice. If you do not have a prior experience with R, I recommend that you should take the Getting Started with R project before taking this project.
Regular Expressions in Python
In this 1-hour long project-based course, you will learn how to construct regex patterns, validate passwords and user input in web forms and extract patterns and replace strings with regex. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Managing Data Analysis
This one-week course describes the process of analyzing data and how to manage that process. We describe the iterative nature of data analysis and the role of stating a sharp question, exploratory data analysis, inference, formal statistical modeling, interpretation, and communication. In addition, we will describe how to direct analytic activities within a team and to drive the data analysis process towards coherent and useful results. This is a focused course designed to rapidly get you up to speed on the process of data analysis and how it can be managed. Our goal was to make this as convenient as possible for you without sacrificing any essential content. We've left the technical information aside so that you can focus on managing your team and moving it forward. After completing this course you will know how to…. 1. Describe the basic data analysis iteration 2. Identify different types of questions and translate them to specific datasets 3. Describe different types of data pulls 4. Explore datasets to determine if data are appropriate for a given question 5. Direct model building efforts in common data analyses 6. Interpret the results from common data analyses 7. Integrate statistical findings to form coherent data analysis presentations Commitment: 1 week of study, 4-6 hours Course cover image by fdecomite. Creative Commons BY https://flic.kr/p/4HjmvD
Google Cloud Big Data and Machine Learning Fundamentals
This course introduces the Google Cloud big data and machine learning products and services that support the data-to-AI lifecycle. It explores the processes, challenges, and benefits of building a big data pipeline and machine learning models with Vertex AI on Google Cloud.
Python and Machine-Learning for Asset Management with Alternative Data Sets
Over-utilization of market and accounting data over the last few decades has led to portfolio crowding, mediocre performance and systemic risks, incentivizing financial institutions which are looking for an edge to quickly adopt alternative data as a substitute to traditional data. This course introduces the core concepts around alternative data, the most recent research in this area, as well as practical portfolio examples and actual applications. The approach of this course is somewhat unique because while the theory covered is still a main component, practical lab sessions and examples of working with alternative datasets are also key. This course is fo you if you are aiming at carreers prospects as a data scientist in financial markets, are looking to enhance your analytics skillsets to the financial markets, or if you are interested in cutting-edge technology and research as they apply to big data. The required background is: Python programming, Investment theory , and Statistics. This course will enable you to learn new data and research techniques applied to the financial markets while strengthening data science and python skills.