Back to Courses

Data Analysis Courses - Page 12

Showing results 111-120 of 998
Tidy Messy Data using tidyr in R
As data enthusiasts and professionals, our work often requires dealing with data in different forms. In particular, messy data can be a big challenge because the quality of your analysis largely depends on the quality of the data. This project-based course, "Tidy Messy Data using tidyr in R," is intended for beginner and intermediate R users with related experiences who are willing to advance their knowledge and skills. In this course, you will learn practical ways for data cleaning, reshaping, and transformation using R. You will learn how to use different tidyr functions like pivot_longer(), pivot_wider(), separate_rows(), separate(), and others to achieve the tidy data principles. By the end of this 2-hour-long project, you will get hands-on massaging data to put in the proper format. By extension, you will learn to create plots using ggplot(). This project-based course is a beginner to an intermediate-level course in R. Therefore, to get the most out of this project, it is essential to have a basic understanding of using R. Specifically, you should be able to load data into R and understand how the pipe function works. It will be helpful to complete my previous project titled "Data Manipulation with dplyr in R."
Data Analyst Career Guide and Interview Preparation
This course is designed to prepare you to enter the job market as a data analyst. It provides guidance about the regular functions and tasks of data analysts and their place in the data ecosystem, as well as the opportunities of the profession and some options for career development. It explains practical techniques for creating essential job-seeking materials such as a resume and a portfolio, as well as auxiliary tools like a cover letter and an elevator pitch. You will learn how to find and assess prospective job positions, apply to them, and lay the groundwork for interviewing. You will also get inside tips and steps you can use to perform professionally and effectively at interviews. Let seasoned professionals share their experience to help you get ahead of the competition.
Data Analysis with Python
Analyzing data with Python is an essential skill for Data Scientists and Data Analysts. This course will take you from the basics of data analysis with Python to building and evaluating data models. Topics covered include: - collecting and importing data - cleaning, preparing & formatting data - data frame manipulation - summarizing data, - building machine learning regression models - model refinement - creating data pipelines You will learn how to import data from multiple sources, clean and wrangle data, perform exploratory data analysis (EDA), and create meaningful data visualizations. You will then predict future trends from data by developing linear, multiple, polynomial regression models & pipelines and learn how to evaluate them. In addition to video lectures you will learn and practice using hands-on labs and projects. You will work with several open source Python libraries, including Pandas and Numpy to load, manipulate, analyze, and visualize cool datasets. You will also work with scipy and scikit-learn, to build machine learning models and make predictions. If you choose to take this course and earn the Coursera course certificate, you will also earn an IBM digital badge.
Excel Time Series Models for Business Forecasting
This course explores different time series business forecasting methods. The course covers a variety of business forecasting methods for different types of components present in time series data — level, trending, and seasonal. We will learn about the theoretical methods and apply these methods to business data using Microsoft Excel. These forecasting methods will be programmed into Microsoft Excel, displayed graphically, and we will optimise these models to produce accurate forecasts. We will compare different models and their forecasts to decide which model best suits our business' needs.
Processing Data with Python
Processing data is used in virtually every field these days. It is used for analyzing web traffic to determine personal preferences, gathering scientific data for biological analysis, analyzing weather patterns, business practices, and on. Data can take on many different forms and come from many different sources. Python is an open-source (free) programming language that is used in web programming, data science, artificial intelligence, and many scientific applications. It has libraries that can be used to parse and quickly analyze the data in whatever form it comes in, whether it be in XML, CSV, or JSON format. Data cleaning is an important aspect of processing data, particularly in the field of data science. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Moneyball and Beyond
The book Moneyball triggered a revolution in the analysis of performance statistics in professional sports, by showing that data analytics could be used to increase team winning percentage. This course shows how to program data using Python to test the claims that lie behind the Moneyball story, and to examine the evolution of Moneyball statistics since the book was published. The learner is led through the process of calculating baseball performance statistics from publicly available datasets. The course progresses from the analysis of on base percentage and slugging percentage to more advanced measures derived using the run expectancy matrix, such as wins above replacement (WAR). By the end of this course the learner will be able to use these statistics to conduct their own team and player analyses.
Perform Feature Analysis with Yellowbrick
Welcome to this project-based course on Performing Feature Analysis with Yellowbrick. In this course, we are going to use visualizations to steer machine learning workflows. The problem we will tackle is to predict whether rooms in apartments are occupied or unoccupied based on passive sensor data such as temperature, humidity, light and CO2 levels. With an emphasis on visual steering of our analysis, we will cover the following topics in our machine learning workflow: feature analysis using methods such as scatter plots, RadViz, parallel coordinates plots, feature ranking, and manifold visualization. This course runs on Coursera's hands-on project platform called Rhyme. On Rhyme, you do projects in a hands-on manner in your browser. You will get instant access to pre-configured cloud desktops containing all of the software and data you need for the project. Everything is already set up directly in your internet browser so you can just focus on learning. For this project, you’ll get instant access to a cloud desktop with Python, Jupyter, Yellowbrick, and scikit-learn pre-installed. Notes: - You will be able to access the cloud desktop 5 times. However, you will be able to access instructions videos as many times as you want. - This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Use Python and Java to Create a GUI Application
By the end of this project, you will implement a Java GUI to read from a user-provided file containing data. The GUI will call Python applications to plot columnar data as X and Y coordinates on a regression graph, and display statistics about the data from each of the selected columns. A graphical user interface can be a nice alternative to using the command line for running programs, as there is no need to memorize how to execute a command with arguments. A label may be added to describe what is needed for the application, for example. There are many choices for building a graphical user interface in Java. Using the Java Swing GUI package is the standard GUI toolkit for Java applications and is widely available on multiple platforms including Windows, Mac, and Linux. The event handlers in Java can then call existing Python applications to analyze the data. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Data Integration with Microsoft Azure Data Factory
In this course, you will learn how to create and manage data pipelines in the cloud using Azure Data Factory. This course is part of a Specialization intended for Data engineers and developers who want to demonstrate their expertise in designing and implementing data solutions that use Microsoft Azure data services. It is ideal for anyone interested in preparing for the DP-203: Data Engineering on Microsoft Azure exam (beta). This is the third course in a program of 10 courses to help prepare you to take the exam so that you can have expertise in designing and implementing data solutions that use Microsoft Azure data services. The Data Engineering on Microsoft Azure exam is an opportunity to prove knowledge expertise in integrating, transforming, and consolidating data from various structured and unstructured data systems into structures that are suitable for building analytics solutions that use Microsoft Azure data services. Each course teaches you the concepts and skills that are measured by the exam. By the end of this Specialization, you will be ready to take and sign-up for the Exam DP-203: Data Engineering on Microsoft Azure (beta).
Advanced Relational Database and SQL
In this 1-hour long project-based course, you will gain hands-on experience and learn about advanced SQL topics such as stored procedures, tiggers, functions, common table expressions and recursion. If you have intermediate level of experience with SQL and want to learn more, this course is for you! Note: This is an advanced level course. Taking my course "Introduction to Relational Database and SQL" and "Intermediate Relational Database and SQL" before taking this course is highly recommended. Especially if you do not have any previous experience with relational database and SQL.