Back to Courses

Data Science Courses - Page 100

Showing results 991-1000 of 1407
Data Storytelling
This course will cover the more complex concepts that become involved when working beyond simple datasets. Exploring the connection between visual aspects and data understanding, we will examine how those concepts work together through data storytelling. After reviewing key points on how to avoid problematic visualizations and data misrepresentation, you will continue working in Tableau performing multivariate descriptive analysis of the S&P 500 stock sectors.
Applied Machine Learning in Python
This course will introduce the learner to applied machine learning, focusing more on the techniques and methods than on the statistics behind these methods. The course will start with a discussion of how machine learning is different than descriptive statistics, and introduce the scikit learn toolkit through a tutorial. The issue of dimensionality of data will be discussed, and the task of clustering data, as well as evaluating those clusters, will be tackled. Supervised approaches for creating predictive models will be described, and learners will be able to apply the scikit learn predictive modelling methods while understanding process issues related to data generalizability (e.g. cross validation, overfitting). The course will end with a look at more advanced techniques, such as building ensembles, and practical limitations of predictive models. By the end of this course, students will be able to identify the difference between a supervised (classification) and unsupervised (clustering) technique, identify which technique they need to apply for a particular dataset and need, engineer features to meet that need, and write python code to carry out an analysis. This course should be taken after Introduction to Data Science in Python and Applied Plotting, Charting & Data Representation in Python and before Applied Text Mining in Python and Applied Social Analysis in Python.
Introduction to Widgets for Data Science
In this 2-hour long project-based course, you will learn what are widgets, how they can used for data science work, types of widgets, linking multiple widgets, basic dashboards of widgets and creating child widgets.
Introduction to Machine Learning in Sports Analytics
In this course students will explore supervised machine learning techniques using the python scikit learn (sklearn) toolkit and real-world athletic data to understand both machine learning algorithms and how to predict athletic outcomes. Building on the previous courses in the specialization, students will apply methods such as support vector machines (SVM), decision trees, random forest, linear and logistic regression, and ensembles of learners to examine data from professional sports leagues such as the NHL and MLB as well as wearable devices such as the Apple Watch and inertial measurement units (IMUs). By the end of the course students will have a broad understanding of how classification and regression techniques can be used to enable sports analytics across athletic activities and events.
How to Create a Program Evaluation for Your Non-Profit
In this 1.5 hour long project-based course, you will learn how to create a program evaluation plan for your non-profit. By the end of the course, you will understand the importance of program evaluation, how to use Logic Models, how to write SMART goals for your program, and how to formulate good questions to gather the data you need. Learning Objectives: Task 1: Understand why evaluation is valuable for your program or organization. Task 2: Learn how to identify and define the typical components of a Logic Model. Task 3: Write SMART goals for a program outcome. Task 4: Identify different measurement tools and how they will be helpful in the evaluation process. Task 5: Understand what makes a good question, identify different types of questions, and practice writing good questions. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Image Super Resolution Using Autoencoders in Keras
Welcome to this 1.5 hours long hands-on project on Image Super Resolution using Autoencoders in Keras. In this project, you’re going to learn what an autoencoder is, use Keras with Tensorflow as its backend to train your own autoencoder, and use this deep learning powered autoencoder to significantly enhance the quality of images. That is, our neural network will create high-resolution images from low-res source images. This course runs on Coursera's hands-on project platform called Rhyme. On Rhyme, you do projects in a hands-on manner in your browser. You will get instant access to pre-configured cloud desktops containing all of the software and data you need for the project. Everything is already set up directly in your internet browser so you can just focus on learning. For this project, you’ll get instant access to a cloud desktop with Python, Jupyter, and Keras pre-installed. Notes: - You will be able to access the cloud desktop 5 times. However, you will be able to access instructions videos as many times as you want. - This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
AI Workflow: Machine Learning, Visual Recognition and NLP
This is the fourth course in the IBM AI Enterprise Workflow Certification specialization.    You are STRONGLY encouraged to complete these courses in order as they are not individual independent courses, but part of a workflow where each course builds on the previous ones.  Course 4 covers the next stage of the workflow, setting up models and their associated data pipelines for a hypothetical streaming media company.  The first topic covers the complex topic of evaluation metrics, where you will learn best practices for a number of different metrics including regression metrics, classification metrics, and multi-class metrics, which you will use to select the best model for your business challenge.  The next topics cover best practices for different types of models including linear models, tree-based models, and neural networks.  Out-of-the-box Watson models for natural language understanding and visual recognition will be used.  There will be case studies focusing on natural language processing and on image analysis to provide realistic context for the model pipelines.   By the end of this course you will be able to: Discuss common regression, classification, and multilabel classification metrics Explain the use of linear and logistic regression in supervised learning applications Describe common strategies for grid searching and cross-validation Employ evaluation metrics to select models for production use Explain the use of tree-based algorithms in supervised learning applications Explain the use of Neural Networks in supervised learning applications Discuss the major variants of neural networks and recent advances Create a neural net model in Tensorflow Create and test an instance of Watson Visual Recognition Create and test an instance of Watson NLU Who should take this course? This course targets existing data science practitioners that have expertise building machine learning models, who want to deepen their skills on building and deploying AI in large enterprises. If you are an aspiring Data Scientist, this course is NOT for you as you need real world expertise to benefit from the content of these courses.   What skills should you have? It is assumed that you have completed Courses 1 through 3 of the IBM AI Enterprise Workflow specialization and you have a solid understanding of the following topics prior to starting this course: Fundamental understanding of Linear Algebra; Understand sampling, probability theory, and probability distributions; Knowledge of descriptive and inferential statistical concepts; General understanding of machine learning techniques and best practices; Practiced understanding of Python and the packages commonly used in data science: NumPy, Pandas, matplotlib, scikit-learn; Familiarity with IBM Watson Studio; Familiarity with the design thinking process.
AI Workflow: AI in Production
This is the sixth course in the IBM AI Enterprise Workflow Certification specialization.   You are STRONGLY encouraged to complete these courses in order as they are not individual independent courses, but part of a workflow where each course builds on the previous ones.     This course focuses on models in production at a hypothetical streaming media company.  There is an introduction to IBM Watson Machine Learning.  You will build your own API in a Docker container and learn how to manage containers with Kubernetes.  The course also introduces  several other tools in the IBM ecosystem designed to help deploy or maintain models in production.  The AI workflow is not a linear process so there is some time dedicated to the most important feedback loops in order to promote efficient iteration on the overall workflow.   By the end of this course you will be able to: 1.  Use Docker to deploy a flask application 2.  Deploy a simple UI to integrate the ML model, Watson NLU, and Watson Visual Recognition 3.  Discuss basic Kubernetes terminology 4.  Deploy a scalable web application on Kubernetes  5.  Discuss the different feedback loops in AI workflow 6.  Discuss the use of unit testing in the context of model production 7.  Use IBM Watson OpenScale to assess bias and performance of production machine learning models. Who should take this course? This course targets existing data science practitioners that have expertise building machine learning models, who want to deepen their skills on building and deploying AI in large enterprises. If you are an aspiring Data Scientist, this course is NOT for you as you need real world expertise to benefit from the content of these courses.   What skills should you have? It is assumed that you have completed Courses 1 through 5 of the IBM AI Enterprise Workflow specialization and you have a solid understanding of the following topics prior to starting this course: Fundamental understanding of Linear Algebra; Understand sampling, probability theory, and probability distributions; Knowledge of descriptive and inferential statistical concepts; General understanding of machine learning techniques and best practices; Practiced understanding of Python and the packages commonly used in data science: NumPy, Pandas, matplotlib, scikit-learn; Familiarity with IBM Watson Studio; Familiarity with the design thinking process.
Validity and Bias in Epidemiology
Epidemiological studies can provide valuable insights about the frequency of a disease, its potential causes and the effectiveness of available treatments. Selecting an appropriate study design can take you a long way when trying to answer such a question. However, this is by no means enough. A study can yield biased results for many different reasons. This course offers an introduction to some of these factors and provides guidance on how to deal with bias in epidemiological research. In this course you will learn about the main types of bias and what effect they might have on your study findings. You will then focus on the concept of confounding and you will explore various methods to identify and control for confounding in different study designs. In the last module of this course we will discuss the phenomenon of effect modification, which is key to understanding and interpreting study results. We will finish the course with a broader discussion of causality in epidemiology and we will highlight how you can utilise all the tools that you have learnt to decide whether your findings indicate a true association and if this can be considered causal.
Predict Ad Clicks Using Logistic Regression and XG-Boost
In this project, we will predict Ads clicks using logistic regression and XG-boost algorithms. In this project, we will assume that you have been hired as a consultant to a start-up that is running a targeted marketing ad campaign on Facebook. The company wants to analyze customer behavior by predicting which customer clicks on the advertisement.