Back to Courses

Machine Learning Courses - Page 13

Showing results 121-130 of 485
Predicting Salaries with Decision Trees
In this 1.5 hour long project-based course, you will tackle a real-world prediction problem using machine learning. The dataset we are going to use comes from the U.S. Census Bureau; they recorded a number of attributes such as gender and occupation as well as the salary range for a sample of more than 32,000 Americans. We will fit a decision tree to this data, and try to predict the salary for a person we haven’t seen before. By the end of this project, you will have created a machine learning model using industry standard tools, including Python and sklearn. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Data Science Fundamentals for Data Analysts
In this course we're going to guide you through the fundamental building blocks of data science, one of the fastest-growing fields in the world! With the help of our industry-leading data scientists, we’ve designed this course to build ready-to-apply data science skills in just 15 hours of learning. First, we’ll give you a quick introduction to data science - what it is and how it is used to solve real-world problems. For the rest of the course, we'll teach you the skills you need to apply foundational data science concepts and techniques to solve these real-world problems. By the end of this course, you'll be able to leverage your existing data analysis skills to design, execute, assess, and communicate the results of your very own data science projects.
Introduction to Recommender Systems: Non-Personalized and Content-Based
This course, which is designed to serve as the first course in the Recommender Systems specialization, introduces the concept of recommender systems, reviews several examples in detail, and leads you through non-personalized recommendation using summary statistics and product associations, basic stereotype-based or demographic recommendations, and content-based filtering recommendations. After completing this course, you will be able to compute a variety of recommendations from datasets using basic spreadsheet tools, and if you complete the honors track you will also have programmed these recommendations using the open source LensKit recommender toolkit. In addition to detailed lectures and interactive exercises, this course features interviews with several leaders in research and practice on advanced topics and current directions in recommender systems.
Mathematics for Machine Learning: PCA
This intermediate-level course introduces the mathematical foundations to derive Principal Component Analysis (PCA), a fundamental dimensionality reduction technique. We'll cover some basic statistics of data sets, such as mean values and variances, we'll compute distances and angles between vectors using inner products and derive orthogonal projections of data onto lower-dimensional subspaces. Using all these tools, we'll then derive PCA as a method that minimizes the average squared reconstruction error between data points and their reconstruction. At the end of this course, you'll be familiar with important mathematical concepts and you can implement PCA all by yourself. If you’re struggling, you'll find a set of jupyter notebooks that will allow you to explore properties of the techniques and walk you through what you need to do to get on track. If you are already an expert, this course may refresh some of your knowledge. The lectures, examples and exercises require: 1. Some ability of abstract thinking 2. Good background in linear algebra (e.g., matrix and vector algebra, linear independence, basis) 3. Basic background in multivariate calculus (e.g., partial derivatives, basic optimization) 4. Basic knowledge in python programming and numpy Disclaimer: This course is substantially more abstract and requires more programming than the other two courses of the specialization. However, this type of abstract thinking, algebraic manipulation and programming is necessary if you want to understand and develop machine learning algorithms.
Google Cloud Big Data and Machine Learning Fundamentals
This course introduces the Google Cloud big data and machine learning products and services that support the data-to-AI lifecycle. It explores the processes, challenges, and benefits of building a big data pipeline and machine learning models with Vertex AI on Google Cloud.
Diabetes Prediction With Pyspark MLLIB
In this 1 hour long project-based course, you will learn to build a logistic regression model using Pyspark MLLIB to classify patients as either diabetic or non-diabetic. We will use the popular Pima Indian Diabetes data set. Our goal is to use a simple logistic regression classifier from the pyspark Machine learning library for diabetes classification. We will be carrying out the entire project on the Google Colab environment with the installation of Pyspark.You will need a free Gmail account to complete this project. Please be aware of the fact that the dataset and the model in this project, can not be used in the real-life. We are only using this data for the educational purpose. By the end of this project, you will be able to build the logistic regression classifier using Pyspark MLlib to classify between the diabetic and nondiabetic patients.You will also be able to setup and work with Pyspark on Google colab environment. Additionally, you will also be able to clean and prepare data for analysis. You should be familiar with the Python Programming language and you should have a theoretical understanding of the Logistic Regression algorithm. You will need a free Gmail account to complete this project. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Capstone Project: Predicting Safety Stock
In this course, we'll make predictions on product usage and calculate optimal safety stock storage. We'll start with a time series of shoe sales across multiple stores on three different continents. To begin, we'll look for unique insights and other interesting things we can find in the data by performing groupings and comparing products within each store. Then, we'll use a seasonal autoregressive integrated moving average (SARIMA) model to make predictions on future sales. In addition to making predictions, we'll analyze the provided statistics (such as p-score) to judge the viability of using the SARIMA model to make predictions. Then, we'll tune the hyper-parameters of the model to garner better results and higher statistical significance. Finally, we'll make predictions on safety stock by looking to the data for monthly usage predictions and calculating safety stock from the formula involving lead times.
Advanced Learning Algorithms
In the second course of the Machine Learning Specialization, you will: • Build and train a neural network with TensorFlow to perform multi-class classification • Apply best practices for machine learning development so that your models generalize to data and tasks in the real world • Build and use decision trees and tree ensemble methods, including random forests and boosted trees The Machine Learning Specialization is a foundational online program created in collaboration between DeepLearning.AI and Stanford Online. In this beginner-friendly program, you will learn the fundamentals of machine learning and how to use these techniques to build real-world AI applications. This Specialization is taught by Andrew Ng, an AI visionary who has led critical research at Stanford University and groundbreaking work at Google Brain, Baidu, and Landing.AI to advance the AI field. This 3-course Specialization is an updated and expanded version of Andrew’s pioneering Machine Learning course, rated 4.9 out of 5 and taken by over 4.8 million learners since it launched in 2012. It provides a broad introduction to modern machine learning, including supervised learning (multiple linear regression, logistic regression, neural networks, and decision trees), unsupervised learning (clustering, dimensionality reduction, recommender systems), and some of the best practices used in Silicon Valley for artificial intelligence and machine learning innovation (evaluating and tuning models, taking a data-centric approach to improving performance, and more.) By the end of this Specialization, you will have mastered key theoretical concepts and gained the practical know-how to quickly and powerfully apply machine learning to challenging real-world problems. If you’re looking to break into AI or build a career in machine learning, the new Machine Learning Specialization is the best place to start.
Deep Learning with PyTorch : Object Localization
Object Localization is the task of locating an instance of a particular object category in an image, typically by specifying a tightly cropped bounding box centered on the instance. In this 2-hour project-based course, you will be able to understand the Object Localization Dataset and you will write a custom dataset class for Image-bounding box dataset. Additionally, you will apply augmentation for localization task to augment images as well as its effect on bounding box. For localization task augmentation you will use albumentation library. We will plot the (image-bounding box) pair. Thereafter, we will load a pretrained state of the art convolutional neural network using timm library.Moreover, we are going to create train function and evaluator function which will be helpful to write training loop. Lastly, you will use best trained model to find bounding box given any image.
Image Classification on Autopilot with AWS AutoGluon
Hello everyone and welcome to this new hands-on project on image classification with Amazon Web Services (AWS) AutoGluon. In this project, we will train several deep neural networks models to classify images using a powerful library known as AutoGluon. AutoGluon is the library behind AWS SageMaker autopilot and it allows for quick prototyping of several powerful models using a few lines of code.