Back to Courses

Data Science Courses - Page 99

Showing results 981-990 of 1407
Data Modeling and Regression Analysis in Business
The course will begin with what is familiar to many business managers and those who have taken the first two courses in this specialization. The first set of tools will explore data description, statistical inference, and regression. We will extend these concepts to other statistical methods used for prediction when the response variable is categorical such as win-don’t win an auction. In the next segment, students will learn about tools used for identifying important features in the dataset that can either reduce the complexity or help identify important features of the data or further help explain behavior. 
Pattern Discovery in Data Mining
Learn the general concepts of data mining along with basic methodologies and applications. Then dive into one subfield in data mining: pattern discovery. Learn in-depth concepts, methods, and applications of pattern discovery in data mining. We will also introduce methods for data-driven phrase mining and some interesting applications of pattern discovery. This course provides you the opportunity to learn skills and content to practice and engage in scalable pattern discovery methods on massive transactional data, discuss pattern evaluation measures, and study methods for mining diverse kinds of patterns, sequential patterns, and sub-graph patterns.
Create digit recognition web app with Streamlit
In this 1-hour long project-based course, you will learn how to create a digit recognition web application using streamlit. This project is divided into two stages. In the first stage, you are going to write the training pipeline in which you will load MNIST Handwritten dataset. You will write the training and validation functions in order to train and validate the dataset. Lastly, in this stage you will do inference. In the second stage, you will use the best trained model from the training pipeline and you will use that in your web app. You will create the web user interface using streamlit python library. In this web app a user will draw a digit and given that drawn digit, the best trained model will output the probabilities.
Calculus through Data & Modelling: Vector Calculus
This course continues your study of calculus by focusing on the applications of integration to vector valued functions, or vector fields. These are functions that assign vectors to points in space, allowing us to develop advanced theories to then apply to real-world problems. We define line integrals, which can be used to fund the work done by a vector field. We culminate this course with Green's Theorem, which describes the relationship between certain kinds of line integrals on closed paths and double integrals. In the discrete case, this theorem is called the Shoelace Theorem and allows us to measure the areas of polygons. We use this version of the theorem to develop more tools of data analysis through a peer reviewed project. Upon successful completion of this course, you have all the tools needed to master any advanced mathematics, computer science, or data science that builds off of the foundations of single or multivariable calculus.
Introduction to Business Analysis Using Spreadsheets: Basics
In this 1-hour 30-mins long project-based course, you will learn the responsibilities of a Business Analyst such as Learn the basic concepts of data analysis and descriptive statistics. Learn how to manipulate, analyze, and visualize data in Google Sheets using functions, aggregation functions, and logical aggregation functions. and present data using different types of charts. This course works best for learners who wish to learn about Business Analysis and wish to learn about the role of a Business Analyst. Note: This project works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Retrieve Data with Multiple-Table SQL Queries
In this course you will be introduced to two methods of writing SQL queries that retrieve data from two or more tables. Since one of the functions of a database is to store data in an organized format, many databases are made up of multiple tables. Often, the data output required from the database is made up of data from more than one table. For example, the data that populates a student transcript might come from the Student, Course, and Section tables. While the Student table may provide the student’s name, the name and number of the course might come from the Course table and the specific grade for that course may come from yet another table. While writing SQL queries in SQLiteStudio, you'll learn the SQL syntax required to join tables together as you develop an understanding of how the relationships among tables come into play. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Geospatial Big Data Visualization with Kepler GL
In this 1-hour long project-based course, you will learn how to easily create beautiful data visualization with Kepler and effectively design different geospatial data visualizations.
Understanding Deepfakes with Keras
In this 2-hour long project-based course, you will learn to implement DCGAN or Deep Convolutional Generative Adversarial Network, and you will train the network to generate realistic looking synthesized images. The term Deepfake is typically associated with synthetic data generated by Neural Networks which is similar to real-world, observed data - often with synthesized images, videos or audio. Through this hands-on project, we will go through the details of how such a network is structured, trained, and will ultimately generate synthetic images similar to hand-written digit 0 from the MNIST dataset. Since this is a practical, project-based course, you will need to have a theoretical understanding of Neural Networks, Convolutional Neural Networks, and optimization algorithms like Gradient Descent. We will focus on the practical aspect of implementing and training DCGAN, but not too much on the theoretical aspect. You will also need some prior experience with Python programming. This course runs on Coursera's hands-on project platform called Rhyme. On Rhyme, you do projects in a hands-on manner in your browser. You will get instant access to pre-configured cloud desktops containing all of the software and data you need for the project. Everything is already set up directly in your internet browser so you can just focus on learning. For this project, you’ll get instant access to a cloud desktop with Python, Jupyter, and Tensorflow pre-installed. Notes: - You will be able to access the cloud desktop 5 times. However, you will be able to access instructions videos as many times as you want. - This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Datastream MySQL to BigQuery
This is a self-paced lab that takes place in the Google Cloud console. Learn to migrate MySQL Databases to BigQuery using Datastream and Dataflow. Datastream is a serverless and easy-to-use Change Data Capture (CDC) and replication service that allows you to synchronize data across heterogeneous databases, storage systems, and applications reliably and with minimal latency. In this lab you'll learn how to replicate data from your OLTP workloads into BigQuery, in real time. You will begin by deploying MySQL on Cloud SQL and import a dataset using the gcloud command line. Then, in the Cloud Console UI, you will create and start a Datastream stream and a Dataflow job for replication. The replication uses a Dataflow template to enable continuous replication of data, along with Cloud Storage and Pub/Sub for buffering data.
Basic Statistics
Understanding statistics is essential to understand research in the social and behavioral sciences. In this course you will learn the basics of statistics; not just how to calculate them, but also how to evaluate them. This course will also prepare you for the next course in the specialization - the course Inferential Statistics. In the first part of the course we will discuss methods of descriptive statistics. You will learn what cases and variables are and how you can compute measures of central tendency (mean, median and mode) and dispersion (standard deviation and variance). Next, we discuss how to assess relationships between variables, and we introduce the concepts correlation and regression. The second part of the course is concerned with the basics of probability: calculating probabilities, probability distributions and sampling distributions. You need to know about these things in order to understand how inferential statistics work. The third part of the course consists of an introduction to methods of inferential statistics - methods that help us decide whether the patterns we see in our data are strong enough to draw conclusions about the underlying population we are interested in. We will discuss confidence intervals and significance tests. You will not only learn about all these statistical concepts, you will also be trained to calculate and generate these statistics yourself using freely available statistical software.