Back to Courses

Data Analysis Courses - Page 8

Showing results 71-80 of 998
Create an infographic with Infogram
In this 2-hour long project-based course, you will learn how to design effectively an infographic with infogram.com, adding line, bar and map charts, and connecting the data story with text and visuals. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
SQL Functions
By the end of this project, you will create a number of examples that will develop your learning around functions in SQL. This course will enable you to take your beginner knowledge of SQL to the next level by incorporating functions into your programming. Thus, you will be able to develop more complex code and be able to solve more difficult problems. Thus, you will be able to develop more complex code and be able to solve more difficult problems. This course will provide students with the knowledge behind different functions in SQL such as string functions, numeric functions, date functions, null SQL functions, stored functions and stored procedures. This project will take students through a number of examples demonstrating SQL functions based on a database. You will gain an understanding of these concepts from the in-depth examples provided. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Hierarchical relational data analysis using python
By the end of this project you will learn how to analyze Hierarchical Data. we are going to work with a dataset related to Mexico toy sales. The dataset contains some hierarchical data about different products sold in different stores in different cities in Mexico. we are going to load this data and after some preprocessing steps, we are going to learn how to analyze this data using different visualization techniques. During this project we are going to learn about a very important concept called Data Granularity. And we will also learn how to use different levels of granularity to answer some analytical question. and at the end we are going to talk about Treemaps and Sunburst Diagram, two handy visualization techniques used for hierarchical data.
Creating Multi Task Models With Keras
In this 1 hour long guided project, you will learn to create and train multi-task, multi-output models with Keras. You will learn to use Keras' functional API to create a multi output model which will be trained to learn two different labels given the same input example. The model will have one input but two outputs. A few of the shallow layers will be shared between the two outputs, you will also use a ResNet style skip connection in the model. If you are familiar with Keras, you have probably come across examples of models that are trained to perform multiple tasks. For example, an object detection model where a CNN is trained to find all class instances in the input images as well as give a regression output to localize the detected class instances in the input. Being able to use Keras' functional API is a first step towards building complex, multi-output models like object detection models. We will be using TensorFlow as our machine learning framework. The project uses the Google Colab environment. You will need prior programming experience in Python. You will also need prior experience with Keras. Consider this to be an intermediate level Keras project. This is a practical, hands on guided project for learners who already have theoretical understanding of Neural Networks, Convolutional Neural Networks, and optimization algorithms like gradient descent but want to understand how to use use Keras to write custom, more complex models than just plain sequential neural networks. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.
Data Science Methodology
If there is a shortcut to becoming a Data Scientist, then learning to think and work like a successful Data Scientist is it. Most of the established data scientists follow a similar methodology for solving Data Science problems. In this course you will learn and then apply this methodology that can be used to tackle any Data Science scenario. The purpose of this course is to share a methodology that can be used within data science, to ensure that the data used in problem solving is relevant and properly manipulated to address the question at hand. Accordingly, in this course, you will learn: - The major steps involved in practicing data science - Forming a business/research problem, collecting, preparing & analyzing data, building a model, deploying a model and understanding the importance of feedback - Apply the 6 stages of the CRISP-DM methodology, the most popular methodology for Data Science and Data Mining problems - How data scientists think! To apply the methodology, you will work on a real-world inspired scenario and work with Jupyter Notebooks using Python to develop hands-on experience.
Modeling Data in the Tidyverse
Developing insights about your organization, business, or research project depends on effective modeling and analysis of the data you collect. Building effective models requires understanding the different types of questions you can ask and how to map those questions to your data. Different modeling approaches can be chosen to detect interesting patterns in the data and identify hidden relationships. This course covers the types of questions you can ask of data and the various modeling approaches that you can apply. Topics covered include hypothesis testing, linear regression, nonlinear modeling, and machine learning. With this collection of tools at your disposal, as well as the techniques learned in the other courses in this specialization, you will be able to make key discoveries from your data for improving decision-making throughout your organization. In this specialization we assume familiarity with the R programming language. If you are not yet familiar with R, we suggest you first complete R Programming before returning to complete this course.
Database Design and Basic SQL in PostgreSQL
In this course you will learn more about the historical design of databases and the use of SQL in the PostgreSQL environment. Using SQL techniques and common commands (INSERT INTO, WHERE, ORDER BY, ON DELETE CASCADE, etc) will enable you to create tables, column types and define the schema of your data in PostgreSQL. You will learn about data modeling and how to represent one-to-many and many-to-many relationships in PostgreSQL. Students will do hands-on assignments creating tables, inserting data, designing data models, creating relational structures and inserting and querying relational data in tables.
Life Expectancy Prediction Using Machine Learning
In this hands-on project, we will train a Linear Regression model to predict life expectancy. The dataset was initially obtained from the World Health Organization (WHO) and United Nations Websites. Data contains features such as year, status, life expectancy, adult mortality, infant deaths, percentage of expenditure, and alcohol consumption.
Applied Data Science for Data Analysts
In this course, you will develop your data science skills while solving real-world problems. You'll work through the data science process to and use unsupervised learning to explore data, engineer and select meaningful features, and solve complex supervised learning problems using tree-based models. You will also learn to apply hyperparameter tuning and cross-validation strategies to improve model performance. NOTE: This is the third and final course in the Data Science with Databricks for Data Analysts Coursera specialization. To be successful in this course we highly recommend taking the first two courses in that specialization prior to taking this course. These courses are: Apache Spark for Data Analysts and Data Science Fundamentals for Data Analysts.
Transfer Learning for NLP with TensorFlow Hub
This is a hands-on project on transfer learning for natural language processing with TensorFlow and TF Hub. By the time you complete this project, you will be able to use pre-trained NLP text embedding models from TensorFlow Hub, perform transfer learning to fine-tune models on real-world data, build and evaluate multiple models for text classification with TensorFlow, and visualize model performance metrics with Tensorboard. Prerequisites: In order to successfully complete this project, you should be competent in the Python programming language, be familiar with deep learning for Natural Language Processing (NLP), and have trained models with TensorFlow or and its Keras API. Note: This course works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.