Back to Courses

Data Science Courses - Page 89

Showing results 881-890 of 1407
Social Media Data Analytics
Learner Outcomes: After taking this course, you will be able to: - Utilize various Application Programming Interface (API) services to collect data from different social media sources such as YouTube, Twitter, and Flickr. - Process the collected data - primarily structured - using methods involving correlation, regression, and classification to derive insights about the sources and people who generated that data. - Analyze unstructured data - primarily textual comments - for sentiments expressed in them. - Use different tools for collecting, analyzing, and exploring social media data for research and development purposes. Sample Learner Story: Data analyst wanting to leverage social media data. Isabella is a Data Analyst working as a consultant for a multinational corporation. She has experience working with Web analysis tools as well as marketing data. She wants to now expand into social media arena, trying to leverage the vast amounts of data available through various social media channels. Specifically, she wants to see how their clients, partners, and competitors view their products/services and talk about them. She hopes to build a new workflow of data analytics that incorporates traditional data processing using Web and marketing tools, as well as newer methods of using social media data. Sample Job Roles requiring these skills: - Social Media Analyst - Web Analyst - Data Analyst - Marketing and Public Relations Final Project Deliverable/ Artifact: The course will have a series of small assignments or mini-projects that involve data collection, analysis, and presentation involving various social media sources using the techniques learned in the class. The course was developed by Dr. Chirag Shah while he was a faculty member at Rutgers University. He is currently a faculty member at University of Washington.
How to Use SQL with Large Datasets
By the end of this project, you will use SQL to manage a large COVID-19 dataset using MySQL Workbench. MySQL is a widely used relational database and can be used with large datasets if it is managed appropriately. This may include using the proper database engine, indexing the database, creating summary tables, and using proper database queries.
Building Stock Returns Heatmap with Tableau
In this 1-hour long project-based course, you will learn how to extract stock data using Google Finance, build a Heat and Treemap in Tableau, build a stock returns dashboard in Tableau. Note: This course works best for learners who are based in the North America region. We're currently working on providing the same experience in other regions. This course's content is not intended to be investment advice and does not constitute an offer to perform any operations in the regulated or unregulated financial market.
Introduction to R Programming and Tidyverse
This course is a gentle introduction to programming in R designed for 3 types of learners. It will be right for you, if: • you want to do data analysis but don’t know programming • you know programming but aren’t familiar with R • you know some R programming but want to learn the tidyverse verbs You will learn to do data visualization and analysis in a reproducible manner and use functions that allow your code to be easily read and understood. You will use RMarkdown to create nice documents and reports that execute your code freshly every time it’s run and that capture your thoughts about the data along the way. This course has been designed for learners from non-STEM backgrounds to help prepare them for more advanced data science courses by providing an introduction to programming and to the R language. I am excited for you to join me on the journey! The course logo was created using images of stickers from the RStudio shop. Please visit https://swag.rstudio.com/s/shop.
Introduction to Computer Vision and Image Processing
Computer Vision is one of the most exciting fields in Machine Learning and AI. It has applications in many industries, such as self-driving cars, robotics, augmented reality, and much more. In this beginner-friendly course, you will understand computer vision and learn about its various applications across many industries. As part of this course, you will utilize Python, Pillow, and OpenCV for basic image processing and perform image classification and object detection. This is a hands-on course and involves several labs and exercises. Labs will combine Jupyter Labs and Computer Vision Learning Studio (CV Studio), a free learning tool for computer vision. CV Studio allows you to upload, train, and test your own custom image classifier and detection models. At the end of the course, you will create your own computer vision web app and deploy it to the Cloud. This course does not require any prior Machine Learning or Computer Vision experience. However, some knowledge of the Python programming language and high school math is necessary.
Sample-based Learning Methods
In this course, you will learn about several algorithms that can learn near optimal policies based on trial and error interaction with the environment---learning from the agent’s own experience. Learning from actual experience is striking because it requires no prior knowledge of the environment’s dynamics, yet can still attain optimal behavior. We will cover intuitively simple but powerful Monte Carlo methods, and temporal difference learning methods including Q-learning. We will wrap up this course investigating how we can get the best of both worlds: algorithms that can combine model-based planning (similar to dynamic programming) and temporal difference updates to radically accelerate learning. By the end of this course you will be able to: - Understand Temporal-Difference learning and Monte Carlo as two strategies for estimating value functions from sampled experience - Understand the importance of exploration, when using sampled experience rather than dynamic programming sweeps within a model - Understand the connections between Monte Carlo and Dynamic Programming and TD. - Implement and apply the TD algorithm, for estimating value functions - Implement and apply Expected Sarsa and Q-learning (two TD methods for control) - Understand the difference between on-policy and off-policy control - Understand planning with simulated experience (as opposed to classic planning strategies) - Implement a model-based approach to RL, called Dyna, which uses simulated experience - Conduct an empirical study to see the improvements in sample efficiency when using Dyna
GIS, Mapping, and Spatial Analysis Capstone
In this capstone course, you will apply everything you have learned by designing and then completing your own GIS project. You will plan out your project by writing a brief proposal that explains what you plan to do and why. You will then find data for a topic and location of your choice, and perform analysis and create maps that allow you to try out different tools and data sets. The results of your work will be assembled into an Esri story map, which is a web site with maps, images, text, and video. The goal is for you to have a finished product that you can share, and that demonstrates what you have learned. Note: software is not provided for this course.
Cloud DNS: Traffic Steering using Geolocation Policy
This is a self-paced lab that takes place in the Google Cloud console. In this lab you will configure and test the Geolocation routing policy.
Reproducible Templates for Analysis and Dissemination
This course will assist you with recreating work that a previous coworker completed, revisiting a project you abandoned some time ago, or simply reproducing a document with a consistent format and workflow. Incomplete information about how the work was done, where the files are, and which is the most recent version can give rise to many complications. This course focuses on the proper documentation creation process, allowing you and your colleagues to easily reproduce the components of your workflow. Throughout this course, you'll receive helpful demonstrations of RStudio and the R Markdown language and engage in active learning opportunities to help you build a professional online portfolio.
Datadog: Getting started with the Helm Chart
This is a self-paced lab that takes place in the Google Cloud console. In this lab, you will learn how to use the Datadog Helm Chart. In this lab you will run the Datadog Agent in a Kubernetes cluster as a DaemonSet in order to start collecting your cluster and applications metrics, traces, and logs. You can deploy a Datadog Agent with a Helm chart or directly with a DaemonSet object YAML definition. In this lab you will be explaining and using those options in a real cluster, checking in real time the features they enable.