Torrent details for "PacktPub | 50 Hours of Big Data, PySpark, AWS, Scala, and Scraping [Video] [FCO]"    Log in to bookmark

Torrent details
Cover
Download
Torrent rating (1 rated)
Controls:
Category:
Language:
English English
Total Size:
17.33 GB
Info Hash:
c5bd420f293768e4c5270514f34e65e770e862e3
Added By:
Added:  
26-04-2022 21:07
Views:
1,187
Health:
Seeds:
7
Leechers:
5
Completed:
853



Thanks for rating :
Portal (5),


Description
Lynda and other Courses >>> https://www.freecoursesonline.me/

Image error


By: AI Sciences
Released: March 2022
Course Source: https://www.packtpub.com/product/50-hours-of-big-data-pyspark-aws-scala-and-scraping-video/9781803237039

Video Details

ISBN 9781803237039
Course Length 54 hours 32 minutes

About

Part 1 is designed to reflect the most in-demand Scala skills. It provides an in-depth understanding of core Scala concepts. We will wrap up with a discussion on Map Reduce and ETL pipelines using Spark from AWS S3 to AWS RDS (includes six mini-projects and one Scala Spark project).

Part 2 covers PySpark to perform data analysis. You will explore Spark RDDs, Dataframes, a bit of Spark SQL queries, transformations, and actions that can be performed on the data using Spark RDDs and dataframes, the ecosystem of Spark and Hadoop, and their underlying architecture. You will also learn how we can leverage AWS storage, databases, computations, and how Spark can communicate with different AWS services.

Part 3 is all about data scraping and data mining. You will cover important concepts such as Internet Browser execution and communication with the server, synchronous and asynchronous, parsing data in response from the server, tools for data scraping, Python requests module, and more.

In Part 4, you will be using MongoDB to develop an understanding of the NoSQL databases. You will explore the basic operations and explore the MongoDB query, project and update operators. We will wind up this section with two projects: Developing a CRUD-based application using Django and MongoDB and implementing an ETL pipeline using PySpark to dump the data in MongoDB.

By the end of this course, you will be able to relate the concepts and practical aspects of learned technologies with real-world problems.

All the resources of this course are available at https://github.com/PacktPublishing/50-Hours-of-Big-Data-PySpark-AWS-Scala-and-Scraping

Author

AI Sciences, The AI Sciences Team is a group of experts, PhDs, and practitioners of artificial intelligence, computer science, machine learning, and statistics. Some of them work in big companies such as Amazon, Google, Facebook, Microsoft, KPMG, BCG, and IBM.

They have produced a series of courses mainly dedicated to beginners and newcomers on the techniques and methods of machine learning, statistics, artificial intelligence, and data science.

Their courses have had phenomenal success and have helped more than 100,000 students master AI and data science.

Image error

  User comments    Sort newest first

No comments have been posted yet.



Post anonymous comment
  • Comments need intelligible text (not only emojis or meaningless drivel).
  • No upload requests, visit the forum or message the uploader for this.
  • Use common sense and try to stay on topic.

  • :) :( :D :P :-) B) 8o :? 8) ;) :-* :-( :| O:-D Party Pirates Yuk Facepalm :-@ :o) Pacman Shit Alien eyes Ass Warn Help Bad Love Joystick Boom Eggplant Floppy TV Ghost Note Msg


    CAPTCHA Image 

    Anonymous comments have a moderation delay and show up after 15 minutes