What's new
Warez.Ge

This is a sample guest message. Register a free account today to become a member! Once signed in, you'll be able to participate on this site by adding your own topics and posts, as well as connect with other members through your own private inbox!

Mastering Large Datasets with Python Parallelize and Distribute Your Python Code

voska89

Moderator
Staff member
Top Poster Of Month
a68080564bd6a936edaba9105162b2b7.jpeg

Mastering Large Datasets with Python: Parallelize and Distribute Your Python Code by John T. Wolohan
English | January 21, 2020 | ISBN: 1617296236 | True EPUB | 312 pages | 7.2 MB
Summary

Modern data science solutions need to be clean, easy to read, and scalable. In Mastering Large Datasets with Python, author J.T. Wolohan teaches you how to take a small project and scale it up using a functionally influenced approach to Python coding. You'll explore methods and built-in Python tools that lend themselves to clarity and scalability, like the high-performing parallelism method, as well as distributed technologies that allow for high data throughput. The abundant hands-on exercises in this practical tutorial will lock in these essential skills for any large-scale data science project.
About the technology
Programming techniques that work well on laptop-sized data can slow to a crawl-or fail altogether-when applied to massive files or distributed datasets. By mastering the powerful map and reduce paradigm, along with the Python-based tools that support it, you can write data-centric applications that scale efficiently without requiring codebase rewrites as your requirements change.
About the book
Mastering Large Datasets with Python teaches you to write code that can handle datasets of any size. You'll start with laptop-sized datasets that teach you to parallelize data analysis by breaking large tasks into smaller ones that can run simultaneously. You'll then scale those same programs to industrial-sized datasets on a cluster of cloud servers. With the map and reduce paradigm firmly in place, you'll explore tools like Hadoop and PySpark to efficiently process massive distributed datasets, speed up decision-making with machine learning, and simplify your data storage with AWS S3.
What's inside
An introduction to the map and reduce paradigm
Parallelization with the multiprocessing module and pathos frameworkHadoop and Spark for distributed computingRunning AWS jobs to process large datasets
About the reader
For Python programmers who need to work faster with more data.


Code:
https://hot4share.com/o1th7m3xkhzk/v6tqi.M.L.D.w.P.P.a.D.Y.P.C.rar.html
Uploadgig
https://uploadgig.com/file/download/af05d1934837CF22/v6tqi.M.L.D.w.P.P.a.D.Y.P.C.rar
Rapidgator
https://rapidgator.net/file/2ef0092a5ba4c1656db99f44ffe7a6d6/v6tqi.M.L.D.w.P.P.a.D.Y.P.C.rar.html
NitroFlare
https://nitro.download/view/6065962CEEB4F03/v6tqi.M.L.D.w.P.P.a.D.Y.P.C.rar
Links are Interchangeable - No Password - Single Extraction
 

Users who are viewing this thread

Back
Top