Author: John Wolohan
Publisher: Simon and Schuster
ISBN: 1638350361
Category : Computers
Languages : en
Pages : 451
Book Description
Summary Modern data science solutions need to be clean, easy to read, and scalable. In Mastering Large Datasets with Python, author J.T. Wolohan teaches you how to take a small project and scale it up using a functionally influenced approach to Python coding. You’ll explore methods and built-in Python tools that lend themselves to clarity and scalability, like the high-performing parallelism method, as well as distributed technologies that allow for high data throughput. The abundant hands-on exercises in this practical tutorial will lock in these essential skills for any large-scale data science project. Purchase of the print book includes a free eBook in PDF, Kindle, and ePub formats from Manning Publications. About the technology Programming techniques that work well on laptop-sized data can slow to a crawl—or fail altogether—when applied to massive files or distributed datasets. By mastering the powerful map and reduce paradigm, along with the Python-based tools that support it, you can write data-centric applications that scale efficiently without requiring codebase rewrites as your requirements change. About the book Mastering Large Datasets with Python teaches you to write code that can handle datasets of any size. You’ll start with laptop-sized datasets that teach you to parallelize data analysis by breaking large tasks into smaller ones that can run simultaneously. You’ll then scale those same programs to industrial-sized datasets on a cluster of cloud servers. With the map and reduce paradigm firmly in place, you’ll explore tools like Hadoop and PySpark to efficiently process massive distributed datasets, speed up decision-making with machine learning, and simplify your data storage with AWS S3. What's inside An introduction to the map and reduce paradigm Parallelization with the multiprocessing module and pathos framework Hadoop and Spark for distributed computing Running AWS jobs to process large datasets About the reader For Python programmers who need to work faster with more data. About the author J. T. Wolohan is a lead data scientist at Booz Allen Hamilton, and a PhD researcher at Indiana University, Bloomington. Table of Contents: PART 1 1 ¦ Introduction 2 ¦ Accelerating large dataset work: Map and parallel computing 3 ¦ Function pipelines for mapping complex transformations 4 ¦ Processing large datasets with lazy workflows 5 ¦ Accumulation operations with reduce 6 ¦ Speeding up map and reduce with advanced parallelization PART 2 7 ¦ Processing truly big datasets with Hadoop and Spark 8 ¦ Best practices for large data with Apache Streaming and mrjob 9 ¦ PageRank with map and reduce in PySpark 10 ¦ Faster decision-making with machine learning and PySpark PART 3 11 ¦ Large datasets in the cloud with Amazon Web Services and S3 12 ¦ MapReduce in the cloud with Amazon’s Elastic MapReduce
Mastering Large Datasets with Python
Author: John Wolohan
Publisher: Simon and Schuster
ISBN: 1638350361
Category : Computers
Languages : en
Pages : 451
Book Description
Summary Modern data science solutions need to be clean, easy to read, and scalable. In Mastering Large Datasets with Python, author J.T. Wolohan teaches you how to take a small project and scale it up using a functionally influenced approach to Python coding. You’ll explore methods and built-in Python tools that lend themselves to clarity and scalability, like the high-performing parallelism method, as well as distributed technologies that allow for high data throughput. The abundant hands-on exercises in this practical tutorial will lock in these essential skills for any large-scale data science project. Purchase of the print book includes a free eBook in PDF, Kindle, and ePub formats from Manning Publications. About the technology Programming techniques that work well on laptop-sized data can slow to a crawl—or fail altogether—when applied to massive files or distributed datasets. By mastering the powerful map and reduce paradigm, along with the Python-based tools that support it, you can write data-centric applications that scale efficiently without requiring codebase rewrites as your requirements change. About the book Mastering Large Datasets with Python teaches you to write code that can handle datasets of any size. You’ll start with laptop-sized datasets that teach you to parallelize data analysis by breaking large tasks into smaller ones that can run simultaneously. You’ll then scale those same programs to industrial-sized datasets on a cluster of cloud servers. With the map and reduce paradigm firmly in place, you’ll explore tools like Hadoop and PySpark to efficiently process massive distributed datasets, speed up decision-making with machine learning, and simplify your data storage with AWS S3. What's inside An introduction to the map and reduce paradigm Parallelization with the multiprocessing module and pathos framework Hadoop and Spark for distributed computing Running AWS jobs to process large datasets About the reader For Python programmers who need to work faster with more data. About the author J. T. Wolohan is a lead data scientist at Booz Allen Hamilton, and a PhD researcher at Indiana University, Bloomington. Table of Contents: PART 1 1 ¦ Introduction 2 ¦ Accelerating large dataset work: Map and parallel computing 3 ¦ Function pipelines for mapping complex transformations 4 ¦ Processing large datasets with lazy workflows 5 ¦ Accumulation operations with reduce 6 ¦ Speeding up map and reduce with advanced parallelization PART 2 7 ¦ Processing truly big datasets with Hadoop and Spark 8 ¦ Best practices for large data with Apache Streaming and mrjob 9 ¦ PageRank with map and reduce in PySpark 10 ¦ Faster decision-making with machine learning and PySpark PART 3 11 ¦ Large datasets in the cloud with Amazon Web Services and S3 12 ¦ MapReduce in the cloud with Amazon’s Elastic MapReduce
Publisher: Simon and Schuster
ISBN: 1638350361
Category : Computers
Languages : en
Pages : 451
Book Description
Summary Modern data science solutions need to be clean, easy to read, and scalable. In Mastering Large Datasets with Python, author J.T. Wolohan teaches you how to take a small project and scale it up using a functionally influenced approach to Python coding. You’ll explore methods and built-in Python tools that lend themselves to clarity and scalability, like the high-performing parallelism method, as well as distributed technologies that allow for high data throughput. The abundant hands-on exercises in this practical tutorial will lock in these essential skills for any large-scale data science project. Purchase of the print book includes a free eBook in PDF, Kindle, and ePub formats from Manning Publications. About the technology Programming techniques that work well on laptop-sized data can slow to a crawl—or fail altogether—when applied to massive files or distributed datasets. By mastering the powerful map and reduce paradigm, along with the Python-based tools that support it, you can write data-centric applications that scale efficiently without requiring codebase rewrites as your requirements change. About the book Mastering Large Datasets with Python teaches you to write code that can handle datasets of any size. You’ll start with laptop-sized datasets that teach you to parallelize data analysis by breaking large tasks into smaller ones that can run simultaneously. You’ll then scale those same programs to industrial-sized datasets on a cluster of cloud servers. With the map and reduce paradigm firmly in place, you’ll explore tools like Hadoop and PySpark to efficiently process massive distributed datasets, speed up decision-making with machine learning, and simplify your data storage with AWS S3. What's inside An introduction to the map and reduce paradigm Parallelization with the multiprocessing module and pathos framework Hadoop and Spark for distributed computing Running AWS jobs to process large datasets About the reader For Python programmers who need to work faster with more data. About the author J. T. Wolohan is a lead data scientist at Booz Allen Hamilton, and a PhD researcher at Indiana University, Bloomington. Table of Contents: PART 1 1 ¦ Introduction 2 ¦ Accelerating large dataset work: Map and parallel computing 3 ¦ Function pipelines for mapping complex transformations 4 ¦ Processing large datasets with lazy workflows 5 ¦ Accumulation operations with reduce 6 ¦ Speeding up map and reduce with advanced parallelization PART 2 7 ¦ Processing truly big datasets with Hadoop and Spark 8 ¦ Best practices for large data with Apache Streaming and mrjob 9 ¦ PageRank with map and reduce in PySpark 10 ¦ Faster decision-making with machine learning and PySpark PART 3 11 ¦ Large datasets in the cloud with Amazon Web Services and S3 12 ¦ MapReduce in the cloud with Amazon’s Elastic MapReduce
Mastering Large Datasets
Author: J. T. Wolohan
Publisher: Manning Publications
ISBN: 9781617296239
Category :
Languages : en
Pages : 350
Book Description
With an emphasis on clarity, style, and performance, author J.T. Wolohan expertly guides you through implementing a functionally-influenced approach to Python coding. You'll get familiar with Python's functional built-ins like the functools operator and itertools modules, as well as the toolz library. Mastering Large Datasets teaches you to write easily readable, easily scalable Python code that can efficiently process large volumes of structured and unstructured data. By the end of this comprehensive guide, you'll have a solid grasp on the tools and methods that will take your code beyond the laptop and your data science career to the next level! Purchase of the print book includes a free eBook in PDF, Kindle, and ePub formats from Manning Publications.
Publisher: Manning Publications
ISBN: 9781617296239
Category :
Languages : en
Pages : 350
Book Description
With an emphasis on clarity, style, and performance, author J.T. Wolohan expertly guides you through implementing a functionally-influenced approach to Python coding. You'll get familiar with Python's functional built-ins like the functools operator and itertools modules, as well as the toolz library. Mastering Large Datasets teaches you to write easily readable, easily scalable Python code that can efficiently process large volumes of structured and unstructured data. By the end of this comprehensive guide, you'll have a solid grasp on the tools and methods that will take your code beyond the laptop and your data science career to the next level! Purchase of the print book includes a free eBook in PDF, Kindle, and ePub formats from Manning Publications.
Mastering Large Language Models with Python
Author: Raj Arun R
Publisher: Orange Education Pvt Ltd
ISBN: 8197081824
Category : Computers
Languages : en
Pages : 547
Book Description
A Comprehensive Guide to Leverage Generative AI in the Modern Enterprise KEY FEATURES ● Gain a comprehensive understanding of LLMs within the framework of Generative AI, from foundational concepts to advanced applications. ● Dive into practical exercises and real-world applications, accompanied by detailed code walkthroughs in Python. ● Explore LLMOps with a dedicated focus on ensuring trustworthy AI and best practices for deploying, managing, and maintaining LLMs in enterprise settings. ● Prioritize the ethical and responsible use of LLMs, with an emphasis on building models that adhere to principles of fairness, transparency, and accountability, fostering trust in AI technologies. DESCRIPTION “Mastering Large Language Models with Python” is an indispensable resource that offers a comprehensive exploration of Large Language Models (LLMs), providing the essential knowledge to leverage these transformative AI models effectively. From unraveling the intricacies of LLM architecture to practical applications like code generation and AI-driven recommendation systems, readers will gain valuable insights into implementing LLMs in diverse projects. Covering both open-source and proprietary LLMs, the book delves into foundational concepts and advanced techniques, empowering professionals to harness the full potential of these models. Detailed discussions on quantization techniques for efficient deployment, operational strategies with LLMOps, and ethical considerations ensure a well-rounded understanding of LLM implementation. Through real-world case studies, code snippets, and practical examples, readers will navigate the complexities of LLMs with confidence, paving the way for innovative solutions and organizational growth. Whether you seek to deepen your understanding, drive impactful applications, or lead AI-driven initiatives, this book equips you with the tools and insights needed to excel in the dynamic landscape of artificial intelligence. WHAT WILL YOU LEARN ● In-depth study of LLM architecture and its versatile applications across industries. ● Harness open-source and proprietary LLMs to craft innovative solutions. ● Implement LLM APIs for a wide range of tasks spanning natural language processing, audio analysis, and visual recognition. ● Optimize LLM deployment through techniques such as quantization and operational strategies like LLMOps, ensuring efficient and scalable model usage. ● Master prompt engineering techniques to fine-tune LLM outputs, enhancing quality and relevance for diverse use cases. ● Navigate the complex landscape of ethical AI development, prioritizing responsible practices to drive impactful technology adoption and advancement. WHO IS THIS BOOK FOR? This book is tailored for software engineers, data scientists, AI researchers, and technology leaders with a foundational understanding of machine learning concepts and programming. It's ideal for those looking to deepen their knowledge of Large Language Models and their practical applications in the field of AI. If you aim to explore LLMs extensively for implementing inventive solutions or spearheading AI-driven projects, this book is tailored to your needs. TABLE OF CONTENTS 1. The Basics of Large Language Models and Their Applications 2. Demystifying Open-Source Large Language Models 3. Closed-Source Large Language Models 4. LLM APIs for Various Large Language Model Tasks 5. Integrating Cohere API in Google Sheets 6. Dynamic Movie Recommendation Engine Using LLMs 7. Document-and Web-based QA Bots with Large Language Models 8. LLM Quantization Techniques and Implementation 9. Fine-tuning and Evaluation of LLMs 10. Recipes for Fine-Tuning and Evaluating LLMs 11. LLMOps - Operationalizing LLMs at Scale 12. Implementing LLMOps in Practice Using MLflow on Databricks 13. Mastering the Art of Prompt Engineering 14. Prompt Engineering Essentials and Design Patterns 15. Ethical Considerations and Regulatory Frameworks for LLMs 16. Towards Trustworthy Generative AI (A Novel Framework Inspired by Symbolic Reasoning) Index
Publisher: Orange Education Pvt Ltd
ISBN: 8197081824
Category : Computers
Languages : en
Pages : 547
Book Description
A Comprehensive Guide to Leverage Generative AI in the Modern Enterprise KEY FEATURES ● Gain a comprehensive understanding of LLMs within the framework of Generative AI, from foundational concepts to advanced applications. ● Dive into practical exercises and real-world applications, accompanied by detailed code walkthroughs in Python. ● Explore LLMOps with a dedicated focus on ensuring trustworthy AI and best practices for deploying, managing, and maintaining LLMs in enterprise settings. ● Prioritize the ethical and responsible use of LLMs, with an emphasis on building models that adhere to principles of fairness, transparency, and accountability, fostering trust in AI technologies. DESCRIPTION “Mastering Large Language Models with Python” is an indispensable resource that offers a comprehensive exploration of Large Language Models (LLMs), providing the essential knowledge to leverage these transformative AI models effectively. From unraveling the intricacies of LLM architecture to practical applications like code generation and AI-driven recommendation systems, readers will gain valuable insights into implementing LLMs in diverse projects. Covering both open-source and proprietary LLMs, the book delves into foundational concepts and advanced techniques, empowering professionals to harness the full potential of these models. Detailed discussions on quantization techniques for efficient deployment, operational strategies with LLMOps, and ethical considerations ensure a well-rounded understanding of LLM implementation. Through real-world case studies, code snippets, and practical examples, readers will navigate the complexities of LLMs with confidence, paving the way for innovative solutions and organizational growth. Whether you seek to deepen your understanding, drive impactful applications, or lead AI-driven initiatives, this book equips you with the tools and insights needed to excel in the dynamic landscape of artificial intelligence. WHAT WILL YOU LEARN ● In-depth study of LLM architecture and its versatile applications across industries. ● Harness open-source and proprietary LLMs to craft innovative solutions. ● Implement LLM APIs for a wide range of tasks spanning natural language processing, audio analysis, and visual recognition. ● Optimize LLM deployment through techniques such as quantization and operational strategies like LLMOps, ensuring efficient and scalable model usage. ● Master prompt engineering techniques to fine-tune LLM outputs, enhancing quality and relevance for diverse use cases. ● Navigate the complex landscape of ethical AI development, prioritizing responsible practices to drive impactful technology adoption and advancement. WHO IS THIS BOOK FOR? This book is tailored for software engineers, data scientists, AI researchers, and technology leaders with a foundational understanding of machine learning concepts and programming. It's ideal for those looking to deepen their knowledge of Large Language Models and their practical applications in the field of AI. If you aim to explore LLMs extensively for implementing inventive solutions or spearheading AI-driven projects, this book is tailored to your needs. TABLE OF CONTENTS 1. The Basics of Large Language Models and Their Applications 2. Demystifying Open-Source Large Language Models 3. Closed-Source Large Language Models 4. LLM APIs for Various Large Language Model Tasks 5. Integrating Cohere API in Google Sheets 6. Dynamic Movie Recommendation Engine Using LLMs 7. Document-and Web-based QA Bots with Large Language Models 8. LLM Quantization Techniques and Implementation 9. Fine-tuning and Evaluation of LLMs 10. Recipes for Fine-Tuning and Evaluating LLMs 11. LLMOps - Operationalizing LLMs at Scale 12. Implementing LLMOps in Practice Using MLflow on Databricks 13. Mastering the Art of Prompt Engineering 14. Prompt Engineering Essentials and Design Patterns 15. Ethical Considerations and Regulatory Frameworks for LLMs 16. Towards Trustworthy Generative AI (A Novel Framework Inspired by Symbolic Reasoning) Index
Extending Power BI with Python and R
Author: Luca Zavarella
Publisher: Packt Publishing Ltd
ISBN: 1801076677
Category : Computers
Languages : en
Pages : 559
Book Description
Perform more advanced analysis and manipulation of your data beyond what Power BI can do to unlock valuable insights using Python and R Key FeaturesGet the most out of Python and R with Power BI by implementing non-trivial codeLeverage the toolset of Python and R chunks to inject scripts into your Power BI dashboardsImplement new techniques for ingesting, enriching, and visualizing data with Python and R in Power BIBook Description Python and R allow you to extend Power BI capabilities to simplify ingestion and transformation activities, enhance dashboards, and highlight insights. With this book, you'll be able to make your artifacts far more interesting and rich in insights using analytical languages. You'll start by learning how to configure your Power BI environment to use your Python and R scripts. The book then explores data ingestion and data transformation extensions, and advances to focus on data augmentation and data visualization. You'll understand how to import data from external sources and transform them using complex algorithms. The book helps you implement personal data de-identification methods such as pseudonymization, anonymization, and masking in Power BI. You'll be able to call external APIs to enrich your data much more quickly using Python programming and R programming. Later, you'll learn advanced Python and R techniques to perform in-depth analysis and extract valuable information using statistics and machine learning. You'll also understand the main statistical features of datasets by plotting multiple visual graphs in the process of creating a machine learning model. By the end of this book, you'll be able to enrich your Power BI data models and visualizations using complex algorithms in Python and R. What you will learnDiscover best practices for using Python and R in Power BI productsUse Python and R to perform complex data manipulations in Power BIApply data anonymization and data pseudonymization in Power BILog data and load large datasets in Power BI using Python and REnrich your Power BI dashboards using external APIs and machine learning modelsExtract insights from your data using linear optimization and other algorithmsHandle outliers and missing values for multivariate and time-series dataCreate any visualization, as complex as you want, using R scriptsWho this book is for This book is for business analysts, business intelligence professionals, and data scientists who already use Microsoft Power BI and want to add more value to their analysis using Python and R. Working knowledge of Power BI is required to make the most of this book. Basic knowledge of Python and R will also be helpful.
Publisher: Packt Publishing Ltd
ISBN: 1801076677
Category : Computers
Languages : en
Pages : 559
Book Description
Perform more advanced analysis and manipulation of your data beyond what Power BI can do to unlock valuable insights using Python and R Key FeaturesGet the most out of Python and R with Power BI by implementing non-trivial codeLeverage the toolset of Python and R chunks to inject scripts into your Power BI dashboardsImplement new techniques for ingesting, enriching, and visualizing data with Python and R in Power BIBook Description Python and R allow you to extend Power BI capabilities to simplify ingestion and transformation activities, enhance dashboards, and highlight insights. With this book, you'll be able to make your artifacts far more interesting and rich in insights using analytical languages. You'll start by learning how to configure your Power BI environment to use your Python and R scripts. The book then explores data ingestion and data transformation extensions, and advances to focus on data augmentation and data visualization. You'll understand how to import data from external sources and transform them using complex algorithms. The book helps you implement personal data de-identification methods such as pseudonymization, anonymization, and masking in Power BI. You'll be able to call external APIs to enrich your data much more quickly using Python programming and R programming. Later, you'll learn advanced Python and R techniques to perform in-depth analysis and extract valuable information using statistics and machine learning. You'll also understand the main statistical features of datasets by plotting multiple visual graphs in the process of creating a machine learning model. By the end of this book, you'll be able to enrich your Power BI data models and visualizations using complex algorithms in Python and R. What you will learnDiscover best practices for using Python and R in Power BI productsUse Python and R to perform complex data manipulations in Power BIApply data anonymization and data pseudonymization in Power BILog data and load large datasets in Power BI using Python and REnrich your Power BI dashboards using external APIs and machine learning modelsExtract insights from your data using linear optimization and other algorithmsHandle outliers and missing values for multivariate and time-series dataCreate any visualization, as complex as you want, using R scriptsWho this book is for This book is for business analysts, business intelligence professionals, and data scientists who already use Microsoft Power BI and want to add more value to their analysis using Python and R. Working knowledge of Power BI is required to make the most of this book. Basic knowledge of Python and R will also be helpful.
Data Engineering with Python
Author: Paul Crickard
Publisher: Packt Publishing Ltd
ISBN: 1839212306
Category : Computers
Languages : en
Pages : 357
Book Description
Build, monitor, and manage real-time data pipelines to create data engineering infrastructure efficiently using open-source Apache projects Key Features Become well-versed in data architectures, data preparation, and data optimization skills with the help of practical examples Design data models and learn how to extract, transform, and load (ETL) data using Python Schedule, automate, and monitor complex data pipelines in production Book DescriptionData engineering provides the foundation for data science and analytics, and forms an important part of all businesses. This book will help you to explore various tools and methods that are used for understanding the data engineering process using Python. The book will show you how to tackle challenges commonly faced in different aspects of data engineering. You’ll start with an introduction to the basics of data engineering, along with the technologies and frameworks required to build data pipelines to work with large datasets. You’ll learn how to transform and clean data and perform analytics to get the most out of your data. As you advance, you'll discover how to work with big data of varying complexity and production databases, and build data pipelines. Using real-world examples, you’ll build architectures on which you’ll learn how to deploy data pipelines. By the end of this Python book, you’ll have gained a clear understanding of data modeling techniques, and will be able to confidently build data engineering pipelines for tracking data, running quality checks, and making necessary changes in production.What you will learn Understand how data engineering supports data science workflows Discover how to extract data from files and databases and then clean, transform, and enrich it Configure processors for handling different file formats as well as both relational and NoSQL databases Find out how to implement a data pipeline and dashboard to visualize results Use staging and validation to check data before landing in the warehouse Build real-time pipelines with staging areas that perform validation and handle failures Get to grips with deploying pipelines in the production environment Who this book is for This book is for data analysts, ETL developers, and anyone looking to get started with or transition to the field of data engineering or refresh their knowledge of data engineering using Python. This book will also be useful for students planning to build a career in data engineering or IT professionals preparing for a transition. No previous knowledge of data engineering is required.
Publisher: Packt Publishing Ltd
ISBN: 1839212306
Category : Computers
Languages : en
Pages : 357
Book Description
Build, monitor, and manage real-time data pipelines to create data engineering infrastructure efficiently using open-source Apache projects Key Features Become well-versed in data architectures, data preparation, and data optimization skills with the help of practical examples Design data models and learn how to extract, transform, and load (ETL) data using Python Schedule, automate, and monitor complex data pipelines in production Book DescriptionData engineering provides the foundation for data science and analytics, and forms an important part of all businesses. This book will help you to explore various tools and methods that are used for understanding the data engineering process using Python. The book will show you how to tackle challenges commonly faced in different aspects of data engineering. You’ll start with an introduction to the basics of data engineering, along with the technologies and frameworks required to build data pipelines to work with large datasets. You’ll learn how to transform and clean data and perform analytics to get the most out of your data. As you advance, you'll discover how to work with big data of varying complexity and production databases, and build data pipelines. Using real-world examples, you’ll build architectures on which you’ll learn how to deploy data pipelines. By the end of this Python book, you’ll have gained a clear understanding of data modeling techniques, and will be able to confidently build data engineering pipelines for tracking data, running quality checks, and making necessary changes in production.What you will learn Understand how data engineering supports data science workflows Discover how to extract data from files and databases and then clean, transform, and enrich it Configure processors for handling different file formats as well as both relational and NoSQL databases Find out how to implement a data pipeline and dashboard to visualize results Use staging and validation to check data before landing in the warehouse Build real-time pipelines with staging areas that perform validation and handle failures Get to grips with deploying pipelines in the production environment Who this book is for This book is for data analysts, ETL developers, and anyone looking to get started with or transition to the field of data engineering or refresh their knowledge of data engineering using Python. This book will also be useful for students planning to build a career in data engineering or IT professionals preparing for a transition. No previous knowledge of data engineering is required.
Mastering Python Scientific Computing
Author: Hemant Kumar Mehta
Publisher: Packt Publishing Ltd
ISBN: 1783288833
Category : Computers
Languages : en
Pages : 301
Book Description
A complete guide for Python programmers to master scientific computing using Python APIs and tools About This Book The basics of scientific computing to advanced concepts involving parallel and large scale computation are all covered. Most of the Python APIs and tools used in scientific computing are discussed in detail The concepts are discussed with suitable example programs Who This Book Is For If you are a Python programmer and want to get your hands on scientific computing, this book is for you. The book expects you to have had exposure to various concepts of Python programming. What You Will Learn Fundamentals and components of scientific computing Scientific computing data management Performing numerical computing using NumPy and SciPy Concepts and programming for symbolic computing using SymPy Using the plotting library matplotlib for data visualization Data analysis and visualization using Pandas, matplotlib, and IPython Performing parallel and high performance computing Real-life case studies and best practices of scientific computing In Detail In today's world, along with theoretical and experimental work, scientific computing has become an important part of scientific disciplines. Numerical calculations, simulations and computer modeling in this day and age form the vast majority of both experimental and theoretical papers. In the scientific method, replication and reproducibility are two important contributing factors. A complete and concrete scientific result should be reproducible and replicable. Python is suitable for scientific computing. A large community of users, plenty of help and documentation, a large collection of scientific libraries and environments, great performance, and good support makes Python a great choice for scientific computing. At present Python is among the top choices for developing scientific workflow and the book targets existing Python developers to master this domain using Python. The main things to learn in the book are the concept of scientific workflow, managing scientific workflow data and performing computation on this data using Python. The book discusses NumPy, SciPy, SymPy, matplotlib, Pandas and IPython with several example programs. Style and approach This book follows a hands-on approach to explain the complex concepts related to scientific computing. It details various APIs using appropriate examples.
Publisher: Packt Publishing Ltd
ISBN: 1783288833
Category : Computers
Languages : en
Pages : 301
Book Description
A complete guide for Python programmers to master scientific computing using Python APIs and tools About This Book The basics of scientific computing to advanced concepts involving parallel and large scale computation are all covered. Most of the Python APIs and tools used in scientific computing are discussed in detail The concepts are discussed with suitable example programs Who This Book Is For If you are a Python programmer and want to get your hands on scientific computing, this book is for you. The book expects you to have had exposure to various concepts of Python programming. What You Will Learn Fundamentals and components of scientific computing Scientific computing data management Performing numerical computing using NumPy and SciPy Concepts and programming for symbolic computing using SymPy Using the plotting library matplotlib for data visualization Data analysis and visualization using Pandas, matplotlib, and IPython Performing parallel and high performance computing Real-life case studies and best practices of scientific computing In Detail In today's world, along with theoretical and experimental work, scientific computing has become an important part of scientific disciplines. Numerical calculations, simulations and computer modeling in this day and age form the vast majority of both experimental and theoretical papers. In the scientific method, replication and reproducibility are two important contributing factors. A complete and concrete scientific result should be reproducible and replicable. Python is suitable for scientific computing. A large community of users, plenty of help and documentation, a large collection of scientific libraries and environments, great performance, and good support makes Python a great choice for scientific computing. At present Python is among the top choices for developing scientific workflow and the book targets existing Python developers to master this domain using Python. The main things to learn in the book are the concept of scientific workflow, managing scientific workflow data and performing computation on this data using Python. The book discusses NumPy, SciPy, SymPy, matplotlib, Pandas and IPython with several example programs. Style and approach This book follows a hands-on approach to explain the complex concepts related to scientific computing. It details various APIs using appropriate examples.
Spark in Action
Author: Jean-Georges Perrin
Publisher: Simon and Schuster
ISBN: 1638351309
Category : Computers
Languages : en
Pages : 574
Book Description
Summary The Spark distributed data processing platform provides an easy-to-implement tool for ingesting, streaming, and processing data from any source. In Spark in Action, Second Edition, you’ll learn to take advantage of Spark’s core features and incredible processing speed, with applications including real-time computation, delayed evaluation, and machine learning. Spark skills are a hot commodity in enterprises worldwide, and with Spark’s powerful and flexible Java APIs, you can reap all the benefits without first learning Scala or Hadoop. Foreword by Rob Thomas. About the technology Analyzing enterprise data starts by reading, filtering, and merging files and streams from many sources. The Spark data processing engine handles this varied volume like a champ, delivering speeds 100 times faster than Hadoop systems. Thanks to SQL support, an intuitive interface, and a straightforward multilanguage API, you can use Spark without learning a complex new ecosystem. About the book Spark in Action, Second Edition, teaches you to create end-to-end analytics applications. In this entirely new book, you’ll learn from interesting Java-based examples, including a complete data pipeline for processing NASA satellite data. And you’ll discover Java, Python, and Scala code samples hosted on GitHub that you can explore and adapt, plus appendixes that give you a cheat sheet for installing tools and understanding Spark-specific terms. What's inside Writing Spark applications in Java Spark application architecture Ingestion through files, databases, streaming, and Elasticsearch Querying distributed datasets with Spark SQL About the reader This book does not assume previous experience with Spark, Scala, or Hadoop. About the author Jean-Georges Perrin is an experienced data and software architect. He is France’s first IBM Champion and has been honored for 12 consecutive years. Table of Contents PART 1 - THE THEORY CRIPPLED BY AWESOME EXAMPLES 1 So, what is Spark, anyway? 2 Architecture and flow 3 The majestic role of the dataframe 4 Fundamentally lazy 5 Building a simple app for deployment 6 Deploying your simple app PART 2 - INGESTION 7 Ingestion from files 8 Ingestion from databases 9 Advanced ingestion: finding data sources and building your own 10 Ingestion through structured streaming PART 3 - TRANSFORMING YOUR DATA 11 Working with SQL 12 Transforming your data 13 Transforming entire documents 14 Extending transformations with user-defined functions 15 Aggregating your data PART 4 - GOING FURTHER 16 Cache and checkpoint: Enhancing Spark’s performances 17 Exporting data and building full data pipelines 18 Exploring deployment
Publisher: Simon and Schuster
ISBN: 1638351309
Category : Computers
Languages : en
Pages : 574
Book Description
Summary The Spark distributed data processing platform provides an easy-to-implement tool for ingesting, streaming, and processing data from any source. In Spark in Action, Second Edition, you’ll learn to take advantage of Spark’s core features and incredible processing speed, with applications including real-time computation, delayed evaluation, and machine learning. Spark skills are a hot commodity in enterprises worldwide, and with Spark’s powerful and flexible Java APIs, you can reap all the benefits without first learning Scala or Hadoop. Foreword by Rob Thomas. About the technology Analyzing enterprise data starts by reading, filtering, and merging files and streams from many sources. The Spark data processing engine handles this varied volume like a champ, delivering speeds 100 times faster than Hadoop systems. Thanks to SQL support, an intuitive interface, and a straightforward multilanguage API, you can use Spark without learning a complex new ecosystem. About the book Spark in Action, Second Edition, teaches you to create end-to-end analytics applications. In this entirely new book, you’ll learn from interesting Java-based examples, including a complete data pipeline for processing NASA satellite data. And you’ll discover Java, Python, and Scala code samples hosted on GitHub that you can explore and adapt, plus appendixes that give you a cheat sheet for installing tools and understanding Spark-specific terms. What's inside Writing Spark applications in Java Spark application architecture Ingestion through files, databases, streaming, and Elasticsearch Querying distributed datasets with Spark SQL About the reader This book does not assume previous experience with Spark, Scala, or Hadoop. About the author Jean-Georges Perrin is an experienced data and software architect. He is France’s first IBM Champion and has been honored for 12 consecutive years. Table of Contents PART 1 - THE THEORY CRIPPLED BY AWESOME EXAMPLES 1 So, what is Spark, anyway? 2 Architecture and flow 3 The majestic role of the dataframe 4 Fundamentally lazy 5 Building a simple app for deployment 6 Deploying your simple app PART 2 - INGESTION 7 Ingestion from files 8 Ingestion from databases 9 Advanced ingestion: finding data sources and building your own 10 Ingestion through structured streaming PART 3 - TRANSFORMING YOUR DATA 11 Working with SQL 12 Transforming your data 13 Transforming entire documents 14 Extending transformations with user-defined functions 15 Aggregating your data PART 4 - GOING FURTHER 16 Cache and checkpoint: Enhancing Spark’s performances 17 Exporting data and building full data pipelines 18 Exploring deployment
Python for Finance
Author: Yves J. Hilpisch
Publisher: "O'Reilly Media, Inc."
ISBN: 1492024295
Category : Computers
Languages : en
Pages : 682
Book Description
The financial industry has recently adopted Python at a tremendous rate, with some of the largest investment banks and hedge funds using it to build core trading and risk management systems. Updated for Python 3, the second edition of this hands-on book helps you get started with the language, guiding developers and quantitative analysts through Python libraries and tools for building financial applications and interactive financial analytics. Using practical examples throughout the book, author Yves Hilpisch also shows you how to develop a full-fledged framework for Monte Carlo simulation-based derivatives and risk analytics, based on a large, realistic case study. Much of the book uses interactive IPython Notebooks.
Publisher: "O'Reilly Media, Inc."
ISBN: 1492024295
Category : Computers
Languages : en
Pages : 682
Book Description
The financial industry has recently adopted Python at a tremendous rate, with some of the largest investment banks and hedge funds using it to build core trading and risk management systems. Updated for Python 3, the second edition of this hands-on book helps you get started with the language, guiding developers and quantitative analysts through Python libraries and tools for building financial applications and interactive financial analytics. Using practical examples throughout the book, author Yves Hilpisch also shows you how to develop a full-fledged framework for Monte Carlo simulation-based derivatives and risk analytics, based on a large, realistic case study. Much of the book uses interactive IPython Notebooks.
Frank Kane's Taming Big Data with Apache Spark and Python
Author: Frank Kane
Publisher: Packt Publishing Ltd
ISBN: 1787288307
Category : Computers
Languages : en
Pages : 289
Book Description
Frank Kane's hands-on Spark training course, based on his bestselling Taming Big Data with Apache Spark and Python video, now available in a book. Understand and analyze large data sets using Spark on a single system or on a cluster. About This Book Understand how Spark can be distributed across computing clusters Develop and run Spark jobs efficiently using Python A hands-on tutorial by Frank Kane with over 15 real-world examples teaching you Big Data processing with Spark Who This Book Is For If you are a data scientist or data analyst who wants to learn Big Data processing using Apache Spark and Python, this book is for you. If you have some programming experience in Python, and want to learn how to process large amounts of data using Apache Spark, Frank Kane's Taming Big Data with Apache Spark and Python will also help you. What You Will Learn Find out how you can identify Big Data problems as Spark problems Install and run Apache Spark on your computer or on a cluster Analyze large data sets across many CPUs using Spark's Resilient Distributed Datasets Implement machine learning on Spark using the MLlib library Process continuous streams of data in real time using the Spark streaming module Perform complex network analysis using Spark's GraphX library Use Amazon's Elastic MapReduce service to run your Spark jobs on a cluster In Detail Frank Kane's Taming Big Data with Apache Spark and Python is your companion to learning Apache Spark in a hands-on manner. Frank will start you off by teaching you how to set up Spark on a single system or on a cluster, and you'll soon move on to analyzing large data sets using Spark RDD, and developing and running effective Spark jobs quickly using Python. Apache Spark has emerged as the next big thing in the Big Data domain – quickly rising from an ascending technology to an established superstar in just a matter of years. Spark allows you to quickly extract actionable insights from large amounts of data, on a real-time basis, making it an essential tool in many modern businesses. Frank has packed this book with over 15 interactive, fun-filled examples relevant to the real world, and he will empower you to understand the Spark ecosystem and implement production-grade real-time Spark projects with ease. Style and approach Frank Kane's Taming Big Data with Apache Spark and Python is a hands-on tutorial with over 15 real-world examples carefully explained by Frank in a step-by-step manner. The examples vary in complexity, and you can move through them at your own pace.
Publisher: Packt Publishing Ltd
ISBN: 1787288307
Category : Computers
Languages : en
Pages : 289
Book Description
Frank Kane's hands-on Spark training course, based on his bestselling Taming Big Data with Apache Spark and Python video, now available in a book. Understand and analyze large data sets using Spark on a single system or on a cluster. About This Book Understand how Spark can be distributed across computing clusters Develop and run Spark jobs efficiently using Python A hands-on tutorial by Frank Kane with over 15 real-world examples teaching you Big Data processing with Spark Who This Book Is For If you are a data scientist or data analyst who wants to learn Big Data processing using Apache Spark and Python, this book is for you. If you have some programming experience in Python, and want to learn how to process large amounts of data using Apache Spark, Frank Kane's Taming Big Data with Apache Spark and Python will also help you. What You Will Learn Find out how you can identify Big Data problems as Spark problems Install and run Apache Spark on your computer or on a cluster Analyze large data sets across many CPUs using Spark's Resilient Distributed Datasets Implement machine learning on Spark using the MLlib library Process continuous streams of data in real time using the Spark streaming module Perform complex network analysis using Spark's GraphX library Use Amazon's Elastic MapReduce service to run your Spark jobs on a cluster In Detail Frank Kane's Taming Big Data with Apache Spark and Python is your companion to learning Apache Spark in a hands-on manner. Frank will start you off by teaching you how to set up Spark on a single system or on a cluster, and you'll soon move on to analyzing large data sets using Spark RDD, and developing and running effective Spark jobs quickly using Python. Apache Spark has emerged as the next big thing in the Big Data domain – quickly rising from an ascending technology to an established superstar in just a matter of years. Spark allows you to quickly extract actionable insights from large amounts of data, on a real-time basis, making it an essential tool in many modern businesses. Frank has packed this book with over 15 interactive, fun-filled examples relevant to the real world, and he will empower you to understand the Spark ecosystem and implement production-grade real-time Spark projects with ease. Style and approach Frank Kane's Taming Big Data with Apache Spark and Python is a hands-on tutorial with over 15 real-world examples carefully explained by Frank in a step-by-step manner. The examples vary in complexity, and you can move through them at your own pace.
Mastering Automated Machine Learning: Concepts, Tools, and Techniques
Author: Peter Jones
Publisher: Walzone Press
ISBN:
Category : Computers
Languages : en
Pages : 214
Book Description
"Mastering Automated Machine Learning: Concepts, Tools, and Techniques" is an essential guide for anyone seeking to unlock the full potential of Automated Machine Learning (AutoML), a groundbreaking technology transforming the field of data science. By automating complex and time-consuming processes, AutoML is making machine learning more efficient and accessible to a broader range of professionals. This book offers an in-depth exploration of core principles, state-of-the-art methodologies, and the practical tools that define AutoML. From data preparation and feature engineering to model selection, tuning, and deployment, readers will acquire a thorough understanding of how AutoML streamlines the entire machine learning pipeline. Whether you're a data scientist, machine learning engineer, or software developer eager to harness the power of automation, "Mastering Automated Machine Learning" provides the insights you need to implement cutting-edge AutoML solutions. With practical examples and guidance on using Python-based frameworks, this book equips you to revolutionize your data science projects. Embrace the future of machine learning and optimize your workflows with "Mastering Automated Machine Learning: Concepts, Tools, and Techniques."
Publisher: Walzone Press
ISBN:
Category : Computers
Languages : en
Pages : 214
Book Description
"Mastering Automated Machine Learning: Concepts, Tools, and Techniques" is an essential guide for anyone seeking to unlock the full potential of Automated Machine Learning (AutoML), a groundbreaking technology transforming the field of data science. By automating complex and time-consuming processes, AutoML is making machine learning more efficient and accessible to a broader range of professionals. This book offers an in-depth exploration of core principles, state-of-the-art methodologies, and the practical tools that define AutoML. From data preparation and feature engineering to model selection, tuning, and deployment, readers will acquire a thorough understanding of how AutoML streamlines the entire machine learning pipeline. Whether you're a data scientist, machine learning engineer, or software developer eager to harness the power of automation, "Mastering Automated Machine Learning" provides the insights you need to implement cutting-edge AutoML solutions. With practical examples and guidance on using Python-based frameworks, this book equips you to revolutionize your data science projects. Embrace the future of machine learning and optimize your workflows with "Mastering Automated Machine Learning: Concepts, Tools, and Techniques."