Getting Started with DuckDB

Getting Started with DuckDB PDF Author: Simon Aubury
Publisher: Packt Publishing Ltd
ISBN: 1803232536
Category : Computers
Languages : en
Pages : 382

Get Book Here

Book Description
Analyze and transform data efficiently with DuckDB, a versatile, modern, in-process SQL database Key Features Use DuckDB to rapidly load, transform, and query data across a range of sources and formats Gain practical experience using SQL, Python, and R to effectively analyze data Learn how open source tools and cloud services in the broader data ecosystem complement DuckDB’s versatile capabilities Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionDuckDB is a fast in-process analytical database. Its ease of use, versatile feature set, and powerful analytical capabilities make DuckDB a valuable addition to the data practitioner’s toolkit. Getting Started with DuckDB offers a practical overview of DuckDB’s fundamentals and guidance for effectively using its powerful capabilities. Through extensive hands-on examples, you’ll learn how to use DuckDB to load, transform, and query a variety of data sources and formats, including CSV, JSON, and Parquet files, semi-structured data, remotely-hosted files, and external databases. You'll also find out how to leverage DuckDB's performance optimizations and friendly SQL enhancements. You'll explore how to use DuckDB’s extensions for specialized applications, such as geospatial analysis and text search over document collections. In addition to working through examples in SQL, Python, and R, you’ll also dive into using DuckDB for analyzing public datasets and discover the wider ecosystem of open-source tools and cloud services that supercharge DuckDB-powered workflows and applications. Whether you’re a seasoned data practitioner or new to working with analytical data, this book will rapidly get you up to speed with DuckDB’s versatile and powerful capabilities, enabling you to apply them in your analytical workflows and projects.What you will learn Understand the properties and applications of a columnar in-process database Use SQL to load, transform, and query a range of data formats Discover DuckDB's rich extensions and learn how to apply them Use nested data types to model semi-structured data and extract and model JSON data Integrate DuckDB into your Python and R analytical workflows Effectively leverage DuckDB's convenient SQL enhancements Explore the wider ecosystem and pathways for building DuckDB-powered data applications Who this book is for If you’re interested in expanding your analytical toolkit, this book is for you. It will be particularly valuable for data analysts wanting to rapidly explore and query complex data, data and software engineers looking for a lean and versatile data processing tool, along with data scientists needing a scalable data manipulation library that integrates seamlessly with Python and R. You will get the most from this book if you have some familiarity with SQL and foundational database concepts, as well as exposure to a programming language such as Python or R.

Getting Started with DuckDB

Getting Started with DuckDB PDF Author: Simon Aubury
Publisher: Packt Publishing Ltd
ISBN: 1803232536
Category : Computers
Languages : en
Pages : 382

Get Book Here

Book Description
Analyze and transform data efficiently with DuckDB, a versatile, modern, in-process SQL database Key Features Use DuckDB to rapidly load, transform, and query data across a range of sources and formats Gain practical experience using SQL, Python, and R to effectively analyze data Learn how open source tools and cloud services in the broader data ecosystem complement DuckDB’s versatile capabilities Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionDuckDB is a fast in-process analytical database. Its ease of use, versatile feature set, and powerful analytical capabilities make DuckDB a valuable addition to the data practitioner’s toolkit. Getting Started with DuckDB offers a practical overview of DuckDB’s fundamentals and guidance for effectively using its powerful capabilities. Through extensive hands-on examples, you’ll learn how to use DuckDB to load, transform, and query a variety of data sources and formats, including CSV, JSON, and Parquet files, semi-structured data, remotely-hosted files, and external databases. You'll also find out how to leverage DuckDB's performance optimizations and friendly SQL enhancements. You'll explore how to use DuckDB’s extensions for specialized applications, such as geospatial analysis and text search over document collections. In addition to working through examples in SQL, Python, and R, you’ll also dive into using DuckDB for analyzing public datasets and discover the wider ecosystem of open-source tools and cloud services that supercharge DuckDB-powered workflows and applications. Whether you’re a seasoned data practitioner or new to working with analytical data, this book will rapidly get you up to speed with DuckDB’s versatile and powerful capabilities, enabling you to apply them in your analytical workflows and projects.What you will learn Understand the properties and applications of a columnar in-process database Use SQL to load, transform, and query a range of data formats Discover DuckDB's rich extensions and learn how to apply them Use nested data types to model semi-structured data and extract and model JSON data Integrate DuckDB into your Python and R analytical workflows Effectively leverage DuckDB's convenient SQL enhancements Explore the wider ecosystem and pathways for building DuckDB-powered data applications Who this book is for If you’re interested in expanding your analytical toolkit, this book is for you. It will be particularly valuable for data analysts wanting to rapidly explore and query complex data, data and software engineers looking for a lean and versatile data processing tool, along with data scientists needing a scalable data manipulation library that integrates seamlessly with Python and R. You will get the most from this book if you have some familiarity with SQL and foundational database concepts, as well as exposure to a programming language such as Python or R.

DuckDB in Action

DuckDB in Action PDF Author: Mark Needham
Publisher: Simon and Schuster
ISBN: 1638355592
Category : Computers
Languages : en
Pages : 310

Get Book Here

Book Description
Dive into DuckDB and start processing gigabytes of data with ease—all with no data warehouse. DuckDB is a cutting-edge SQL database that makes it incredibly easy to analyze big data sets right from your laptop. In DuckDB in Action you’ll learn everything you need to know to get the most out of this awesome tool, keep your data secure on prem, and save you hundreds on your cloud bill. From data ingestion to advanced data pipelines, you’ll learn everything you need to get the most out of DuckDB—all through hands-on examples. Open up DuckDB in Action and learn how to: • Read and process data from CSV, JSON and Parquet sources both locally and remote • Write analytical SQL queries, including aggregations, common table expressions, window functions, special types of joins, and pivot tables • Use DuckDB from Python, both with SQL and its "Relational"-API, interacting with databases but also data frames • Prepare, ingest and query large datasets • Build cloud data pipelines • Extend DuckDB with custom functionality Pragmatic and comprehensive, DuckDB in Action introduces the DuckDB database and shows you how to use it to solve common data workflow problems. You won’t need to read through pages of documentation—you’ll learn as you work. Get to grips with DuckDB's unique SQL dialect, learning to seamlessly load, prepare, and analyze data using SQL queries. Extend DuckDB with both Python and built-in tools such as MotherDuck, and gain practical insights into building robust and automated data pipelines. About the technology DuckDB makes data analytics fast and fun! You don’t need to set up a Spark or run a cloud data warehouse just to process a few hundred gigabytes of data. DuckDB is easily embeddable in any data analytics application, runs on a laptop, and processes data from almost any source, including JSON, CSV, Parquet, SQLite and Postgres. About the book DuckDB in Action guides you example-by-example from setup, through your first SQL query, to advanced topics like building data pipelines and embedding DuckDB as a local data store for a Streamlit web app. You’ll explore DuckDB’s handy SQL extensions, get to grips with aggregation, analysis, and data without persistence, and use Python to customize DuckDB. A hands-on project accompanies each new topic, so you can see DuckDB in action. What's inside • Prepare, ingest and query large datasets • Build cloud data pipelines • Extend DuckDB with custom functionality • Fast-paced SQL recap: From simple queries to advanced analytics About the reader For data pros comfortable with Python and CLI tools. About the author Mark Needham is a blogger and video creator at @?LearnDataWithMark. Michael Hunger leads product innovation for the Neo4j graph database. Michael Simons is a Java Champion, author, and Engineer at Neo4j.

In-Memory Analytics with Apache Arrow

In-Memory Analytics with Apache Arrow PDF Author: Matthew Topol
Publisher: Packt Publishing Ltd
ISBN: 183546968X
Category : Computers
Languages : en
Pages : 406

Get Book Here

Book Description
Harness the power of Apache Arrow to optimize tabular data processing and develop robust, high-performance data systems with its standardized, language-independent columnar memory format Key Features Explore Apache Arrow's data types and integration with pandas, Polars, and Parquet Work with Arrow libraries such as Flight SQL, Acero compute engine, and Dataset APIs for tabular data Enhance and accelerate machine learning data pipelines using Apache Arrow and its subprojects Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionApache Arrow is an open source, columnar in-memory data format designed for efficient data processing and analytics. This book harnesses the author’s 15 years of experience to show you a standardized way to work with tabular data across various programming languages and environments, enabling high-performance data processing and exchange. This updated second edition gives you an overview of the Arrow format, highlighting its versatility and benefits through real-world use cases. It guides you through enhancing data science workflows, optimizing performance with Apache Parquet and Spark, and ensuring seamless data translation. You’ll explore data interchange and storage formats, and Arrow's relationships with Parquet, Protocol Buffers, FlatBuffers, JSON, and CSV. You’ll also discover Apache Arrow subprojects, including Flight, SQL, Database Connectivity, and nanoarrow. You’ll learn to streamline machine learning workflows, use Arrow Dataset APIs, and integrate with popular analytical data systems such as Snowflake, Dremio, and DuckDB. The latter chapters provide real-world examples and case studies of products powered by Apache Arrow, providing practical insights into its applications. By the end of this book, you’ll have all the building blocks to create efficient and powerful analytical services and utilities with Apache Arrow.What you will learn Use Apache Arrow libraries to access data files, both locally and in the cloud Understand the zero-copy elements of the Apache Arrow format Improve the read performance of data pipelines by memory-mapping Arrow files Produce and consume Apache Arrow data efficiently by sharing memory with the C API Leverage the Arrow compute engine, Acero, to perform complex operations Create Arrow Flight servers and clients for transferring data quickly Build the Arrow libraries locally and contribute to the community Who this book is for This book is for developers, data engineers, and data scientists looking to explore the capabilities of Apache Arrow from the ground up. Whether you’re building utilities for data analytics and query engines, or building full pipelines with tabular data, this book can help you out regardless of your preferred programming language. A basic understanding of data analysis concepts is needed, but not necessary. Code examples are provided using C++, Python, and Go throughout the book.

Elastic Stack 8.x Cookbook

Elastic Stack 8.x Cookbook PDF Author: Huage Chen
Publisher: Packt Publishing Ltd
ISBN: 1837633509
Category : Computers
Languages : en
Pages : 688

Get Book Here

Book Description
Unlock the full potential of Elastic Stack for search, analytics, security, and observability and manage substantial data workloads in both on-premise and cloud environments Key Features Explore the diverse capabilities of the Elastic Stack through a comprehensive set of recipes Build search applications, analyze your data, and observe cloud-native applications Harness powerful machine learning and AI features to create data science and search applications Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionLearn how to make the most of the Elastic Stack (ELK Stack) products—including Elasticsearch, Kibana, Elastic Agent, and Logstash—to take data reliably and securely from any source, in any format, and then search, analyze, and visualize it in real-time. This cookbook takes a practical approach to unlocking the full potential of Elastic Stack through detailed recipes step by step. Starting with installing and ingesting data using Elastic Agent and Beats, this book guides you through data transformation and enrichment with various Elastic components and explores the latest advancements in search applications, including semantic search and Generative AI. You'll then visualize and explore your data and create dashboards using Kibana. As you progress, you'll advance your skills with machine learning for data science, get to grips with natural language processing, and discover the power of vector search. The book covers Elastic Observability use cases for log, infrastructure, and synthetics monitoring, along with essential strategies for securing the Elastic Stack. Finally, you'll gain expertise in Elastic Stack operations to effectively monitor and manage your system.What you will learn Discover techniques for collecting data from diverse sources Visualize data and create dashboards using Kibana to extract business insights Explore machine learning, vector search, and AI capabilities of Elastic Stack Handle data transformation and data formatting Build search solutions from the ingested data Leverage data science tools for in-depth data exploration Monitor and manage your system with Elastic Stack Who this book is for This book is for Elastic Stack users, developers, observability practitioners, and data professionals ranging from beginner to expert level. If you’re a developer, you’ll benefit from the easy-to-follow recipes for using APIs and features to build powerful applications, and if you’re an observability practitioner, this book will help you with use cases covering APM, Kubernetes, and cloud monitoring. For data engineers and AI enthusiasts, the book covers dedicated recipes on vector search and machine learning. No prior knowledge of the Elastic Stack is required.

Polars Cookbook

Polars Cookbook PDF Author: Yuki Kakegawa
Publisher: Packt Publishing Ltd
ISBN: 180512515X
Category : Computers
Languages : en
Pages : 394

Get Book Here

Book Description
Leverage a lightning fast DataFrame library for efficient data wrangling in Python Key Features Unlock the power of Python Polars for faster and more efficient data analysis workflows Master the fundamentals of Python Polars with step-by-step recipes Discover data manipulation techniques to apply across multiple data problems Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionPolars Cookbook is a complete guide that not only helps you get started with Python Polars but also gives you effective solutions to your day-to-day data problems. Dive into the world of Polars, a high-performance DataFrame library designed for efficient data processing and analysis. This cookbook takes a practical approach to unlocking the full potential of Polars through detailed, step-by-step recipes. Starting with installation and basic operations, this book guides you through data manipulation, advanced querying, and performance optimization techniques. You’ll learn how to handle large datasets, perform complex transformations, and leverage Polars’ powerful features for data science tasks. As you progress, you’ll explore Polars’ integration with other tools and libraries, and discover how to deploy Polars in both onpremises and cloud environments. You’ll also explore use cases for data engineering, time series analysis, statistical analysis, and machine learning, providing essential strategies for securing and optimizing your Polars workflows. By the end of this book, you’ll have acquired the knowledge and skills to build scalable, efficient, and reliable data processing solutions using Polars.What you will learn Read from different data sources and write to various files and databases Apply aggregations, window functions, and string manipulations Perform common data tasks such as handling missing values and performing list and array operations Discover how to reshape and tidy your data by pivoting, joining, and concatenating Analyze your time series data in Python Polars Create better workflows with testing and debugging Who this book is for This book is for data analysts, data scientists, and data engineers who want to learn how to use Polars in their workflows. Working knowledge of the Python programming language is required. Experience working with a DataFrame library such as pandas or PySpark will also be helpful.

Amazon DynamoDB - The Definitive Guide

Amazon DynamoDB - The Definitive Guide PDF Author: Aman Dhingra
Publisher: Packt Publishing Ltd
ISBN: 1803248327
Category : Computers
Languages : en
Pages : 415

Get Book Here

Book Description
Harness the potential and scalability of DynamoDB to effortlessly construct resilient, low-latency databases Key Features Discover how DynamoDB works behind the scenes to make the most of its features Learn how to keep latency and costs minimal even when scaling up Integrate DynamoDB with other AWS services to create a full data analytics system Purchase of the print or Kindle book includes a free PDF eBook Book DescriptionThis book will help you master Amazon DynamoDB, the fully managed, serverless, NoSQL database service designed for high performance at any scale. Authored by Aman Dhingra, senior DynamoDB specialist solutions architect at AWS, and Mike Mackay, former senior NoSQL specialist solutions architect at AWS, this guide draws on their expertise to equip you with the knowledge and skills needed to harness DynamoDB's full potential. This book not only introduces you to DynamoDB's core features and real-world applications, but also provides in-depth guidance on transitioning from traditional relational databases to the NoSQL world. You'll learn essential data modeling techniques, such as vertical partitioning, and explore the nuances of DynamoDB's indexing capabilities, capacity modes, and consistency models. The chapters also help you gain a solid understanding of advanced topics such as enhanced analytical patterns, implementing caching with DynamoDB Accelerator (DAX), and integrating DynamoDB with other AWS services to optimize your data strategies. By the end of this book, you’ll be able to design, build, and deliver low-latency, high-throughput DynamoDB solutions, driving new levels of efficiency and performance for your applications.What you will learn Master key-value data modeling in DynamoDB for efficiency Transition from RDBMSs to NoSQL with optimized strategies Implement read consistency and ACID transactions effectively Explore vertical partitioning for specific data access patterns Optimize data retrieval using secondary indexes in DynamoDB Manage capacity modes, backup strategies, and core components Enhance DynamoDB with caching, analytics, and global tables Evaluate and design your DynamoDB migration strategy Who this book is for This book is for software architects designing scalable systems, developers optimizing performance with DynamoDB, and engineering managers guiding decision-making. Data engineers will learn to integrate DynamoDB into workflows, while product owners will explore its innovative capabilities. DBAs transitioning to NoSQL will find valuable insights on DynamoDB and RDBMS integration. Basic knowledge of software engineering, Python, and cloud computing is helpful. Hands-on AWS or DynamoDB experience is beneficial but not required.

DevOps for Data Science

DevOps for Data Science PDF Author: Alex Gold
Publisher: CRC Press
ISBN: 104003442X
Category : Business & Economics
Languages : en
Pages : 274

Get Book Here

Book Description
Data Scientists are experts at analyzing, modelling and visualizing data but, at one point or another, have all encountered difficulties in collaborating with or delivering their work to the people and systems that matter. Born out of the agile software movement, DevOps is a set of practices, principles and tools that help software engineers reliably deploy work to production. This book takes the lessons of DevOps and aplies them to creating and delivering production-grade data science projects in Python and R. This book’s first section explores how to build data science projects that deploy to production with no frills or fuss. Its second section covers the rudiments of administering a server, including Linux, application, and network administration before concluding with a demystification of the concerns of enterprise IT/Administration in its final section, making it possible for data scientists to communicate and collaborate with their organization’s security, networking, and administration teams. Key Features: • Start-to-finish labs take readers through creating projects that meet DevOps best practices and creating a server-based environment to work on and deploy them. • Provides an appendix of cheatsheets so that readers will never be without the reference they need to remember a Git, Docker, or Command Line command. • Distills what a data scientist needs to know about Docker, APIs, CI/CD, Linux, DNS, SSL, HTTP, Auth, and more. • Written specifically to address the concern of a data scientist who wants to take their Python or R work to production. There are countless books on creating data science work that is correct. This book, on the otherhand, aims to go beyond this, targeted at data scientists who want their work to be than merely accurate and deliver work that matters.

Duckdb: Up and Running

Duckdb: Up and Running PDF Author: Wei-Meng Lee
Publisher:
ISBN: 9781098159696
Category : Computers
Languages : en
Pages : 0

Get Book Here

Book Description


Analytics Engineering with SQL and dbt

Analytics Engineering with SQL and dbt PDF Author: Rui Pedro Machado
Publisher: "O'Reilly Media, Inc."
ISBN: 1098142349
Category : Computers
Languages : en
Pages : 324

Get Book Here

Book Description
With the shift from data warehouses to data lakes, data now lands in repositories before it's been transformed, enabling engineers to model raw data into clean, well-defined datasets. dbt (data build tool) helps you take data further. This practical book shows data analysts, data engineers, BI developers, and data scientists how to create a true self-service transformation platform through the use of dynamic SQL. Authors Rui Machado from Monstarlab and Hélder Russa from Jumia show you how to quickly deliver new data products by focusing more on value delivery and less on architectural and engineering aspects. If you know your business well and have the technical skills to model raw data into clean, well-defined datasets, you'll learn how to design and deliver data models without any technical influence. With this book, you'll learn: What dbt is and how a dbt project is structured How dbt fits into the data engineering and analytics worlds How to collaborate on building data models The main tools and architectures for building useful, functional data models How to fit dbt into data warehousing and laking architecture How to build tests for data transformations

Analyzing Baseball Data with R

Analyzing Baseball Data with R PDF Author: Jim Albert
Publisher: CRC Press
ISBN: 104009712X
Category : Mathematics
Languages : en
Pages : 418

Get Book Here

Book Description
“Our community has continued to grow exponentially, thanks to those who inspire the next generation. And inspiring the next generation is what the authors of Analyzing Baseball Data with R are doing. They are setting the career path for still thousands more. We all need some sort of kickstart to take that first or second step. You may be a beginner R coder, but you need access to baseball data. How do you access this data, how do you manipulate it, how do you analyze it? This is what this book does for you. But it does more, by doing what sabermetrics does best: it asks baseball questions. Throughout the book, baseball questions are asked, some straightforward, and others more thought-provoking.” From the Foreword by Tom Tango Analyzing Baseball Data with R Third Edition introduces R to sabermetricians, baseball enthusiasts, and students interested in exploring the richness of baseball data. It equips you with the necessary skills and software tools to perform all the analysis steps, from importing the data to transforming them into an appropriate format to visualizing the data via graphs to performing a statistical analysis. The authors first present an overview of publicly available baseball datasets and a gentle introduction to the type of data structures and exploratory and data management capabilities of R. They also cover the ggplot2 graphics functions and employ a tidyverse-friendly workflow throughout. Much of the book illustrates the use of R through popular sabermetrics topics, including the Pythagorean formula, runs expectancy, catcher framing, career trajectories, simulation of games and seasons, patterns of streaky behavior of players, and launch angles and exit velocities. All the datasets and R code used in the text are available for download online. New to the third edition is the revised R code to make use of new functions made available through the tidyverse. The third edition introduces three chapters of new material, focusing on communicating results via presentations using the Quarto publishing system, web applications using the Shiny package, and working with large data files. An online version of this book is hosted at https://beanumber.github.io/abdwr3e/.