Author: April Reeve
Publisher: Newnes
ISBN: 0123977916
Category : Computers
Languages : en
Pages : 203
Book Description
Managing Data in Motion describes techniques that have been developed for significantly reducing the complexity of managing system interfaces and enabling scalable architectures. Author April Reeve brings over two decades of experience to present a vendor-neutral approach to moving data between computing environments and systems. Readers will learn the techniques, technologies, and best practices for managing the passage of data between computer systems and integrating disparate data together in an enterprise environment. The average enterprise's computing environment is comprised of hundreds to thousands computer systems that have been built, purchased, and acquired over time. The data from these various systems needs to be integrated for reporting and analysis, shared for business transaction processing, and converted from one format to another when old systems are replaced and new systems are acquired. The management of the "data in motion" in organizations is rapidly becoming one of the biggest concerns for business and IT management. Data warehousing and conversion, real-time data integration, and cloud and "big data" applications are just a few of the challenges facing organizations and businesses today. Managing Data in Motion tackles these and other topics in a style easily understood by business and IT managers as well as programmers and architects. - Presents a vendor-neutral overview of the different technologies and techniques for moving data between computer systems including the emerging solutions for unstructured as well as structured data types - Explains, in non-technical terms, the architecture and components required to perform data integration - Describes how to reduce the complexity of managing system interfaces and enable a scalable data architecture that can handle the dimensions of "Big Data"
Managing Data in Motion
Author: April Reeve
Publisher: Newnes
ISBN: 0123977916
Category : Computers
Languages : en
Pages : 203
Book Description
Managing Data in Motion describes techniques that have been developed for significantly reducing the complexity of managing system interfaces and enabling scalable architectures. Author April Reeve brings over two decades of experience to present a vendor-neutral approach to moving data between computing environments and systems. Readers will learn the techniques, technologies, and best practices for managing the passage of data between computer systems and integrating disparate data together in an enterprise environment. The average enterprise's computing environment is comprised of hundreds to thousands computer systems that have been built, purchased, and acquired over time. The data from these various systems needs to be integrated for reporting and analysis, shared for business transaction processing, and converted from one format to another when old systems are replaced and new systems are acquired. The management of the "data in motion" in organizations is rapidly becoming one of the biggest concerns for business and IT management. Data warehousing and conversion, real-time data integration, and cloud and "big data" applications are just a few of the challenges facing organizations and businesses today. Managing Data in Motion tackles these and other topics in a style easily understood by business and IT managers as well as programmers and architects. - Presents a vendor-neutral overview of the different technologies and techniques for moving data between computer systems including the emerging solutions for unstructured as well as structured data types - Explains, in non-technical terms, the architecture and components required to perform data integration - Describes how to reduce the complexity of managing system interfaces and enable a scalable data architecture that can handle the dimensions of "Big Data"
Publisher: Newnes
ISBN: 0123977916
Category : Computers
Languages : en
Pages : 203
Book Description
Managing Data in Motion describes techniques that have been developed for significantly reducing the complexity of managing system interfaces and enabling scalable architectures. Author April Reeve brings over two decades of experience to present a vendor-neutral approach to moving data between computing environments and systems. Readers will learn the techniques, technologies, and best practices for managing the passage of data between computer systems and integrating disparate data together in an enterprise environment. The average enterprise's computing environment is comprised of hundreds to thousands computer systems that have been built, purchased, and acquired over time. The data from these various systems needs to be integrated for reporting and analysis, shared for business transaction processing, and converted from one format to another when old systems are replaced and new systems are acquired. The management of the "data in motion" in organizations is rapidly becoming one of the biggest concerns for business and IT management. Data warehousing and conversion, real-time data integration, and cloud and "big data" applications are just a few of the challenges facing organizations and businesses today. Managing Data in Motion tackles these and other topics in a style easily understood by business and IT managers as well as programmers and architects. - Presents a vendor-neutral overview of the different technologies and techniques for moving data between computer systems including the emerging solutions for unstructured as well as structured data types - Explains, in non-technical terms, the architecture and components required to perform data integration - Describes how to reduce the complexity of managing system interfaces and enable a scalable data architecture that can handle the dimensions of "Big Data"
Executing Data Quality Projects
Author: Danette McGilvray
Publisher: Academic Press
ISBN: 0128180161
Category : Computers
Languages : en
Pages : 378
Book Description
Executing Data Quality Projects, Second Edition presents a structured yet flexible approach for creating, improving, sustaining and managing the quality of data and information within any organization. Studies show that data quality problems are costing businesses billions of dollars each year, with poor data linked to waste and inefficiency, damaged credibility among customers and suppliers, and an organizational inability to make sound decisions. Help is here! This book describes a proven Ten Step approach that combines a conceptual framework for understanding information quality with techniques, tools, and instructions for practically putting the approach to work – with the end result of high-quality trusted data and information, so critical to today's data-dependent organizations. The Ten Steps approach applies to all types of data and all types of organizations – for-profit in any industry, non-profit, government, education, healthcare, science, research, and medicine. This book includes numerous templates, detailed examples, and practical advice for executing every step. At the same time, readers are advised on how to select relevant steps and apply them in different ways to best address the many situations they will face. The layout allows for quick reference with an easy-to-use format highlighting key concepts and definitions, important checkpoints, communication activities, best practices, and warnings. The experience of actual clients and users of the Ten Steps provide real examples of outputs for the steps plus highlighted, sidebar case studies called Ten Steps in Action. This book uses projects as the vehicle for data quality work and the word broadly to include: 1) focused data quality improvement projects, such as improving data used in supply chain management, 2) data quality activities in other projects such as building new applications and migrating data from legacy systems, integrating data because of mergers and acquisitions, or untangling data due to organizational breakups, and 3) ad hoc use of data quality steps, techniques, or activities in the course of daily work. The Ten Steps approach can also be used to enrich an organization's standard SDLC (whether sequential or Agile) and it complements general improvement methodologies such as six sigma or lean. No two data quality projects are the same but the flexible nature of the Ten Steps means the methodology can be applied to all. The new Second Edition highlights topics such as artificial intelligence and machine learning, Internet of Things, security and privacy, analytics, legal and regulatory requirements, data science, big data, data lakes, and cloud computing, among others, to show their dependence on data and information and why data quality is more relevant and critical now than ever before. - Includes concrete instructions, numerous templates, and practical advice for executing every step of The Ten Steps approach - Contains real examples from around the world, gleaned from the author's consulting practice and from those who implemented based on her training courses and the earlier edition of the book - Allows for quick reference with an easy-to-use format highlighting key concepts and definitions, important checkpoints, communication activities, and best practices - A companion Web site includes links to numerous data quality resources, including many of the templates featured in the text, quick summaries of key ideas from the Ten Steps methodology, and other tools and information that are available online
Publisher: Academic Press
ISBN: 0128180161
Category : Computers
Languages : en
Pages : 378
Book Description
Executing Data Quality Projects, Second Edition presents a structured yet flexible approach for creating, improving, sustaining and managing the quality of data and information within any organization. Studies show that data quality problems are costing businesses billions of dollars each year, with poor data linked to waste and inefficiency, damaged credibility among customers and suppliers, and an organizational inability to make sound decisions. Help is here! This book describes a proven Ten Step approach that combines a conceptual framework for understanding information quality with techniques, tools, and instructions for practically putting the approach to work – with the end result of high-quality trusted data and information, so critical to today's data-dependent organizations. The Ten Steps approach applies to all types of data and all types of organizations – for-profit in any industry, non-profit, government, education, healthcare, science, research, and medicine. This book includes numerous templates, detailed examples, and practical advice for executing every step. At the same time, readers are advised on how to select relevant steps and apply them in different ways to best address the many situations they will face. The layout allows for quick reference with an easy-to-use format highlighting key concepts and definitions, important checkpoints, communication activities, best practices, and warnings. The experience of actual clients and users of the Ten Steps provide real examples of outputs for the steps plus highlighted, sidebar case studies called Ten Steps in Action. This book uses projects as the vehicle for data quality work and the word broadly to include: 1) focused data quality improvement projects, such as improving data used in supply chain management, 2) data quality activities in other projects such as building new applications and migrating data from legacy systems, integrating data because of mergers and acquisitions, or untangling data due to organizational breakups, and 3) ad hoc use of data quality steps, techniques, or activities in the course of daily work. The Ten Steps approach can also be used to enrich an organization's standard SDLC (whether sequential or Agile) and it complements general improvement methodologies such as six sigma or lean. No two data quality projects are the same but the flexible nature of the Ten Steps means the methodology can be applied to all. The new Second Edition highlights topics such as artificial intelligence and machine learning, Internet of Things, security and privacy, analytics, legal and regulatory requirements, data science, big data, data lakes, and cloud computing, among others, to show their dependence on data and information and why data quality is more relevant and critical now than ever before. - Includes concrete instructions, numerous templates, and practical advice for executing every step of The Ten Steps approach - Contains real examples from around the world, gleaned from the author's consulting practice and from those who implemented based on her training courses and the earlier edition of the book - Allows for quick reference with an easy-to-use format highlighting key concepts and definitions, important checkpoints, communication activities, and best practices - A companion Web site includes links to numerous data quality resources, including many of the templates featured in the text, quick summaries of key ideas from the Ten Steps methodology, and other tools and information that are available online
Data and Reality
Author: William Kent
Publisher: North Holland
ISBN:
Category : Computers
Languages : en
Pages : 238
Book Description
The nature of an information system; Naming; Relationships; Attributes; Types and categories and sets; Models; The record model; The other three popular models; The modelling of relationships; Elementary concepts; Philosophy.
Publisher: North Holland
ISBN:
Category : Computers
Languages : en
Pages : 238
Book Description
The nature of an information system; Naming; Relationships; Attributes; Types and categories and sets; Models; The record model; The other three popular models; The modelling of relationships; Elementary concepts; Philosophy.
Kafka: The Definitive Guide
Author: Gwen Shapira
Publisher: "O'Reilly Media, Inc."
ISBN: 1492043036
Category : Computers
Languages : en
Pages : 500
Book Description
Every enterprise application creates data, whether it consists of log messages, metrics, user activity, or outgoing messages. Moving all this data is just as important as the data itself. With this updated edition, application architects, developers, and production engineers new to the Kafka streaming platform will learn how to handle data in motion. Additional chapters cover Kafka's AdminClient API, transactions, new security features, and tooling changes. Engineers from Confluent and LinkedIn responsible for developing Kafka explain how to deploy production Kafka clusters, write reliable event-driven microservices, and build scalable stream processing applications with this platform. Through detailed examples, you'll learn Kafka's design principles, reliability guarantees, key APIs, and architecture details, including the replication protocol, the controller, and the storage layer. You'll examine: Best practices for deploying and configuring Kafka Kafka producers and consumers for writing and reading messages Patterns and use-case requirements to ensure reliable data delivery Best practices for building data pipelines and applications with Kafka How to perform monitoring, tuning, and maintenance tasks with Kafka in production The most critical metrics among Kafka's operational measurements Kafka's delivery capabilities for stream processing systems
Publisher: "O'Reilly Media, Inc."
ISBN: 1492043036
Category : Computers
Languages : en
Pages : 500
Book Description
Every enterprise application creates data, whether it consists of log messages, metrics, user activity, or outgoing messages. Moving all this data is just as important as the data itself. With this updated edition, application architects, developers, and production engineers new to the Kafka streaming platform will learn how to handle data in motion. Additional chapters cover Kafka's AdminClient API, transactions, new security features, and tooling changes. Engineers from Confluent and LinkedIn responsible for developing Kafka explain how to deploy production Kafka clusters, write reliable event-driven microservices, and build scalable stream processing applications with this platform. Through detailed examples, you'll learn Kafka's design principles, reliability guarantees, key APIs, and architecture details, including the replication protocol, the controller, and the storage layer. You'll examine: Best practices for deploying and configuring Kafka Kafka producers and consumers for writing and reading messages Patterns and use-case requirements to ensure reliable data delivery Best practices for building data pipelines and applications with Kafka How to perform monitoring, tuning, and maintenance tasks with Kafka in production The most critical metrics among Kafka's operational measurements Kafka's delivery capabilities for stream processing systems
Microsoft Cybersecurity Architect Exam Ref SC-100
Author: Dwayne Natwick
Publisher: Packt Publishing Ltd
ISBN: 1836208502
Category : Computers
Languages : en
Pages : 301
Book Description
Unlock your potential to pass the SC-100 exam by mastering advanced cloud security strategies, designing zero-trust architectures, and evaluating cybersecurity frameworks with this latest exam guide Purchase of this book unlocks access to web-based exam prep resources such as mock exams, flashcards, exam tips, the eBook PDF Key Features Gain a deep understanding of all topics covered in the latest SC-100 exam Advance your knowledge of architecting and evaluating cybersecurity services to tackle day-to-day challenges Get certified with ease through mock tests with exam-level difficulty Benefit from practical examples that will help you put your new knowledge to work Book DescriptionThis Second Edition of Microsoft Cybersecurity Architect Exam Ref SC-100 is a comprehensive guide that will help cybersecurity professionals design and evaluate the cybersecurity architecture of Microsoft cloud services. Packed with practice questions, mock exams, interactive flashcards, and invaluable exam tips, this comprehensive resource gives you everything you need to conquer the SC-100 exam with confidence. This book will take you through designing a strategy for a cybersecurity architecture and evaluating the governance, risk, and compliance (GRC) of the architecture of both cloud-only and hybrid infrastructures. You'll discover how to implement zero trust principles, enhance security operations, and elevate your organization's security posture. By the end of this book, you'll be fully equipped to plan, design, and assess cybersecurity frameworks for Microsoft cloud environments—and pass the SC-100 exam with flying colors. Ready to take your cybersecurity expertise to the next level? This guide is your key to success.What you will learn Design a zero-trust strategy and architecture Evaluate GRC technical and security operation strategies Apply encryption standards for data protection Utilize Microsoft Defender tools to assess and enhance security posture Translate business goals into actionable security requirements Assess and mitigate security risks using industry benchmarks and threat intelligence Optimize security operations using SIEM and SOAR technologies Securely manage secrets, keys, and certificates in cloud environments Who this book is for This book targets is for IT professionals pursuing the Microsoft Cybersecurity Architect Expert SC-100 certification. Familiarity with the principles of administering core features and services within Microsoft Azure, Microsoft 365 and on-premises related technologies (server, active directory, networks) are needed. Prior knowledge of integration of these technologies with each other will also be beneficial.
Publisher: Packt Publishing Ltd
ISBN: 1836208502
Category : Computers
Languages : en
Pages : 301
Book Description
Unlock your potential to pass the SC-100 exam by mastering advanced cloud security strategies, designing zero-trust architectures, and evaluating cybersecurity frameworks with this latest exam guide Purchase of this book unlocks access to web-based exam prep resources such as mock exams, flashcards, exam tips, the eBook PDF Key Features Gain a deep understanding of all topics covered in the latest SC-100 exam Advance your knowledge of architecting and evaluating cybersecurity services to tackle day-to-day challenges Get certified with ease through mock tests with exam-level difficulty Benefit from practical examples that will help you put your new knowledge to work Book DescriptionThis Second Edition of Microsoft Cybersecurity Architect Exam Ref SC-100 is a comprehensive guide that will help cybersecurity professionals design and evaluate the cybersecurity architecture of Microsoft cloud services. Packed with practice questions, mock exams, interactive flashcards, and invaluable exam tips, this comprehensive resource gives you everything you need to conquer the SC-100 exam with confidence. This book will take you through designing a strategy for a cybersecurity architecture and evaluating the governance, risk, and compliance (GRC) of the architecture of both cloud-only and hybrid infrastructures. You'll discover how to implement zero trust principles, enhance security operations, and elevate your organization's security posture. By the end of this book, you'll be fully equipped to plan, design, and assess cybersecurity frameworks for Microsoft cloud environments—and pass the SC-100 exam with flying colors. Ready to take your cybersecurity expertise to the next level? This guide is your key to success.What you will learn Design a zero-trust strategy and architecture Evaluate GRC technical and security operation strategies Apply encryption standards for data protection Utilize Microsoft Defender tools to assess and enhance security posture Translate business goals into actionable security requirements Assess and mitigate security risks using industry benchmarks and threat intelligence Optimize security operations using SIEM and SOAR technologies Securely manage secrets, keys, and certificates in cloud environments Who this book is for This book targets is for IT professionals pursuing the Microsoft Cybersecurity Architect Expert SC-100 certification. Familiarity with the principles of administering core features and services within Microsoft Azure, Microsoft 365 and on-premises related technologies (server, active directory, networks) are needed. Prior knowledge of integration of these technologies with each other will also be beneficial.
Big Data For Dummies
Author: Judith S. Hurwitz
Publisher: John Wiley & Sons
ISBN: 1118644174
Category : Computers
Languages : en
Pages : 336
Book Description
Find the right big data solution for your business or organization Big data management is one of the major challenges facing business, industry, and not-for-profit organizations. Data sets such as customer transactions for a mega-retailer, weather patterns monitored by meteorologists, or social network activity can quickly outpace the capacity of traditional data management tools. If you need to develop or manage big data solutions, you'll appreciate how these four experts define, explain, and guide you through this new and often confusing concept. You'll learn what it is, why it matters, and how to choose and implement solutions that work. Effectively managing big data is an issue of growing importance to businesses, not-for-profit organizations, government, and IT professionals Authors are experts in information management, big data, and a variety of solutions Explains big data in detail and discusses how to select and implement a solution, security concerns to consider, data storage and presentation issues, analytics, and much more Provides essential information in a no-nonsense, easy-to-understand style that is empowering Big Data For Dummies cuts through the confusion and helps you take charge of big data solutions for your organization.
Publisher: John Wiley & Sons
ISBN: 1118644174
Category : Computers
Languages : en
Pages : 336
Book Description
Find the right big data solution for your business or organization Big data management is one of the major challenges facing business, industry, and not-for-profit organizations. Data sets such as customer transactions for a mega-retailer, weather patterns monitored by meteorologists, or social network activity can quickly outpace the capacity of traditional data management tools. If you need to develop or manage big data solutions, you'll appreciate how these four experts define, explain, and guide you through this new and often confusing concept. You'll learn what it is, why it matters, and how to choose and implement solutions that work. Effectively managing big data is an issue of growing importance to businesses, not-for-profit organizations, government, and IT professionals Authors are experts in information management, big data, and a variety of solutions Explains big data in detail and discusses how to select and implement a solution, security concerns to consider, data storage and presentation issues, analytics, and much more Provides essential information in a no-nonsense, easy-to-understand style that is empowering Big Data For Dummies cuts through the confusion and helps you take charge of big data solutions for your organization.
Data Virtualization for Business Intelligence Systems
Author: Rick van der Lans
Publisher: Elsevier
ISBN: 0123944252
Category : Business & Economics
Languages : en
Pages : 297
Book Description
Annotation In this book, Rick van der Lans explains how data virtualization servers work, what techniques to use to optimize access to various data sources and how these products can be applied in different projects.
Publisher: Elsevier
ISBN: 0123944252
Category : Business & Economics
Languages : en
Pages : 297
Book Description
Annotation In this book, Rick van der Lans explains how data virtualization servers work, what techniques to use to optimize access to various data sources and how these products can be applied in different projects.
Hands-On Big Data Modeling
Author: James Lee
Publisher: Packt Publishing Ltd
ISBN: 1788626087
Category : Computers
Languages : en
Pages : 293
Book Description
Solve all big data problems by learning how to create efficient data models Key FeaturesCreate effective models that get the most out of big dataApply your knowledge to datasets from Twitter and weather data to learn big dataTackle different data modeling challenges with expert techniques presented in this bookBook Description Modeling and managing data is a central focus of all big data projects. In fact, a database is considered to be effective only if you have a logical and sophisticated data model. This book will help you develop practical skills in modeling your own big data projects and improve the performance of analytical queries for your specific business requirements. To start with, you’ll get a quick introduction to big data and understand the different data modeling and data management platforms for big data. Then you’ll work with structured and semi-structured data with the help of real-life examples. Once you’ve got to grips with the basics, you’ll use the SQL Developer Data Modeler to create your own data models containing different file types such as CSV, XML, and JSON. You’ll also learn to create graph data models and explore data modeling with streaming data using real-world datasets. By the end of this book, you’ll be able to design and develop efficient data models for varying data sizes easily and efficiently. What you will learnGet insights into big data and discover various data modelsExplore conceptual, logical, and big data modelsUnderstand how to model data containing different file typesRun through data modeling with examples of Twitter, Bitcoin, IMDB and weather data modelingCreate data models such as Graph Data and Vector SpaceModel structured and unstructured data using Python and RWho this book is for This book is great for programmers, geologists, biologists, and every professional who deals with spatial data. If you want to learn how to handle GIS, GPS, and remote sensing data, then this book is for you. Basic knowledge of R and QGIS would be helpful.
Publisher: Packt Publishing Ltd
ISBN: 1788626087
Category : Computers
Languages : en
Pages : 293
Book Description
Solve all big data problems by learning how to create efficient data models Key FeaturesCreate effective models that get the most out of big dataApply your knowledge to datasets from Twitter and weather data to learn big dataTackle different data modeling challenges with expert techniques presented in this bookBook Description Modeling and managing data is a central focus of all big data projects. In fact, a database is considered to be effective only if you have a logical and sophisticated data model. This book will help you develop practical skills in modeling your own big data projects and improve the performance of analytical queries for your specific business requirements. To start with, you’ll get a quick introduction to big data and understand the different data modeling and data management platforms for big data. Then you’ll work with structured and semi-structured data with the help of real-life examples. Once you’ve got to grips with the basics, you’ll use the SQL Developer Data Modeler to create your own data models containing different file types such as CSV, XML, and JSON. You’ll also learn to create graph data models and explore data modeling with streaming data using real-world datasets. By the end of this book, you’ll be able to design and develop efficient data models for varying data sizes easily and efficiently. What you will learnGet insights into big data and discover various data modelsExplore conceptual, logical, and big data modelsUnderstand how to model data containing different file typesRun through data modeling with examples of Twitter, Bitcoin, IMDB and weather data modelingCreate data models such as Graph Data and Vector SpaceModel structured and unstructured data using Python and RWho this book is for This book is great for programmers, geologists, biologists, and every professional who deals with spatial data. If you want to learn how to handle GIS, GPS, and remote sensing data, then this book is for you. Basic knowledge of R and QGIS would be helpful.
Life and Motion of Socio-Economic Units
Author: Andrew Frank
Publisher: CRC Press
ISBN: 0203305701
Category : Technology & Engineering
Languages : en
Pages : 361
Book Description
One of the ongoing problems researchers in geography and GIS have is studying data that is inherently spatial over a long period of time. One of the main hurdles they have to overcome is the study of groups of people classified by their socio-economic status (one of the main means for governments, companies and research organisations to group together segments of the population). The amount of data collected by governments, business and research organisations has increased markedly in recent years. Geographic Information Systems have been more widely used than ever before for the storage and analysis of this information. Most GIS can handle this information spatially rather than temporally, and have difficulty with the management of socio-economic time series, which relate to spatial units. Accordingly, this book covers the issues ranging from the formal model to differentiate aspects of spatio-temporal data, through philosophical and fundamental reconsideration of time and space to the development of practical solutions to the problem. This book draws together an interdisciplinary group of scientists in the field of geography, computing, surveying and philosophy. It presents the definitive sourcebook on temporal GIS as applied to socio-economic units.
Publisher: CRC Press
ISBN: 0203305701
Category : Technology & Engineering
Languages : en
Pages : 361
Book Description
One of the ongoing problems researchers in geography and GIS have is studying data that is inherently spatial over a long period of time. One of the main hurdles they have to overcome is the study of groups of people classified by their socio-economic status (one of the main means for governments, companies and research organisations to group together segments of the population). The amount of data collected by governments, business and research organisations has increased markedly in recent years. Geographic Information Systems have been more widely used than ever before for the storage and analysis of this information. Most GIS can handle this information spatially rather than temporally, and have difficulty with the management of socio-economic time series, which relate to spatial units. Accordingly, this book covers the issues ranging from the formal model to differentiate aspects of spatio-temporal data, through philosophical and fundamental reconsideration of time and space to the development of practical solutions to the problem. This book draws together an interdisciplinary group of scientists in the field of geography, computing, surveying and philosophy. It presents the definitive sourcebook on temporal GIS as applied to socio-economic units.
Technology Made Simple for the Technical Recruiter, Second Edition
Author: Obi Ogbanufe
Publisher: iUniverse
ISBN: 1532064985
Category : Computers
Languages : en
Pages : 225
Book Description
If you’re a technical recruiter who wants to keep your skills up to date in the competitive field of technical resource placement, you need a detailed guidebook to outpace competitors. This technical skills primer focuses on technology fundamentals—from basic programming terms to big data vocabulary, network lingo, operating system jargon, and other crucial skill sets. Topics covered include: •sample questions to ask candidates, •types of networks and operating systems, •software development strategies, •cloud systems administration and DevOps, •data science and database job roles, and •information security job roles. Armed with indispensable information, the alphabet soup of technology acronyms will no longer be intimidating, and you will be able to analyze client and candidate requirements with confidence. Written in clear and concise prose, Technology Made Simple for the Technical Recruiter is an invaluable resource for any technical recruiter.
Publisher: iUniverse
ISBN: 1532064985
Category : Computers
Languages : en
Pages : 225
Book Description
If you’re a technical recruiter who wants to keep your skills up to date in the competitive field of technical resource placement, you need a detailed guidebook to outpace competitors. This technical skills primer focuses on technology fundamentals—from basic programming terms to big data vocabulary, network lingo, operating system jargon, and other crucial skill sets. Topics covered include: •sample questions to ask candidates, •types of networks and operating systems, •software development strategies, •cloud systems administration and DevOps, •data science and database job roles, and •information security job roles. Armed with indispensable information, the alphabet soup of technology acronyms will no longer be intimidating, and you will be able to analyze client and candidate requirements with confidence. Written in clear and concise prose, Technology Made Simple for the Technical Recruiter is an invaluable resource for any technical recruiter.