In this talk we'll talk in details on how modern databases are capable of performing backups without downtimes and how these backups later can be used to restore database to any point in time.
While the talk describes generally applicable approach, a Litestream (SQLite backup service) is used for reference implementation.
A Step-By-Step Disaster Recovery Blueprint & Best Practices for Your NetBacku...Symantec
In this technical session we will share a few customer tested blueprints for implementing DR strategies with NetBackup appliances showing support for onsite and offsite disaster recovery. This includes the architecture design with Symantec best practices, down to execution of the wizards and command lines needed to implement the solution.
Watch the recording of this Google+ Hangout: http://bit.ly/13oTjvp
Apache Kafka is a distributed streaming platform that allows building event-driven architectures. It provides high throughput and low latency for processing streaming data. Key features include event logging, publish-subscribe messaging, and stream processing capabilities. Some advantages are eventual consistency, scalability, fault tolerance and being more agile to maintain compared to traditional databases. It requires Zookeeper and the Java client API has undergone changes. Performance can be very high, with examples of LinkedIn processing 1.1 trillion messages per day and 2 million writes per second on modest hardware.
Slide 1 - Parallels Plesk Control Panel 8.6.0webhostingguy
The document discusses various maintenance items and PTFs for IBM DB2 including:
- PTFs for DB2 Version 8 and z/OS to fix various issues like performance problems, errors, and serviceability enhancements
- New features in recent DB2 releases including support for longer SQL statements in ODBC, improved monitoring of real storage usage, and preliminary support for IBM's Enterprise Workload Manager
- Details on fixes for specific problems like encrypting passwords for distributed data, diagnosing hung threads, and monitoring when dynamic SQL exceeds resource limits.
FOSDEM MySQL & Friends Devroom, February 2018 MySQL Point-in-Time Recovery l...Frederic Descamps
The document describes how to perform point-in-time recovery (PITR) with MySQL to restore data to a past state. It requires binary logs to be enabled and backups kept along with binlogs. The procedure involves restoring the last backup, finding the binlog position, and replaying binlog events from that position up to the desired point in time. An example demonstrates restoring data after an accidental update, identifying the binlog position to recover to using SHOW BINLOG EVENTS.
DockerCon 2017 - Cilium - Network and Application Security with BPF and XDPThomas Graf
This talk will start with a deep dive and hands on examples of BPF, possibly the most promising low level technology to address challenges in application and network security, tracing, and visibility. We will discuss how BPF evolved from a simple bytecode language to filter raw sockets for tcpdump to the a JITable virtual machine capable of universally extending and instrumenting both the Linux kernel and user space applications. The introduction is followed by a concrete example of how the Cilium open source project applies BPF to solve networking, security, and load balancing for highly distributed applications. We will discuss and demonstrate how Cilium with the help of BPF can be combined with distributed system orchestration such as Docker to simplify security, operations, and troubleshooting of distributed applications.
Cilium - Network and Application Security with BPF and XDP Thomas Graf, Cova...Docker, Inc.
This talk will start with a deep dive and hands-on examples of BPF, possibly the most promising low-level technology to address challenges in application and network security, tracing, and visibility. We will discuss how BPF evolved from a simple bytecode language to filter raw sockets for tcpdump to the a JITable virtual machine capable of universally extending and instrumenting both the Linux kernel and user space applications. The introduction is followed by a concrete example of how the Cilium open source project applies BPF to solve networking, security and load balancing for highly distributed applications. We will discuss and demonstrate how Cilium with the help of BPF can be combined with distributed system orchestration such as Docker to simplify security, operations, and troubleshooting of distributed applications.
Buckle Up! With Valerie Burchby and Xinran Waibe | Current 2022HostedbyConfluent
Buckle Up! With Valerie Burchby and Xinran Waibe | Current 2022
Are you considering converting your daily batch ETLs into a new and exhilarating realtime framework? We’ll help you look before you leap as we take a deep dive into the unique operational challenges entailed in transitioning data processing paradigms.
As batched data pipelines consume data from well defined time intervals and write results to partitioned data storage, batched jobs are often idempotent, so the failure recovery is simply rerunning the faulty job instances. Batched data processes are triggered at a certain frequency (e.g. daily or hourly), so the data latency is determined by both the job scheduler and job run time. Therefore, many advanced data use cases, such as frequency capping, requires event streaming to enable real-time data insights. Event streaming applications process unbounded input data in real-time and append output to message queues and/or tables to be further processed. However, real-time data insights are no free meal - because event streaming comes with many unique engineering challenges, such as handling late-arriving and duplicate events, implementing event-time partitioning, and backfilling historical data after failures. In addition, batched-driven and even streaming are not incompatible to each other but can often be better together, as the Delta and Kappa Architecture are commonly adopted in modern data systems.
This document discusses database backup and recovery strategies. It outlines different backup types including logical, physical, hot, and cold backups. It describes how backups can protect a database from failures, increase uptime, and minimize data loss. The document also categorizes different types of failures and whether recovery is needed. It provides details on enabling archive logging mode and performing physical database backups in both open and closed states. Logical backups using Oracle Export and Import utilities are also covered.
A Step-By-Step Disaster Recovery Blueprint & Best Practices for Your NetBacku...Symantec
In this technical session we will share a few customer tested blueprints for implementing DR strategies with NetBackup appliances showing support for onsite and offsite disaster recovery. This includes the architecture design with Symantec best practices, down to execution of the wizards and command lines needed to implement the solution.
Watch the recording of this Google+ Hangout: http://bit.ly/13oTjvp
Apache Kafka is a distributed streaming platform that allows building event-driven architectures. It provides high throughput and low latency for processing streaming data. Key features include event logging, publish-subscribe messaging, and stream processing capabilities. Some advantages are eventual consistency, scalability, fault tolerance and being more agile to maintain compared to traditional databases. It requires Zookeeper and the Java client API has undergone changes. Performance can be very high, with examples of LinkedIn processing 1.1 trillion messages per day and 2 million writes per second on modest hardware.
Slide 1 - Parallels Plesk Control Panel 8.6.0webhostingguy
The document discusses various maintenance items and PTFs for IBM DB2 including:
- PTFs for DB2 Version 8 and z/OS to fix various issues like performance problems, errors, and serviceability enhancements
- New features in recent DB2 releases including support for longer SQL statements in ODBC, improved monitoring of real storage usage, and preliminary support for IBM's Enterprise Workload Manager
- Details on fixes for specific problems like encrypting passwords for distributed data, diagnosing hung threads, and monitoring when dynamic SQL exceeds resource limits.
FOSDEM MySQL & Friends Devroom, February 2018 MySQL Point-in-Time Recovery l...Frederic Descamps
The document describes how to perform point-in-time recovery (PITR) with MySQL to restore data to a past state. It requires binary logs to be enabled and backups kept along with binlogs. The procedure involves restoring the last backup, finding the binlog position, and replaying binlog events from that position up to the desired point in time. An example demonstrates restoring data after an accidental update, identifying the binlog position to recover to using SHOW BINLOG EVENTS.
DockerCon 2017 - Cilium - Network and Application Security with BPF and XDPThomas Graf
This talk will start with a deep dive and hands on examples of BPF, possibly the most promising low level technology to address challenges in application and network security, tracing, and visibility. We will discuss how BPF evolved from a simple bytecode language to filter raw sockets for tcpdump to the a JITable virtual machine capable of universally extending and instrumenting both the Linux kernel and user space applications. The introduction is followed by a concrete example of how the Cilium open source project applies BPF to solve networking, security, and load balancing for highly distributed applications. We will discuss and demonstrate how Cilium with the help of BPF can be combined with distributed system orchestration such as Docker to simplify security, operations, and troubleshooting of distributed applications.
Cilium - Network and Application Security with BPF and XDP Thomas Graf, Cova...Docker, Inc.
This talk will start with a deep dive and hands-on examples of BPF, possibly the most promising low-level technology to address challenges in application and network security, tracing, and visibility. We will discuss how BPF evolved from a simple bytecode language to filter raw sockets for tcpdump to the a JITable virtual machine capable of universally extending and instrumenting both the Linux kernel and user space applications. The introduction is followed by a concrete example of how the Cilium open source project applies BPF to solve networking, security and load balancing for highly distributed applications. We will discuss and demonstrate how Cilium with the help of BPF can be combined with distributed system orchestration such as Docker to simplify security, operations, and troubleshooting of distributed applications.
Buckle Up! With Valerie Burchby and Xinran Waibe | Current 2022HostedbyConfluent
Buckle Up! With Valerie Burchby and Xinran Waibe | Current 2022
Are you considering converting your daily batch ETLs into a new and exhilarating realtime framework? We’ll help you look before you leap as we take a deep dive into the unique operational challenges entailed in transitioning data processing paradigms.
As batched data pipelines consume data from well defined time intervals and write results to partitioned data storage, batched jobs are often idempotent, so the failure recovery is simply rerunning the faulty job instances. Batched data processes are triggered at a certain frequency (e.g. daily or hourly), so the data latency is determined by both the job scheduler and job run time. Therefore, many advanced data use cases, such as frequency capping, requires event streaming to enable real-time data insights. Event streaming applications process unbounded input data in real-time and append output to message queues and/or tables to be further processed. However, real-time data insights are no free meal - because event streaming comes with many unique engineering challenges, such as handling late-arriving and duplicate events, implementing event-time partitioning, and backfilling historical data after failures. In addition, batched-driven and even streaming are not incompatible to each other but can often be better together, as the Delta and Kappa Architecture are commonly adopted in modern data systems.
This document discusses database backup and recovery strategies. It outlines different backup types including logical, physical, hot, and cold backups. It describes how backups can protect a database from failures, increase uptime, and minimize data loss. The document also categorizes different types of failures and whether recovery is needed. It provides details on enabling archive logging mode and performing physical database backups in both open and closed states. Logical backups using Oracle Export and Import utilities are also covered.
1. The document discusses considerations for building a streaming service using Apache Flink, including an overview of Flink's dataflow model, streaming concepts, APIs, operations and monitoring.
2. It provides details on Flink's streaming APIs like ParDo, GroupByKey, windows, process functions and connectors. Monitoring with the Flink dashboard and REST APIs is also covered.
3. Methods for detecting abnormal statuses through metrics and rules are outlined, along with channels for alerts like email, SMS and Slack. The importance of only alerting on meaningful issues is discussed.
CTO Karl Anderson discusses the state of Kazoo. This includes integrations with FreeSWITCH, erlang, and Kamailio. Reseller milestones include the release of whitelabeling, webhooks, migration, carriers, debugging, account management and more.
3 Ways to Improve Performance from a Storage PerspectivePerforce
In this session, get three takeaways about Perforce performance benchmarks and their results across varying storage protocols, using NetApp storage as an example. Learn how to use Perforce benchmarks and tools to validate the performance of your Perforce deployment; understand Perforce performance across different storage protocols; and get tips and tricks for deploying Perforce on varying storage technologies.
Intel® Xeon® Processor E5-2600 v4 Big Data Analytics Applications ShowcaseIntel IT Center
This document showcases the Intel Xeon processor E5-2600 v4 product family and its performance benefits for big data analytics workloads. It provides examples of several software vendors who saw performance improvements of up to 34% on their applications when testing the new Intel Xeon processors. The document also outlines new processor technologies in areas like performance, orchestration and security. Configuration details are provided for the software vendor testing.
This document discusses using data virtualization to accelerate application projects by 50%. It outlines some common problems with physical data copies, such as bottlenecks, bugs due to old data, difficulty creating subsets, and delays. The document then introduces the concept of using a data virtualization appliance to take snapshots of production data and create thin clones for development and testing environments. This allows for fast, full-sized, self-service clones that can be refreshed quickly. Use cases discussed include improved development and testing workflows, faster production support like recovery and migration, and enabling continuous business intelligence functions.
The document describes Oracle's new parallel upgrade process for multitenant databases using catctl.pl. It allows upgrading the containers within a multitenant container database (CDB) simultaneously, reducing downtime. Key steps include running preupgrd.sql to generate fixup scripts, starting the CDB in upgrade mode, and using catctl.pl to run the upgrade scripts across containers in parallel. This promises to significantly improve upon previous sequential upgrade methods.
This document is the student guide for Oracle9i DBA Fundamentals II. It covers networking concepts, Oracle Net configuration, backup and recovery strategies, and recovery techniques using both user-managed and RMAN-based approaches. The guide contains chapters on topics such as basic Oracle Net architecture, configuring the database archiving mode, and RMAN backups. It is intended to teach students advanced database administration skills.
Chicago Docker Meetup Presentation - MediaflyMediafly
This document discusses how Bryan Murphy uses Docker at his company Mediafly. It begins by introducing Bryan and his background. It then describes what Mediafly does, including content management systems, secure content delivery, document and video processing, and customizable user interfaces. The document highlights aspects of Mediafly that make it interesting, such as being multi-device, multi-tenant, service oriented, and distributed. It provides examples of technologies used at Mediafly and some key metrics. The document then discusses why Docker is used at Mediafly, covering benefits like being developer friendly, enabling faster iteration and testing, managing dependencies, sharing environments, standardization, isolation, and infrastructure freedom.
- Fully managed SQL database service hosted on Microsoft Azure that provides predictable performance and pricing with 99.99% availability.
- Offers elastic database pools and data protection services like geo-replication and point-in-time restore.
- Compatible with SQL Server 2014 databases and comes in multiple service tiers based on resources like CPU cores and IOPS.
The document discusses the planned key rollover of the DNSSEC Key Signing Key (KSK) for the root zone from the current KSK-2010 to a new KSK-2017. It provides details on the milestones, approach, and state of the rollover process according to the Automated Updates of DNSSEC Trust Anchors protocol. The rollover was paused in 2017 due to uncertainty in measurement data, but progress has since been made to complete the rollover in 2018.
This document provides a list of media needed to install or upgrade SAP ERP software with Enhancement Package 6. It includes media for Java components, languages, ERP components, enhancement packages, databases like DB2, MaxDB, SQL Server, and operating system kernels for Linux, Windows, AIX, HP-UX, and Solaris. Each media item listed provides the material number, label, and name to use to download it from the SAP software distribution center.
Industry leaders Cisco, NetApp, VMware and Symantec have teamed up to develop a best practice framework and performance benchmark based on the VMware vSphere® Storage APIs - Data Protection (VADP). The test configuration uses the popular NetApp FlexPod environment, and the result proves that you can easily protect over 4 TB of virtual machine data per hour. And improved backup performance creates more reliable backups, shorter backup windows and less impact on the vSphere infrastructure.
In this session, we will show how these performance numbers can be easily obtained with minimal hardware and a small budget. In addition to backup performance, we will also discuss restore performance considerations.
Key topics include:
• How to select the correct hardware for the best ROI
• Strategies for minimizing backup impact and maximizing backup throughput
• Performance characteristics of VADP
• SAN or NBD (network) transports: which is recommended?
• Configurations for the fastest possible restores
Behind story in publisher that developer doesn't knowDavid Kim
The document discusses various technical aspects of managing a game development project and live game operations. It touches on topics like patching processes, crash reporting, in-game events, and server maintenance. It provides recommendations to simplify processes like patching by only downloading differential files, have servers automatically restart when resource usage exceeds a threshold, and handle in-game events and item giveaways through triggered processes rather than taking servers offline. The overall aim appears to be improving stability, player experience, and developer work-life balance.
Following our EBS R12.1.3 upgrade, we experienced inconsistent runtime and resource utilization with the Accrual Reconciliation Load program. During one month end close, a plant accountant could run accrual reconciliation for three years of data in under 10 hours. The next month, the same amount of data would take upwards of 30 hours to load. This behavior made it hard to plan for month end closure; successful accrual load being crucial to the finals for the month. Through multiple rounds of testing, we were able to complete the reconciliation process with nearly 15 years of data in just 1 hour and 45 minutes.
Implementing SharePoint on Azure, Lessons Learnt from a Real World ProjectK.Mohamed Faizal
This document discusses lessons learned from implementing SharePoint on Azure. It covers Azure architecture concepts like virtual networks, cloud services, availability sets, and load balancing. It provides an example reference architecture for a hybrid on-premises and Azure environment. It also discusses topics like database planning, disk performance, server topology with multiple tiers, and reserving IP addresses. The presentation aims to share best practices for deploying SharePoint on Azure based on a real-world project.
[NetApp] Managing Big Workspaces with Storage MagicPerforce
If you work with large volumes of data—multimedia assets, video game art, or firmware designs—you understand the pain of trying to quickly get a copy of source and build assets. But if you have the right storage system, you can be up and running with a new Perforce workspace in minutes instead of hours. See a simple procedure for fast workspace cloning using a few Perforce commands and NetApp FlexClone.
DataEngConf SF16 - Collecting and Moving Data at Scale Hakka Labs
This document summarizes Sada Furuhashi's presentation on Fluentd, an open source data collector. Fluentd provides a centralized way to collect, filter, and output log data from various sources like applications, servers, and databases. It addresses challenges with typical log collection architectures that have high latency, complex parsing, and a combination explosion of connections. Fluentd uses a plugin-based architecture with input, filter, and output components to flexibly collect, transform, and deliver log data at scale to targets like files, databases and visualization tools. Many large companies like Microsoft, Atlassian and Amazon use Fluentd for log collection and analytics in production environments.
The document discusses monitoring input/output (IO) performance in Oracle Exadata systems. It covers write-back flash cache (WBFC), various methods for monitoring IO using Automatic Workload Repository (AWR) data and cell-level scripts, correlating IO to workload, and scaling monitoring using metric extensions and Business Intelligence Publisher (BIP). The presentation provides examples of visualizing IO performance trends over time using AWR and cell data and measuring the impact of initialization parameters on latency. It also addresses reference bands for disk IO capacity and visualizing workload activity storage area activity by day per node.
Postgres indexes: how to make them work for your applicationBartosz Sypytkowski
Indexes are one of the most crucial structures of any relational database. In this talk we'll explain how to use them efficiently, how to read query plans and what do they mean for us. We'll also cover a variety of different indexing structures available in PostgreSQL database and build up some intuition about which one to pick depending on the situation.
This presentation covers HyParView and Plumtree - protocols used to build highly scalable clusters of data capable of gossiping messages between thousands of clients.
Más contenido relacionado
Similar a How do databases perform live backups and point-in-time recovery
1. The document discusses considerations for building a streaming service using Apache Flink, including an overview of Flink's dataflow model, streaming concepts, APIs, operations and monitoring.
2. It provides details on Flink's streaming APIs like ParDo, GroupByKey, windows, process functions and connectors. Monitoring with the Flink dashboard and REST APIs is also covered.
3. Methods for detecting abnormal statuses through metrics and rules are outlined, along with channels for alerts like email, SMS and Slack. The importance of only alerting on meaningful issues is discussed.
CTO Karl Anderson discusses the state of Kazoo. This includes integrations with FreeSWITCH, erlang, and Kamailio. Reseller milestones include the release of whitelabeling, webhooks, migration, carriers, debugging, account management and more.
3 Ways to Improve Performance from a Storage PerspectivePerforce
In this session, get three takeaways about Perforce performance benchmarks and their results across varying storage protocols, using NetApp storage as an example. Learn how to use Perforce benchmarks and tools to validate the performance of your Perforce deployment; understand Perforce performance across different storage protocols; and get tips and tricks for deploying Perforce on varying storage technologies.
Intel® Xeon® Processor E5-2600 v4 Big Data Analytics Applications ShowcaseIntel IT Center
This document showcases the Intel Xeon processor E5-2600 v4 product family and its performance benefits for big data analytics workloads. It provides examples of several software vendors who saw performance improvements of up to 34% on their applications when testing the new Intel Xeon processors. The document also outlines new processor technologies in areas like performance, orchestration and security. Configuration details are provided for the software vendor testing.
This document discusses using data virtualization to accelerate application projects by 50%. It outlines some common problems with physical data copies, such as bottlenecks, bugs due to old data, difficulty creating subsets, and delays. The document then introduces the concept of using a data virtualization appliance to take snapshots of production data and create thin clones for development and testing environments. This allows for fast, full-sized, self-service clones that can be refreshed quickly. Use cases discussed include improved development and testing workflows, faster production support like recovery and migration, and enabling continuous business intelligence functions.
The document describes Oracle's new parallel upgrade process for multitenant databases using catctl.pl. It allows upgrading the containers within a multitenant container database (CDB) simultaneously, reducing downtime. Key steps include running preupgrd.sql to generate fixup scripts, starting the CDB in upgrade mode, and using catctl.pl to run the upgrade scripts across containers in parallel. This promises to significantly improve upon previous sequential upgrade methods.
This document is the student guide for Oracle9i DBA Fundamentals II. It covers networking concepts, Oracle Net configuration, backup and recovery strategies, and recovery techniques using both user-managed and RMAN-based approaches. The guide contains chapters on topics such as basic Oracle Net architecture, configuring the database archiving mode, and RMAN backups. It is intended to teach students advanced database administration skills.
Chicago Docker Meetup Presentation - MediaflyMediafly
This document discusses how Bryan Murphy uses Docker at his company Mediafly. It begins by introducing Bryan and his background. It then describes what Mediafly does, including content management systems, secure content delivery, document and video processing, and customizable user interfaces. The document highlights aspects of Mediafly that make it interesting, such as being multi-device, multi-tenant, service oriented, and distributed. It provides examples of technologies used at Mediafly and some key metrics. The document then discusses why Docker is used at Mediafly, covering benefits like being developer friendly, enabling faster iteration and testing, managing dependencies, sharing environments, standardization, isolation, and infrastructure freedom.
- Fully managed SQL database service hosted on Microsoft Azure that provides predictable performance and pricing with 99.99% availability.
- Offers elastic database pools and data protection services like geo-replication and point-in-time restore.
- Compatible with SQL Server 2014 databases and comes in multiple service tiers based on resources like CPU cores and IOPS.
The document discusses the planned key rollover of the DNSSEC Key Signing Key (KSK) for the root zone from the current KSK-2010 to a new KSK-2017. It provides details on the milestones, approach, and state of the rollover process according to the Automated Updates of DNSSEC Trust Anchors protocol. The rollover was paused in 2017 due to uncertainty in measurement data, but progress has since been made to complete the rollover in 2018.
This document provides a list of media needed to install or upgrade SAP ERP software with Enhancement Package 6. It includes media for Java components, languages, ERP components, enhancement packages, databases like DB2, MaxDB, SQL Server, and operating system kernels for Linux, Windows, AIX, HP-UX, and Solaris. Each media item listed provides the material number, label, and name to use to download it from the SAP software distribution center.
Industry leaders Cisco, NetApp, VMware and Symantec have teamed up to develop a best practice framework and performance benchmark based on the VMware vSphere® Storage APIs - Data Protection (VADP). The test configuration uses the popular NetApp FlexPod environment, and the result proves that you can easily protect over 4 TB of virtual machine data per hour. And improved backup performance creates more reliable backups, shorter backup windows and less impact on the vSphere infrastructure.
In this session, we will show how these performance numbers can be easily obtained with minimal hardware and a small budget. In addition to backup performance, we will also discuss restore performance considerations.
Key topics include:
• How to select the correct hardware for the best ROI
• Strategies for minimizing backup impact and maximizing backup throughput
• Performance characteristics of VADP
• SAN or NBD (network) transports: which is recommended?
• Configurations for the fastest possible restores
Behind story in publisher that developer doesn't knowDavid Kim
The document discusses various technical aspects of managing a game development project and live game operations. It touches on topics like patching processes, crash reporting, in-game events, and server maintenance. It provides recommendations to simplify processes like patching by only downloading differential files, have servers automatically restart when resource usage exceeds a threshold, and handle in-game events and item giveaways through triggered processes rather than taking servers offline. The overall aim appears to be improving stability, player experience, and developer work-life balance.
Following our EBS R12.1.3 upgrade, we experienced inconsistent runtime and resource utilization with the Accrual Reconciliation Load program. During one month end close, a plant accountant could run accrual reconciliation for three years of data in under 10 hours. The next month, the same amount of data would take upwards of 30 hours to load. This behavior made it hard to plan for month end closure; successful accrual load being crucial to the finals for the month. Through multiple rounds of testing, we were able to complete the reconciliation process with nearly 15 years of data in just 1 hour and 45 minutes.
Implementing SharePoint on Azure, Lessons Learnt from a Real World ProjectK.Mohamed Faizal
This document discusses lessons learned from implementing SharePoint on Azure. It covers Azure architecture concepts like virtual networks, cloud services, availability sets, and load balancing. It provides an example reference architecture for a hybrid on-premises and Azure environment. It also discusses topics like database planning, disk performance, server topology with multiple tiers, and reserving IP addresses. The presentation aims to share best practices for deploying SharePoint on Azure based on a real-world project.
[NetApp] Managing Big Workspaces with Storage MagicPerforce
If you work with large volumes of data—multimedia assets, video game art, or firmware designs—you understand the pain of trying to quickly get a copy of source and build assets. But if you have the right storage system, you can be up and running with a new Perforce workspace in minutes instead of hours. See a simple procedure for fast workspace cloning using a few Perforce commands and NetApp FlexClone.
DataEngConf SF16 - Collecting and Moving Data at Scale Hakka Labs
This document summarizes Sada Furuhashi's presentation on Fluentd, an open source data collector. Fluentd provides a centralized way to collect, filter, and output log data from various sources like applications, servers, and databases. It addresses challenges with typical log collection architectures that have high latency, complex parsing, and a combination explosion of connections. Fluentd uses a plugin-based architecture with input, filter, and output components to flexibly collect, transform, and deliver log data at scale to targets like files, databases and visualization tools. Many large companies like Microsoft, Atlassian and Amazon use Fluentd for log collection and analytics in production environments.
The document discusses monitoring input/output (IO) performance in Oracle Exadata systems. It covers write-back flash cache (WBFC), various methods for monitoring IO using Automatic Workload Repository (AWR) data and cell-level scripts, correlating IO to workload, and scaling monitoring using metric extensions and Business Intelligence Publisher (BIP). The presentation provides examples of visualizing IO performance trends over time using AWR and cell data and measuring the impact of initialization parameters on latency. It also addresses reference bands for disk IO capacity and visualizing workload activity storage area activity by day per node.
Similar a How do databases perform live backups and point-in-time recovery (20)
Postgres indexes: how to make them work for your applicationBartosz Sypytkowski
Indexes are one of the most crucial structures of any relational database. In this talk we'll explain how to use them efficiently, how to read query plans and what do they mean for us. We'll also cover a variety of different indexing structures available in PostgreSQL database and build up some intuition about which one to pick depending on the situation.
This presentation covers HyParView and Plumtree - protocols used to build highly scalable clusters of data capable of gossiping messages between thousands of clients.
In this talk we'll discuss technical foundations behind Conflict-free Replicated Data Types (CRDT), which let us create collaborative client applications - systems where no reliance on central servers and offline-first capabilities are one of the founding principles. We'll cover some of the challenges bound to this approach and how to address them. Finally we'll present Yrs - Rust library, that allows us to build rich collaborative applications on desktop and browser.
The document provides an overview of PostgreSQL indexes, including the different types: B-Tree, Hash, BRIN, Bloom, GiST, SP-GiST, GIN, and RUM indexes. It explains how each index type stores and organizes data, as well as when each type is best suited in terms of performance, size, and supported query types such as equality scans, range scans, and full-text search. The document also covers index-only scans, bitmap scans, and tuple identifiers to help explain how indexes are used during query execution.
The document discusses modern concurrency primitives like threads, thread pools, coroutines, and schedulers. It covers why asynchronous programming with async/await is preferred over traditional threading. It also discusses challenges like sharing data across threads and blocking on I/O calls. Some solutions covered include using thread pools with dedicated I/O threads, work stealing, and introducing interruption points in long-running tasks.
During this presentation we'll quickly cover the core principles of eventsourced systems and different approaches to scalling event log to distributed workload. We'll focus on peer-to-peer variants of such: what are their advantages and disadvantages and how we can use them.
During this talk we'll cover the theory and practical implementation behind most common patterns in modern multi-threaded programming. How our everyday libraries and frameworks optimize use of operating system resources for maximum efficiency. We'll also try to understand differences between various approaches and what tradeoffs do they infer. Finally we'll take a look at how they are supported by various compilers and runtimes.
Strongly consistent databases are dominating world of software. However, with increasing scale and global availability of our services, many developers often prefer to loose their constraints in favor of an eventual consistency.
During this presentation we'll talk about Conflict-free Replicated Data Types (CRDT) - an eventually-consistent structures, that can be found in many modern day multi-master, geo-distributed databases such as CosmosDB, DynamoDB, Riak, Cassandra or Redis: how do they work and what makes them so interesting choice in highly available systems.
This is presentation from WG.NET (May 2019), where I'm discussing different aspects of virtualization, mainly in the context of programming languages. We'll covering up what stack vs. register based virtual machines are, what is interpreter and compiler and how to build our own bytecode interpreter for a toy programming language.
This document discusses timekeeping in distributed systems. It begins by explaining how different types of clocks work, from pendulum clocks to atomic clocks. It then discusses key concepts like UTC, leap seconds, and how time is represented in Unix. The document also covers challenges of keeping time across distributed systems and algorithms like NTP, vector clocks, and logical clocks that help order events in a distributed system.
This document provides an introduction to Akka.NET Streams and Reactive Streams. It discusses key concepts like observables, async enumerables, and reactive streams. It also demonstrates how to build workflows with Akka.NET streams, including examples of building a TCP server. The document introduces core Akka.NET streams concepts like sources, flows, and sinks, and how they compose together in a runnable graph. It also covers testing streams with probes and materialization.
This is presentiation for Lambda Days 2019, in which I describe details behind building collaborative text editing experience using Replicated Growable Array CRDTs. Later on we come to defining its issues and how to solve them.
1. The document discusses different database storage structures like B+ trees, LSM trees, and their pros and cons for storing structured data on disk.
2. B+ trees are optimized for read performance but require copy-on-write or write-ahead logging for updates. LSM trees prioritize write performance using an append-only structure but require background merging.
3. Bloom filters can help optimize look ups in LSM trees by quickly checking if an element is not present in a collection without accessing all files.
Slides from presentation, I've made on the BuildStuff LT 2018. Here I'm talking about issues, many people have found when using RESTful APIs and how GraphQL addresses them. Also I'm trying to cover the tradeoffs made by the standard, solutions proposed by different implementations and some ideas for the future.
Mobile App Development Company In Noida | Drona InfotechDrona Infotech
Drona Infotech is a premier mobile app development company in Noida, providing cutting-edge solutions for businesses.
Visit Us For : https://www.dronainfotech.com/mobile-application-development/
Consistent toolbox talks are critical for maintaining workplace safety, as they provide regular opportunities to address specific hazards and reinforce safe practices.
These brief, focused sessions ensure that safety is a continual conversation rather than a one-time event, which helps keep safety protocols fresh in employees' minds. Studies have shown that shorter, more frequent training sessions are more effective for retention and behavior change compared to longer, infrequent sessions.
Engaging workers regularly, toolbox talks promote a culture of safety, empower employees to voice concerns, and ultimately reduce the likelihood of accidents and injuries on site.
The traditional method of conducting safety talks with paper documents and lengthy meetings is not only time-consuming but also less effective. Manual tracking of attendance and compliance is prone to errors and inconsistencies, leading to gaps in safety communication and potential non-compliance with OSHA regulations. Switching to a digital solution like Safelyio offers significant advantages.
Safelyio automates the delivery and documentation of safety talks, ensuring consistency and accessibility. The microlearning approach breaks down complex safety protocols into manageable, bite-sized pieces, making it easier for employees to absorb and retain information.
This method minimizes disruptions to work schedules, eliminates the hassle of paperwork, and ensures that all safety communications are tracked and recorded accurately. Ultimately, using a digital platform like Safelyio enhances engagement, compliance, and overall safety performance on site. https://safelyio.com/
Hand Rolled Applicative User ValidationCode KataPhilip Schwarz
Could you use a simple piece of Scala validation code (granted, a very simplistic one too!) that you can rewrite, now and again, to refresh your basic understanding of Applicative operators <*>, <*, *>?
The goal is not to write perfect code showcasing validation, but rather, to provide a small, rough-and ready exercise to reinforce your muscle-memory.
Despite its grandiose-sounding title, this deck consists of just three slides showing the Scala 3 code to be rewritten whenever the details of the operators begin to fade away.
The code is my rough and ready translation of a Haskell user-validation program found in a book called Finding Success (and Failure) in Haskell - Fall in love with applicative functors.
How Can Hiring A Mobile App Development Company Help Your Business Grow?ToXSL Technologies
ToXSL Technologies is an award-winning Mobile App Development Company in Dubai that helps businesses reshape their digital possibilities with custom app services. As a top app development company in Dubai, we offer highly engaging iOS & Android app solutions. https://rb.gy/necdnt
Malibou Pitch Deck For Its €3M Seed Roundsjcobrien
French start-up Malibou raised a €3 million Seed Round to develop its payroll and human resources
management platform for VSEs and SMEs. The financing round was led by investors Breega, Y Combinator, and FCVC.
14 th Edition of International conference on computer visionShulagnaSarkar2
About the event
14th Edition of International conference on computer vision
Computer conferences organized by ScienceFather group. ScienceFather takes the privilege to invite speakers participants students delegates and exhibitors from across the globe to its International Conference on computer conferences to be held in the Various Beautiful cites of the world. computer conferences are a discussion of common Inventions-related issues and additionally trade information share proof thoughts and insight into advanced developments in the science inventions service system. New technology may create many materials and devices with a vast range of applications such as in Science medicine electronics biomaterials energy production and consumer products.
Nomination are Open!! Don't Miss it
Visit: computer.scifat.com
Award Nomination: https://x-i.me/ishnom
Conference Submission: https://x-i.me/anicon
For Enquiry: Computer@scifat.com
What to do when you have a perfect model for your software but you are constrained by an imperfect business model?
This talk explores the challenges of bringing modelling rigour to the business and strategy levels, and talking to your non-technical counterparts in the process.
Preparing Non - Technical Founders for Engaging a Tech AgencyISH Technologies
Preparing non-technical founders before engaging a tech agency is crucial for the success of their projects. It starts with clearly defining their vision and goals, conducting thorough market research, and gaining a basic understanding of relevant technologies. Setting realistic expectations and preparing a detailed project brief are essential steps. Founders should select a tech agency with a proven track record and establish clear communication channels. Additionally, addressing legal and contractual considerations and planning for post-launch support are vital to ensure a smooth and successful collaboration. This preparation empowers non-technical founders to effectively communicate their needs and work seamlessly with their chosen tech agency.Visit our site to get more details about this. Contact us today www.ishtechnologies.com.au
UI5con 2024 - Boost Your Development Experience with UI5 Tooling ExtensionsPeter Muessig
The UI5 tooling is the development and build tooling of UI5. It is built in a modular and extensible way so that it can be easily extended by your needs. This session will showcase various tooling extensions which can boost your development experience by far so that you can really work offline, transpile your code in your project to use even newer versions of EcmaScript (than 2022 which is supported right now by the UI5 tooling), consume any npm package of your choice in your project, using different kind of proxies, and even stitching UI5 projects during development together to mimic your target environment.
Most important New features of Oracle 23c for DBAs and Developers. You can get more idea from my youtube channel video from https://youtu.be/XvL5WtaC20A
Everything You Need to Know About X-Sign: The eSign Functionality of XfilesPr...XfilesPro
Wondering how X-Sign gained popularity in a quick time span? This eSign functionality of XfilesPro DocuPrime has many advancements to offer for Salesforce users. Explore them now!
11. SQLITE
ROLLBACK
JOURNAL P1
P2 P3
UPDATE users SET name = ‘Joe’ WHERE id = 1;
B+Tree
Pager
Database
File
P1
P2
P3
Rollback
Journal
Locate page
with the
record
12. SQLITE
ROLLBACK
JOURNAL P1
P2 P3
UPDATE users SET name = ‘Joe’ WHERE id = 1;
B+Tree
Pager
Database
File
P1
P2
P3
Rollback
Journal
P2
Copy pages to
Rollback
Journal
P1
13. SQLITE
ROLLBACK
JOURNAL P1
P2 P3
UPDATE users SET name = ‘Joe’ WHERE id = 1;
B+Tree
Pager
Database
File
P1
P2
P3
Rollback
Journal
P2
P1
Modify
affected
pages
17. SQLITE
WRITE-AHEAD
LOG P1
P2 P3
UPDATE users SET name = ‘Joe’ WHERE id = 1;
B+Tree
Pager
Database
File
P1
P2
P3
Write-Ahead
Log
Locate page
with the
record
18. SQLITE
WRITE-AHEAD
LOG P1
P2 P3
UPDATE users SET name = ‘Joe’ WHERE id = 1;
B+Tree
Pager
Database
File
P1
P2
P3
Write-Ahead
Log
P1
P2
Write new
page versions
19. SQLITE
WRITE-AHEAD
LOG P1
P2 P3
UPDATE users SET name = ‘Joe’ WHERE id = 1;
B+Tree
Pager
Database
File
P1
P2
P3
Write-Ahead
Log
P1
P2
Pager
redirects
pages to new
versions
24. EVERY CHANGED PAGE MUST BE
FIRST APPENDED AT THE END OF
THE WRITE-AHEAD LOG FILE
25. EVERY CHANGE IN THE DATABASE
FILE COMES FROM READING
FRONT-TO-BACK WRITE-AHEAD
LOG FILE
EVERY CHANGED PAGE MUST BE
FIRST APPENDED AT THE END OF
THE WRITE-AHEAD LOG FILE
38. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
restore(2024/02/11/10:00:00)
39. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1
P4
F2
P3
F3
F1-F3-2024/02/10/07:54:00
40. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1 P3
F1 P3
F1
41. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1 P3
F1 P3
F1
P1
F4
P2
F5
F4-F5-2024/02/10/07:54:10
42. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1 P3
F1 P3
F1 P1
F4 P2
F5
43. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1 P3
F1 P3
F1 P1
F4 P2
F5
PRAGMA CHECKPOINT(TRUNCATE);
44. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1 P3
F1 P3
F1 P1
F4 P2
F5
45. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1 P3
F1 P3
F1 P1
F4 P2
F5
46. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1 P3
F1 P3
F1 P1
F4 P2
F5
47. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1 P3
F1 P3
F1 P1
F4 P2
F5
48. POINT-IN-TIME RECOVRERY
101
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1 P3
F1 P3
F1 P1
F4 P2
F5
60. TRANSACTION
COMMIT & ROLLBACK
Write-Ahead Log
Database process
P3
F1 P1
F2
size_after=0 size_after=3
T1
P3
F3 P1
F4
size_after=0 size_after=0
ROLLBACK;
EOF
what if we already backed up these
frames? :/
62. POINT-IN-TIME RECOVRERY PROBLEM
Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
restore(2024/02/11/10:00:00)
63. Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
P3
F1 P3
F1 P3
F1 P1
F4 P2
F5
POINT-IN-TIME RECOVRERY PROBLEM
F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
64. Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
POINT-IN-TIME RECOVRERY PROBLEM
INSERT INTO t(name)
VALUES(‘John Doe’);
65. Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
POINT-IN-TIME RECOVRERY PROBLEM
P5
F1 P1
F2
66. Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
POINT-IN-TIME RECOVRERY PROBLEM
P5
F1 P1
F2
P5
F1
P1
F2
F1-F2-2024/02/13/09:30:00
67. Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
POINT-IN-TIME RECOVRERY PROBLEM
P5
F1 P1
F2
P5
F1
P1
F2
F1-F2-2024/02/13/09:30:00
68. Write-Ahead Log
Database process
Backup service Backup drive F1-F3-2024/02/10/07:54:00
F4-F5-2024/02/10/07:54:10
F6-F7-2024/02/11/11:02:00
F8-F9-2024/02/12/20:30:00
POINT-IN-TIME RECOVRERY PROBLEM
P5
F1 P1
F2
P5
F1
P1
F2
F1-F2-2024/02/13/09:30:00
Database history is no longer linear!
87. How does continuous backup and point-in-time recovery work in databases:
https://www.bartoszsypytkowski.com/db-backup-point-in-time-recovery
Litestream: https://litestream.io/
SQLite write-ahead log docs: https://www.sqlite.org/wal.html
REFERENCES