For a long time, relational database management systems have been the only solution for persistent data store. However, with the phenomenal growth of data, this conventional way of storing has become problematic.
To manage the exponentially growing data traffic, largest information technology companies such as Google, Amazon and Yahoo have developed alternative solutions that store data in what have come to be known as NoSQL databases.
Some of the NoSQL features are flexible schema, horizontal scaling and no ACID support. NoSQL databases store and replicate data in distributed systems, often across datacenters, to achieve scalability and reliability.
The CAP theorem states that any networked shared-data system (e.g. NoSQL) can have at most two of three desirable properties:
• consistency(C) - equivalent to having a single up-to-date copy of the data
• availability(A) of that data (for reads and writes)
• tolerance to network partitions(P)
Because of this inherent tradeoff, it is necessary to sacrifice one of these properties. The general belief is that designers cannot sacrifice P and therefore have a difficult choice between C and A.
In this seminar two NoSQL databases are presented: Amazon's Dynamo, which sacrifices consistency thereby achieving very high availability and Google's BigTable, which guarantees strong consistency while provides only best-effort availability.
For a long time, relational database management systems have been the only solution for persistent data store. However, with the phenomenal growth of data, this conventional way of storing has become problematic.
To manage the exponentially growing data traffic, largest information technology companies such as Google, Amazon and Yahoo have developed alternative solutions that store data in what have come to be known as NoSQL databases.
Some of the NoSQL features are flexible schema, horizontal scaling and no ACID support. NoSQL databases store and replicate data in distributed systems, often across datacenters, to achieve scalability and reliability.
The CAP theorem states that any networked shared-data system (e.g. NoSQL) can have at most two of three desirable properties:
• consistency(C) - equivalent to having a single up-to-date copy of the data
• availability(A) of that data (for reads and writes)
• tolerance to network partitions(P)
Because of this inherent tradeoff, it is necessary to sacrifice one of these properties. The general belief is that designers cannot sacrifice P and therefore have a difficult choice between C and A.
In this seminar two NoSQL databases are presented: Amazon's Dynamo, which sacrifices consistency thereby achieving very high availability and Google's BigTable, which guarantees strong consistency while provides only best-effort availability.
The Google File System (GFS) presented in 2003 is the inspiration for the Hadoop Distributed File System (HDFS). Let's take a deep dive into GFS to better understand Hadoop.
A Distributed File System(DFS) is simply a classical model of a file system distributed across multiple machines.The purpose is to promote sharing of dispersed files.
Designed by Sanjay Ghemawat , Howard Gobioff and Shun-Tak Leung of Google in 2002-03.
Provides fault tolerance, serving large number of clients with high aggregate performance.
The field of Google is beyond the searching.
Google store the data in more than 15 thousands commodity hardware.
Handles the exceptions of Google and other Google specific challenges in their distributed file system.
The TCP/IP protocol suite does not define any protocol in the data-link layer or
physical layer. These two layers are territories of networks that when connected
make up the Internet. These networks, wired or wireless, provide services to the upper
three layers of the TCP/IP suite. This may give us a clue that there are several standard
protocols in the market today. For this reason, we discuss the data-link layer in several
chapters. This chapter is an introduction that gives the general idea and common issues
in the data-link layer that relate to all networks.
❑ The first section introduces the data-link layer. It starts with defining the concept
of links and nodes. The section then lists and briefly describes the services provided
by the data-link layer. It next defines two categories of links: point-to-point
and broadcast links. The section finally defines two sublayers at the data-link layer
that will be elaborated on in the next few chapters.
❑ The second section discusses link-layer addressing. It first explains the rationale
behind the existence of an addressing mechanism at the data-link layer. It then
describes three types of link-layer addresses to be found in some link-layer protocols.
The section discusses the Address Resolution Protocol (ARP), which maps
the addresses at the network layer to addresses at the data-link layer. This protocol
helps a packet at the network layer find the link-layer address of the next node for
delivery of the frame that encapsulates the packet. To show how the network layer
helps us to find the data-link-layer addresses, a long example is included in this
section that shows what happens at each node when a packet is travelling through
the Internet.
In this presentation we describe the design and implementation of Kafka Connect, Kafka’s new tool for scalable, fault-tolerant data import and export. First we’ll discuss some existing tools in the space and why they fall short when applied to data integration at large scale. Next, we will explore Kafka Connect’s design and how it compares to systems with similar goals, discussing key design decisions that trade off between ease of use for connector developers, operational complexity, and reuse of existing connectors. Finally, we’ll discuss how standardizing on Kafka Connect can ultimately lead to simplifying your entire data pipeline, making ETL into your data warehouse and enabling stream processing applications as simple as adding another Kafka connector.
Google has designed and implemented a scalable distributed file system for their large distributed data intensive applications. They named it Google File System, GFS.
PowerPoint Presentation on Distributed Operating Systems,reasons for opting for distributed systems over centralized systems,types of Distributed Systems,Process Migration and its advantages.
The Google File System (GFS) presented in 2003 is the inspiration for the Hadoop Distributed File System (HDFS). Let's take a deep dive into GFS to better understand Hadoop.
A Distributed File System(DFS) is simply a classical model of a file system distributed across multiple machines.The purpose is to promote sharing of dispersed files.
Designed by Sanjay Ghemawat , Howard Gobioff and Shun-Tak Leung of Google in 2002-03.
Provides fault tolerance, serving large number of clients with high aggregate performance.
The field of Google is beyond the searching.
Google store the data in more than 15 thousands commodity hardware.
Handles the exceptions of Google and other Google specific challenges in their distributed file system.
The TCP/IP protocol suite does not define any protocol in the data-link layer or
physical layer. These two layers are territories of networks that when connected
make up the Internet. These networks, wired or wireless, provide services to the upper
three layers of the TCP/IP suite. This may give us a clue that there are several standard
protocols in the market today. For this reason, we discuss the data-link layer in several
chapters. This chapter is an introduction that gives the general idea and common issues
in the data-link layer that relate to all networks.
❑ The first section introduces the data-link layer. It starts with defining the concept
of links and nodes. The section then lists and briefly describes the services provided
by the data-link layer. It next defines two categories of links: point-to-point
and broadcast links. The section finally defines two sublayers at the data-link layer
that will be elaborated on in the next few chapters.
❑ The second section discusses link-layer addressing. It first explains the rationale
behind the existence of an addressing mechanism at the data-link layer. It then
describes three types of link-layer addresses to be found in some link-layer protocols.
The section discusses the Address Resolution Protocol (ARP), which maps
the addresses at the network layer to addresses at the data-link layer. This protocol
helps a packet at the network layer find the link-layer address of the next node for
delivery of the frame that encapsulates the packet. To show how the network layer
helps us to find the data-link-layer addresses, a long example is included in this
section that shows what happens at each node when a packet is travelling through
the Internet.
In this presentation we describe the design and implementation of Kafka Connect, Kafka’s new tool for scalable, fault-tolerant data import and export. First we’ll discuss some existing tools in the space and why they fall short when applied to data integration at large scale. Next, we will explore Kafka Connect’s design and how it compares to systems with similar goals, discussing key design decisions that trade off between ease of use for connector developers, operational complexity, and reuse of existing connectors. Finally, we’ll discuss how standardizing on Kafka Connect can ultimately lead to simplifying your entire data pipeline, making ETL into your data warehouse and enabling stream processing applications as simple as adding another Kafka connector.
Google has designed and implemented a scalable distributed file system for their large distributed data intensive applications. They named it Google File System, GFS.
PowerPoint Presentation on Distributed Operating Systems,reasons for opting for distributed systems over centralized systems,types of Distributed Systems,Process Migration and its advantages.
Distributed systems involve complex interactions among many components. This increases the possibilities of failures that could turn a whole system down. Software architects, designers, and developers need to architect, design, and program functional requirements thinking about possibility of failures and the need for a system to keep running despite failures. This presentation tackles but part of the problem, focusing on redundancy, different types of groups, replication, and eventual consistency, finishing with the presentation of CAP theorem.
Presentation delivered at IV Cloud Computing and Big Data Ent at Universdad Nacional de La Plata http://www.jcc.info.unlp.edu.ar/jcc2016/wordpress/index.php/cronograma/
Google is a multi-billion dollar company. It's one of the big power players on the World Wide Web and beyond. The company relies on a distributed computing system to provide users with the infrastructure they need to access, create and alter data.
Surely Google buys state-of-the-art computers and servers to keep things running smoothly, right?
Wrong. The machines that power Google's operations aren't cutting-edge power computers with lots of bells and whistles. In fact, they're relatively inexpensive machines running on Linux operating systems. How can one of the most influential companies on the Web rely on cheap hardware? It's due to the Google File System (GFS), which capitalizes on the strengths of off-the-shelf servers while compensating for any hardware weaknesses. It's all in the design.
Google uses the GFS to organize and manipulate huge files and to allow application developers the research and development resources they require. The GFS is unique to Google and isn't for sale. But it could serve as a model for file systems for organizations with similar needs.
Transferring files from one computer to another is one of the most common tasks expected from a networking or internetworking environment. In reality, the greatest volume of data exchange in the Internet today is due to file transfer.
This module has been created to answer all the questions on how IPFS can be used for dynamic real-time applications. In this module, you will learn:
- how to reason about dynamic data on IPFS,
- IPNS, the simplest construction for naming in IPFS,
- how PubSub can offer subsecond speeds for interactive applications,
- how CRDTs are a fundamental building block for distributed applications,
- what is available in the ecosystem.
February 2017 HUG: Exactly-once end-to-end processing with Apache ApexYahoo Developer Network
Apache Apex (http://apex.apache.org/) is a stream processing platform that helps organizations to build processing pipelines with fault tolerance and strong processing guarantees. It was built to support low processing latency, high throughput, scalability, interoperability, high availability and security. The platform comes with Malhar library - an extensive collection of processing operators and a wide range of input and output connectors for out-of-the-box integration with an existing infrastructure. In the talk I am going to describe how connectors together with the distributed checkpointing (a mechanism used by the Apex to support fault tolerance and high availability) provide exactly-once end-to-end processing guarantees.
Speakers:
Vlad Rozov is Apache Apex PMC member and back-end engineer at DataTorrent where he focuses on the buffer server, Apex platform network layer, benchmarks and optimizing the core components for low latency and high throughput. Prior to DataTorrent Vlad worked on distributed BI platform at Huawei and on multi-dimensional database (OLAP) at Hyperion Solutions and Oracle.
The Building Blocks of QuestDB, a Time Series Databasejavier ramirez
Talk Delivered at Valencia Codes Meetup 2024-06.
Traditionally, databases have treated timestamps just as another data type. However, when performing real-time analytics, timestamps should be first class citizens and we need rich time semantics to get the most out of our data. We also need to deal with ever growing datasets while keeping performant, which is as fun as it sounds.
It is no wonder time-series databases are now more popular than ever before. Join me in this session to learn about the internal architecture and building blocks of QuestDB, an open source time-series database designed for speed. We will also review a history of some of the changes we have gone over the past two years to deal with late and unordered data, non-blocking writes, read-replicas, or faster batch ingestion.
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...2023240532
Quantitative data Analysis
Overview
Reliability Analysis (Cronbach Alpha)
Common Method Bias (Harman Single Factor Test)
Frequency Analysis (Demographic)
Descriptive Analysis
Adjusting OpenMP PageRank : SHORT REPORT / NOTESSubhajit Sahu
For massive graphs that fit in RAM, but not in GPU memory, it is possible to take
advantage of a shared memory system with multiple CPUs, each with multiple cores, to
accelerate pagerank computation. If the NUMA architecture of the system is properly taken
into account with good vertex partitioning, the speedup can be significant. To take steps in
this direction, experiments are conducted to implement pagerank in OpenMP using two
different approaches, uniform and hybrid. The uniform approach runs all primitives required
for pagerank in OpenMP mode (with multiple threads). On the other hand, the hybrid
approach runs certain primitives in sequential mode (i.e., sumAt, multiply).
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
3. WASHINGTON
STATE
UNIVERSITY
PROBLEM EXISTING
3
❖ Frequent Failures of Nodes
❖ Files are huge – multi-GB
❖ Access Patterns
➢ Most of them were sequential Read/Write/append
❖ Scalability
➢ Concurrent Writers
Fig: Andrew File System Architecture
5. WASHINGTON
STATE
UNIVERSITY
ASSUMPTIONS
5
❖ Designed using cheap hardwares
➢ needs constant monitor, detect, tolerate & recover regularly as failure
is obvious
❖ Files are huge ( +100MB to GB)
❖ Workloads:
➢ Large Streaming Reads: Read only(1 MB+)
➢ Small Random Reads: unmodified after written (KB)
➢ includes many writes than append data to files
❖Throughput is more valued than individual request latency
6. WASHINGTON
STATE
UNIVERSITY
INTERFACE
6
❖ Doesn’t support Portable Operating System Interface (POSIX)
➢ supports typical file system operations: create, delete, open, close, read, and
write
❖ Supports Snapshot
➢ creates a copy of a file or a directory tree at low cost
➢ Duplicate metadata
❖ Supports Record append
➢ allows multiple clients to append data to the same file concurrently
8. WASHINGTON
STATE
UNIVERSITY
CHUNK
8
❖ Files are divided into fixed size blocks called chunk
❖ 64 MB; greater than typical file system block size
❖ Each chunk is replicated 3 or more times
❖ Each chunk is identified by 64-bit chunk handle
9. WASHINGTON
STATE
UNIVERSITY
META DATA
9
❖ Three major types of metadata
➢ The file and chunk namespaces
➢ The mapping from files to chunks
➢ Locations of each chunk’s replicas
❖ All the metadata is kept in the Master’s memory
❖ 64MB chunk has 64 bytes of metadata
❖ Chunk Location are updated on every restart & heartbeat message
❖ Operation log contains a historical record of critical metadata changes.
12. WASHINGTON
STATE
UNIVERSITY
GFS MASTER & CLIENT
12
❖ Single Master : maintains all file system metadata
➢ Namespaces, Access Control, mappings from files to chunks, and current
locations of chunks
❖ Clients never read and write file data through the master (Bottleneck)
➢ asks the master which chunk servers it should contact
❖ Communicates with each chunkserver in HeartBeat messages
➢ Is chunkserver up or down?
➢ Are there any disk failures on chunkserver?
➢ Are any replicas corrupted?
21. WASHINGTON
STATE
UNIVERSITY
Leases and Mutation Order
21
❖ Mutation : operation that changes the contents or metadata of a chunk
❖ Lease : to maintain a consistent mutation order across replicas
❖ If the master receives a modification operation for a particular chunk
➢ Master finds the chunk servers that have the chunk and grants a chunk lease to
one of them (primary)
➢ The primary determines the serialization order for all of the chunk’s modifications,
and the secondaries follow that order
➢ Leases timeout at 60 seconds.(also possible to extend the timeout)
24. WASHINGTON
STATE
UNIVERSITY
24
Fig: Data Flow & Control Flow
● Master replies with the identity of the primary and the locations of the other
(secondary) replicas if available
● client caches this data for future mutations & contact again if unreachable
Step
2
26. WASHINGTON
STATE
UNIVERSITY
26
Fig: Data Flow & Control Flow
Client sends write request to primary. Primary decides serialization order for
all incoming modifications and applies them to the chunk
Step
4
29. WASHINGTON
STATE
UNIVERSITY
29
Fig: Data Flow & Control Flow
Primary replies back to the client, either with success or error
If Fails, Client can retry steps (3) through (7)
Step
7
31. WASHINGTON
STATE
UNIVERSITY
NAMESPACE MANAGEMENT &
LOCKING
31
❖ Master maintain a table which map full path
name to metadata
❖ Locks are used over namespaces to ensure
proper serialization
❖ Each node in the namespace has
associated read-write lock
❖ Concurrent operations can be properly
serialized by locking mechanism
32. WASHINGTON
STATE
UNIVERSITY
OTHER MASTER OPERATIONS
32
❖ Replica Placement
➢ GFS place replicas over different racks for reliability and availability
➢ Maximum network bandwidth utilization
❖ Creation, Re-replication, Rebalancing
➢ New replicas are created on below average disk utilization
➢ Re-replication is done when available replicas fall below user specialized goal
➢ Rebalancing is done periodically for better disk space and load balancing
33. WASHINGTON
STATE
UNIVERSITY
OTHER MASTER OPERATIONS
33
❖ Garbage Collection
➢ Deletion operation is logged
➢ Filename is renamed to a hidden name; can be later deleted or recovered
➢ Orphan chunks are removed during regular scan of chunk namespace
❖ Stale Replica Detection
➢ Stale if a chunk server fails and misses mutations to the chunk while it is down
➢ Master stored chunk version used to identify
➢ The master removes stale replicas in its regular garbage collection
35. WASHINGTON
STATE
UNIVERSITY
HIGH AVAILABILITY
35
❖ Fast Recovery
➢ restore their state and start in seconds no matter how they terminated.
➢ Operation logs & Checkpoints
❖ Chunk Replication
➢ each chunk is replicated on multiple chunk servers (3 or more) on different racks
❖ Master Replication
➢ Operation logs & Checkpoints are replicated on various machines
➢ “shadow” masters provide read-only access when the primary master is down (
depends on create delete updates from primary)
37. WASHINGTON
STATE
UNIVERSITY
DATA INTEGRITY
37
❖ Checksum
➢ to check if there is corrupted data
➢ Each chunkserver independently verifies integrity
❖ A chunk is broken up into 64 KB blocks has 32 bit checksum
❖ Checksums are kept in memory and stored persistently with logging,
separate from user data.
38. WASHINGTON
STATE
UNIVERSITY
Example : CheckSum
38
carryout is added
Data added
Data To be sent
Checksum :1’s compliment
Note: Assume receiver receives the same data sent by sender, the data is said to be correct when
sum of checksum & received data contains all value 1
45. WASHINGTON
STATE
UNIVERSITY
ATOMIC RECORD APPEND
45
❖ Primary Chunk Server checks if append exceeds max chunk size
❖ If so, it pads the chunk to max chunk size
❖ Secondary chunk servers do the same
❖ On failure, Client retries the operation
❖GFS appends data at least once automatically