• Save
Dale Wickizer, CTO, NetApp Public Sector
Upcoming SlideShare
Loading in...5
×
 

Dale Wickizer, CTO, NetApp Public Sector

on

  • 1,199 views

The “Tech”Tonic Shift

The “Tech”Tonic Shift

Statistics

Views

Total Views
1,199
Views on SlideShare
1,105
Embed Views
94

Actions

Likes
0
Downloads
0
Comments
0

2 Embeds 94

http://fedscoop.com 59
http://storify.com 35

Accessibility

Categories

Upload Details

Uploaded via as Microsoft PowerPoint

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment
  • How many people were here for the earthquake we had on August 23 of this year? I was actually downtown hosting a panel discussion at an event where the theme was…I kid you not… “Shake IT up”…you just can’t make this stuff up. :^)Today I want to talk to you about a different “Tech”tonic shift: One that is definitely going to shake up traditional enterprise applications as well as the IT organizations that manage them.
  • That shift isbeing driven by an explosion of data being generated and consumed in the world. Data has grown by a factor of 9 over the past 5 years, crossing 1.2 ZB for the first time! (If anyone wonders what 1.2 ZB is, Wikibon has this great graphic, showing it is the equivalent of 75 billion fully loaded iPads, stacked end-to-end and side-by-side, covering Wembley stadium, in a column more than 4 miles high). This year it will grow to 1.8 ZB. More than 90% of this data was unstructured and much of it machine generated, in response to data stored by end users. Over the next decade, this data growth is expected to accelerate, increasing by a factor of 50. Over the same time, the number of files is expected to increase by more than a factor of 75, which will break many traditional applications and file systems. Sadly, the number of IT professionals will only grow by less than 50% and IT budgets will remain relatively flat!
  • We already see enterprises beginning to adopt High Speed Analytics and and scale out storage technologies to provide new Decision Support methodologies.We see new non-traditional NoSQL and columnar database applications coming online everyday, like MongoDB, Cassandra, ParAccel and Hbase. These applications leverage the fast search and index capabilities of Hadoop. They are able to process workloads 100 times faster than Teradata, Netezza or Oracle Exadata over all types of data (structured and unstructured) at a fraction of the cost. In addition, we see the traditional players scrambling to have their own MapReduce functionality (GreenPlum, Teradata, Netezza, Oracle Exadata), but in reality they are using it only as a glorified ETL front-end. Their legacy structured database still sits behind it all. They don’t yet understand that with all this unstructured data coming in, nightly batch processing to do ETL may be going away. The data will be ingested and stored in its native unstructured format and processed directly from there for decision support.
  • The large volumes of data and metadata are going to drive the need for denser equipment racks, physically, electrically and thermally), placing a great strain to today’s data centers.Fortunately, containerized data centers have come of age and offer an economical and “green” alternative to traditional data centers. They are built to handle these densities, as well as the rigors of transport. Pictured here are some 20’ foot containers built by WWT. Yes, that is a C17 on the right.Some organizations, constrained by law or economics, have begun to invest in this technology and repurpose their old data centers and server rooms as office space, obviating the need to build new data centers and office buildings. Frankly, most organizations should probably give this approach more careful consideration, rather than spending tens of millions of dollars to refit old data centers or build new ones to keep pace with technology.
  • So What? Well there are several trends that Big Data could drive in large enterprises: Higher bandwidth data will result in more, large distributed repositories as network bandwidth lags behind. Tendency will be access data where it is generated and lives, vs moving it to a central location to save time, money and bandwidth. As already mentioned, unstructured data puts a strain on existing enterprise applications. There is already a shift beginning towards disruptive solutions based on NoSQL and MapReduce; distributed repositories put even more strain on the metadata layer. Look for a shakeup in the application vendor space as traditional vendors get kicked to the curb. Similarly the glut of more data will put strain on IT organizations, both from a management and an environmentals standpoint. Dense, object-based storage and new containerized data centers are cost-effective and efficient means that can help.

Dale Wickizer, CTO, NetApp Public Sector Dale Wickizer, CTO, NetApp Public Sector Presentation Transcript

  • Dale Wickizer
    Chief Technology Officer
    U.S. Public Sector
    The “Tech”tonic Shift
  • World Data Explosion
    2
    Growth Over the Next Decade:
    Servers (Phys/VM): 10x
    Data/Information: 50x
    #Files: 75x
    IT Professionals: <1.5x
    Source: Revisited: The Rapid Growth in Unstructured Data « Wikibon Blog http://bit.ly/oRSdXm
    • Growing 9x in 5 yrs! (1.8 ZB in 2011)
    • > 90% unstructured data
    • End user and machine generated
    Source: Gantz, John and Reinsel, David, “Extracting Value from Chaos”, IDC IVIEW, June 2011, page 4.
  • 3
    The Earth Is Moving Beneath Traditional Applications (“Tech”tonic Shift)
    High
    Tier 1 BP
    OLTP
    “Decision Support”
    High Speed
    Analytics
    Tier-2 BP
    OLTP
    “Enterprise Applications”
    • Shared, Virtualized Infrastructure
    • Integrated Data Protection
    • Secure Multi-Tenancy
    DSS/DW
    Collaboration
    No SQL
    Columnar
    DBs
    Web Infra
    App Dev
    Data Structure
    Sat Ground Stations
    “Big Data”
    • High Bandwidth Throughput
    • Big Data Content
    IT Infra
    FMV
    DVS
    Content
    Repositories
    Tech
    Comp
    HPC
    Home Dirs
    Low
    Performance
    Small Block,
    Random I/O (100s KIOPS)
    Large Block,
    Sequential I/O
    100s GB/sec
  • Big Data Landscape – Industry View
    4
    Big
    Analytics
    Big
    Content
    ParAccel
    Teradata
    ExaSol
    Greenplum
    Netezza
    PaaS
    New Apps
    StorageGrid
    Isilon/Atmos,
    StorNext.
    Azure
    GoogleApps
    VMware
    HDFS
    NoSQL
    FacebookHaystack
    E-Series
    Cassandra,
    Hbase
    MongoDB
    ETL
    Workflow
    Hadoop
    Cloudera
    HortonWorks
    MapR
  • “Green” Containerized Data Centers
    Mobile
    Resilient (MIL-STD-810f Shock & Vibe)
    Dense ( > 9 PB, 1000s of cores, > 15KW/Rack)
    Energy Efficient (PUE ~ 1.2)
    5
    NetApp Confidential — Limited Use
  • Ethernet’s Relentless March
    6
  • So What?
    Impacts of Big Data In Large Enterprises:
    Increased tendency toward distributed repositories
    Network bandwidth will lag behind demand
    Access data where it “lives” vs. from central repository, saving time, bandwidth and money
    Metadata Management
    Distributed repositories put more strain on the metadata layer
    Shift away from traditional database / enterprise applications
    Shift toward NoSQL / MapReduce based metadata solutions
    Data Management
    Shift toward self-describing objects and object-based storage, along with POSIX-compliant & CDMI access, to overcome inode limitations
    Data Center Challenges
    Denser (> 1.8 PB), heavier (> 3000 lb), power hungry (> 15 KW) racks
    Containerized DCs more efficient cost effective than traditional DCs
    7
  • Thank You
    8
    Dale Wickizer
    Chief Technology Officer
    U.S. Public Sector, NetApp, Inc.
    Dale.Wickizer@netapp.com
    (703) 864-4805