In depth description and analysis of Wireless Lan 802.11 Tech and its impact on Networking given at Glocom in Japan August 2002. Interesting to look back and see what predictions were right on and others not so...
Presentation gives an insight into Moore's law and it's successful 50 years.
An account on what Moore's law is, how we keep pace with Moore's law, and what future holds for it is detailed out in the slides.
The document discusses the history and progression of various technologies such as cameras, cell phones, storage devices, transistors, and computers over time. It notes key developments like the first mobile phone in 1973, Moore's prediction in 1965 about transistors doubling every two years (Moore's Law), and the increasing number of transistors that can be placed on integrated circuits. The document also explores what drives this continued rapid advancement, with quotes suggesting it is fueled by economic factors and human belief that progress can be sustained.
The document discusses how nanotechnology can be used for future electronics. It notes that while microelectronics currently solve many problems, they have limitations in physical size and increasing fabrication costs. Nanotechnology allows for electronics to be developed at the nanoscale level, enabling flexible electronics, improved wireless devices, and molecular devices. Some key applications of nanotechnology include using graphene for flexible electronics, developing improved sensors and memory storage for wireless devices, and creating molecular devices that function like diodes or switches at the nanoscale level. The document concludes that nanotechnology has promise to continue chip miniaturization and enable new flexible and wearable electronic devices.
Imagine a world where everything you have doubles every two years without fail, this is the world Gordon Moore created for us in 1965. Today it quietly governs everything we do by controlling everyday items from toasters to our cars all while deflating our economy at rate which man has never seen before, but for how long can Gordon’s magic continue before Moore’s Law becomes Moore’s Wall?
Telecommunication Systems: How is Technology Change Creating New Opportunitie...Jeffrey Funk
These slides discuss how improvements in the data rates of wireline and wireless systems have and continue to occur. For wireline systems, these improvements are driven by the use of better glass fiber, lasers, amplifiers, and wavelength division multiplexing and there appears to be few limits to these improvements. For wireless systems, these improvements are primarily driven by the use of better ICs. As long as these improvements in ICs continue to occur, improvements in data rates along with improvements in the use of the frequency spectrum continue to be possible. Improvements in both wireless and wireline systems will also make new forms of Internet content possible. Furthermore, these improvements in ICs along with the improvements in MEMS that are discussed in a related set of slides are gradually making cognitive radio economically feasible. All of these improvements are creating various kinds of entrepreneurial opportunities. These slides are based on a forthcoming book entitled “Technology Change and the Rise of New Industries and they are the sixth session in a course entitled “Analyzing Hi-Tech Opportunities.”
The document discusses Wordnet-Affect, an extension of Wordnet that adds affective information. It was built by manually adding affect labels and information to 1903 terms, then projecting this information to Wordnet synsets. The extension was done semi-automatically using Wordnet relations. Possible applications of Wordnet-Affect include sentiment analysis, verbal expressiveness for conversational agents, and computer-assisted creativity. In general, affective computing aims to enable computers to perceive and express emotion and can be applied in areas like adaptive entertainment and modifying environments.
Operation management Telecom Sector in India Saurabh Tiwari
The document discusses the liberalization and growth of India's telecom sector over the past few decades. It outlines key policies and initiatives that opened the sector to private participation, including the National Telecom Policy of 1994 and 1999. The Telecom Regulatory Authority of India was established in 1997 to regulate tariffs and promote fair competition. National long distance and international long distance services were opened to private operators in 2000 and 2002, respectively, accelerating expansion of telecom infrastructure across India.
In depth description and analysis of Wireless Lan 802.11 Tech and its impact on Networking given at Glocom in Japan August 2002. Interesting to look back and see what predictions were right on and others not so...
Presentation gives an insight into Moore's law and it's successful 50 years.
An account on what Moore's law is, how we keep pace with Moore's law, and what future holds for it is detailed out in the slides.
The document discusses the history and progression of various technologies such as cameras, cell phones, storage devices, transistors, and computers over time. It notes key developments like the first mobile phone in 1973, Moore's prediction in 1965 about transistors doubling every two years (Moore's Law), and the increasing number of transistors that can be placed on integrated circuits. The document also explores what drives this continued rapid advancement, with quotes suggesting it is fueled by economic factors and human belief that progress can be sustained.
The document discusses how nanotechnology can be used for future electronics. It notes that while microelectronics currently solve many problems, they have limitations in physical size and increasing fabrication costs. Nanotechnology allows for electronics to be developed at the nanoscale level, enabling flexible electronics, improved wireless devices, and molecular devices. Some key applications of nanotechnology include using graphene for flexible electronics, developing improved sensors and memory storage for wireless devices, and creating molecular devices that function like diodes or switches at the nanoscale level. The document concludes that nanotechnology has promise to continue chip miniaturization and enable new flexible and wearable electronic devices.
Imagine a world where everything you have doubles every two years without fail, this is the world Gordon Moore created for us in 1965. Today it quietly governs everything we do by controlling everyday items from toasters to our cars all while deflating our economy at rate which man has never seen before, but for how long can Gordon’s magic continue before Moore’s Law becomes Moore’s Wall?
Telecommunication Systems: How is Technology Change Creating New Opportunitie...Jeffrey Funk
These slides discuss how improvements in the data rates of wireline and wireless systems have and continue to occur. For wireline systems, these improvements are driven by the use of better glass fiber, lasers, amplifiers, and wavelength division multiplexing and there appears to be few limits to these improvements. For wireless systems, these improvements are primarily driven by the use of better ICs. As long as these improvements in ICs continue to occur, improvements in data rates along with improvements in the use of the frequency spectrum continue to be possible. Improvements in both wireless and wireline systems will also make new forms of Internet content possible. Furthermore, these improvements in ICs along with the improvements in MEMS that are discussed in a related set of slides are gradually making cognitive radio economically feasible. All of these improvements are creating various kinds of entrepreneurial opportunities. These slides are based on a forthcoming book entitled “Technology Change and the Rise of New Industries and they are the sixth session in a course entitled “Analyzing Hi-Tech Opportunities.”
The document discusses Wordnet-Affect, an extension of Wordnet that adds affective information. It was built by manually adding affect labels and information to 1903 terms, then projecting this information to Wordnet synsets. The extension was done semi-automatically using Wordnet relations. Possible applications of Wordnet-Affect include sentiment analysis, verbal expressiveness for conversational agents, and computer-assisted creativity. In general, affective computing aims to enable computers to perceive and express emotion and can be applied in areas like adaptive entertainment and modifying environments.
Operation management Telecom Sector in India Saurabh Tiwari
The document discusses the liberalization and growth of India's telecom sector over the past few decades. It outlines key policies and initiatives that opened the sector to private participation, including the National Telecom Policy of 1994 and 1999. The Telecom Regulatory Authority of India was established in 1997 to regulate tariffs and promote fair competition. National long distance and international long distance services were opened to private operators in 2000 and 2002, respectively, accelerating expansion of telecom infrastructure across India.
Augmented Reality and Education - Infographic (Leapfrog Initiatives)Alfonso Sintjago
Augmented reality (AR) technology is the result of using convergence technology to greatly expand our functional reality, i.e. that which is in our expanded environment, that we know about, and that we are able to act on. The technologies making this possible include, smartphones, GPS, phone cameras, always-on and always-available data networks, etc. A crude, but widespread, example of AR technology is Google Maps. Google Maps makes it possible for us to know about places of interest that are in our environment, or an environment relevant to us, far beyond that which our biological senses can reveal. Other examples include the services provided by apps such as Layars and Wikitude that provide detailed data on objects that we experience, Google’s recently released AR game, Ingress, among many other projects. AR is likely to be one of the most transformative technological developments that we will see over the next decade because it radically changes the world that we live in, how we perceive that world, and how we interact with it and other individuals in it. With affordable HUD displays (ex. Google Glass) this technology will take on a whole other dimension.
Multimedia presentation to board of education on augmented reality for diffus...Jeannie Frazier
The document proposes implementing augmented reality in classrooms and provides supporting research. It discusses how AR can engage students by bringing lessons to life through virtual overlays of real objects. Research shows AR helps learning and several developers have created educational AR modules. While the technology faces challenges, early adopter teachers are using AR successfully. The presentation recommends providing teacher training and resources to help AR gain acceptance and improve education.
1. Wireless communication has existed for millennia using smoke signals, light signals, and flags but long-distance wireless communication was not possible until the 19th century.
2. In the late 19th century, scientific discoveries by Maxwell, Hertz, Tesla, and Marconi laid the foundation for modern radio-based wireless communication using electromagnetic waves.
3. The 20th century saw rapid advances including the first radio broadcasts, development of cell phone networks beginning in the late 1940s-1950s, and the introduction of digital cellular standards like GSM in the 1980s-1990s that enabled international roaming.
Fiber optic systems are important telecommunication
infrastructure for world-wide broadband networks. Wide
bandwidth signal transmission with low delay is a key
requirement in present day applications. Optical fibers provide
enormous and unsurpassed transmission bandwidth with
negligible latency, and are now the transmission medium of
choice for long distance and high data rate transmission in
telecommunication networks.
A bit about Augmented Reality http://k3hamilton.com/AR/
Based on a presentation given on May 27, 2010 by Karen Hamilton and Jorge Olenenwa
Website has moved to http://k3hamilton.com/AR/ due to closing of wikispaces
The document discusses augmented reality (AR), how it differs from virtual reality and RFID, common uses of AR, and examples of AR architectures. It provides an example of how AR could be used in an automated car parking system to improve security and identification. The document outlines advantages of AR such as improved performance and accuracy, as well as disadvantages like security and interoperability issues. It concludes that AR provides a new way of interacting with user interfaces and will likely be used more widely in the future.
3 Things Every Sales Team Needs to Be Thinking About in 2017Drift
Thinking about your sales team's goals for 2017? Drift's VP of Sales shares 3 things you can do to improve conversion rates and drive more revenue.
Read the full story on the Drift blog here: http://blog.drift.com/sales-team-tips
How to Become a Thought Leader in Your NicheLeslie Samuel
Are bloggers thought leaders? Here are some tips on how you can become one. Provide great value, put awesome content out there on a regular basis, and help others.
This document discusses VLSI (Very Large Scale Integration) technology and Moore's Law. It covers key topics like transistor scaling, breakthroughs in transistor size and wafer size, challenges in VLSI design, and examples of integrated circuit cost metrics from 1994. Moore's Law, which states that the number of transistors on a chip doubles every 18 months, is explained. The scaling of features sizes over time and its impact on improving chip performance and reducing costs is also summarized.
This document provides an overview of the history and development of nanotechnology. It discusses key early developments like Richard Feynman's 1959 talk envisioning atomic engineering. It also covers the invention of the scanning tunneling microscope in 1981 and the discovery of buckyballs in 1985. The document then discusses the growth of the nanotechnology industry and funding increases over time. It provides examples of potential applications of nanotechnology and how properties become size-dependent at the nanoscale. Finally, it defines integrated circuits and microelectromechanical systems to provide context around miniaturization.
A presentation about nanoelectronics-what it is and why it is used widely nowadays, its advantages and industrial applications and the future use. Also describes some problems faced by nanoelectronics.
The document discusses the history and development of electronics engineering and technology from vacuum tubes to modern integrated circuits and systems on a chip. Key developments include the invention of the transistor in 1947, the first integrated circuit in 1958, and advancements in microprocessors, memory, and computing power that have enabled technologies like smartphones, smart homes, wireless communications, and consumer electronics. These rapid advances have revolutionized many aspects of modern life.
Electronic manufacturing and the integrated circuitBenediktusMadika1
The document discusses the history and development of electronic manufacturing and integrated circuits. It begins with the invention of the transistor in 1947 and its replacement of germanium with silicon. The integrated circuit was developed in 1958, combining multiple transistors on a single silicon chip. Gordon Moore observed that the number of transistors on integrated circuits doubled every year (Moore's Law). Key developments included the microprocessor in 1971 and dynamic random access memory. The document also describes the various levels of interconnection and packaging used to produce electronic devices and circuits.
Unit – 1 discusses the historical background of low power requirements in integrated circuits. As technology has scaled down, power dissipation has become a major concern due to higher integration densities and leakage currents. There are two main sources of power dissipation - dynamic power which is consumed during switching activity, and static power which is consumed even when no switching is occurring. Dynamic power has three components - switching power due to charging/discharging of capacitances, short circuit power due to direct paths between supply rails during switching, and glitching power due to unnecessary transitions. Low power design aims to reduce both dynamic and static power consumption.
The document discusses the history and development of VLSI (Very Large Scale Integration) technology and Moore's Law over time. It describes how transistors have gotten smaller through scaling, allowing more to fit on chips. This doubling of transistors every couple years is known as Moore's Law. 3D VLSI is presented as a potential solution to continue following Moore's Law by building chips in three dimensions rather than just two. Key challenges of 3D integration are also outlined.
The document discusses the history and development of nanotechnology. It begins with key early milestones like Richard Feynman's 1959 talk envisioning atomic engineering and the 1981 invention of the scanning tunneling microscope. The 1985 discovery of buckyballs also represented an important early discovery. The document then discusses topics like integrated circuits, nanotechnology funding levels, applications of nanotechnology, and how properties change at the nanoscale.
Roberto Siagri presented at Eurotech's 45th Annual Meeting on accelerating technological change. He discussed how Moore's Law and human ingenuity have led to exponential increases in computing power over decades. Eurotech's strategy is to provide platforms that reduce customers' total cost of ownership through scalable software over scalable hardware. Siagri argued that emerging technologies like pervasive computing and the Internet of Things will continue advancing and becoming indistinguishable from everyday life through innovation.
Vacuum tubes were the first electronic switches used in computers but were large, inefficient, and prone to failures. The invention of the transistor in 1947 revolutionized digital electronics by providing a reliable and efficient alternative. Integrated circuits enabled even smaller, cheaper, and more powerful devices by incorporating multiple transistors onto a single chip. The first microprocessor in 1971 was a milestone that led to microprocessors now containing billions of transistors and powering most electronic devices today. Digital signal processors are specialized microprocessors designed to efficiently process digital audio, video, and other signals.
Nanotechnology involves engineering materials at the nanoscale, around 1 to 100 nanometers. Richard Feynman is considered the father of nanotechnology. Nanotechnology has applications in many fields including electronics, computing, medicine, cosmetics, foods, the military, and energy. By 2020, products with nanotechnology components could be worth $1 trillion. Materials behave differently at the nanoscale compared to larger scales due to statistical mechanics and quantum effects. Nanotechnology is approached through top-down methods like lithography or bottom-up methods like self-assembly.
This document provides an overview of VLSI (Very Large Scale Integration) and its applications. It discusses the history of integrated circuits from their inception in the late 1940s to today's advanced nanoscale technologies. Key topics covered include Moore's law of transistor scaling, digital circuit design challenges, CMOS fabrication processes, and examples of how VLSI is used in various electronic systems and devices.
This document provides an overview of nanotechnology, including its definition, history, current applications, and future potential. It defines nanotechnology as the manipulation of matter at the nanoscale (1 billionth of a meter) to create new materials and devices. Some key points:
1) Nanotechnology is inspired by structures found in nature and was pioneered in the 1950s. Current applications include graphene for electronics, organic solar cells, printed electronic displays, and molecular robots for medical applications.
2) Future applications could include ultra-strong lightweight materials for construction, self-cleaning adaptive buildings, highly efficient solar energy, early disease detection chips, artificial organs produced with nanomedicine, and technologies to reverse climate change
Augmented Reality and Education - Infographic (Leapfrog Initiatives)Alfonso Sintjago
Augmented reality (AR) technology is the result of using convergence technology to greatly expand our functional reality, i.e. that which is in our expanded environment, that we know about, and that we are able to act on. The technologies making this possible include, smartphones, GPS, phone cameras, always-on and always-available data networks, etc. A crude, but widespread, example of AR technology is Google Maps. Google Maps makes it possible for us to know about places of interest that are in our environment, or an environment relevant to us, far beyond that which our biological senses can reveal. Other examples include the services provided by apps such as Layars and Wikitude that provide detailed data on objects that we experience, Google’s recently released AR game, Ingress, among many other projects. AR is likely to be one of the most transformative technological developments that we will see over the next decade because it radically changes the world that we live in, how we perceive that world, and how we interact with it and other individuals in it. With affordable HUD displays (ex. Google Glass) this technology will take on a whole other dimension.
Multimedia presentation to board of education on augmented reality for diffus...Jeannie Frazier
The document proposes implementing augmented reality in classrooms and provides supporting research. It discusses how AR can engage students by bringing lessons to life through virtual overlays of real objects. Research shows AR helps learning and several developers have created educational AR modules. While the technology faces challenges, early adopter teachers are using AR successfully. The presentation recommends providing teacher training and resources to help AR gain acceptance and improve education.
1. Wireless communication has existed for millennia using smoke signals, light signals, and flags but long-distance wireless communication was not possible until the 19th century.
2. In the late 19th century, scientific discoveries by Maxwell, Hertz, Tesla, and Marconi laid the foundation for modern radio-based wireless communication using electromagnetic waves.
3. The 20th century saw rapid advances including the first radio broadcasts, development of cell phone networks beginning in the late 1940s-1950s, and the introduction of digital cellular standards like GSM in the 1980s-1990s that enabled international roaming.
Fiber optic systems are important telecommunication
infrastructure for world-wide broadband networks. Wide
bandwidth signal transmission with low delay is a key
requirement in present day applications. Optical fibers provide
enormous and unsurpassed transmission bandwidth with
negligible latency, and are now the transmission medium of
choice for long distance and high data rate transmission in
telecommunication networks.
A bit about Augmented Reality http://k3hamilton.com/AR/
Based on a presentation given on May 27, 2010 by Karen Hamilton and Jorge Olenenwa
Website has moved to http://k3hamilton.com/AR/ due to closing of wikispaces
The document discusses augmented reality (AR), how it differs from virtual reality and RFID, common uses of AR, and examples of AR architectures. It provides an example of how AR could be used in an automated car parking system to improve security and identification. The document outlines advantages of AR such as improved performance and accuracy, as well as disadvantages like security and interoperability issues. It concludes that AR provides a new way of interacting with user interfaces and will likely be used more widely in the future.
3 Things Every Sales Team Needs to Be Thinking About in 2017Drift
Thinking about your sales team's goals for 2017? Drift's VP of Sales shares 3 things you can do to improve conversion rates and drive more revenue.
Read the full story on the Drift blog here: http://blog.drift.com/sales-team-tips
How to Become a Thought Leader in Your NicheLeslie Samuel
Are bloggers thought leaders? Here are some tips on how you can become one. Provide great value, put awesome content out there on a regular basis, and help others.
This document discusses VLSI (Very Large Scale Integration) technology and Moore's Law. It covers key topics like transistor scaling, breakthroughs in transistor size and wafer size, challenges in VLSI design, and examples of integrated circuit cost metrics from 1994. Moore's Law, which states that the number of transistors on a chip doubles every 18 months, is explained. The scaling of features sizes over time and its impact on improving chip performance and reducing costs is also summarized.
This document provides an overview of the history and development of nanotechnology. It discusses key early developments like Richard Feynman's 1959 talk envisioning atomic engineering. It also covers the invention of the scanning tunneling microscope in 1981 and the discovery of buckyballs in 1985. The document then discusses the growth of the nanotechnology industry and funding increases over time. It provides examples of potential applications of nanotechnology and how properties become size-dependent at the nanoscale. Finally, it defines integrated circuits and microelectromechanical systems to provide context around miniaturization.
A presentation about nanoelectronics-what it is and why it is used widely nowadays, its advantages and industrial applications and the future use. Also describes some problems faced by nanoelectronics.
The document discusses the history and development of electronics engineering and technology from vacuum tubes to modern integrated circuits and systems on a chip. Key developments include the invention of the transistor in 1947, the first integrated circuit in 1958, and advancements in microprocessors, memory, and computing power that have enabled technologies like smartphones, smart homes, wireless communications, and consumer electronics. These rapid advances have revolutionized many aspects of modern life.
Electronic manufacturing and the integrated circuitBenediktusMadika1
The document discusses the history and development of electronic manufacturing and integrated circuits. It begins with the invention of the transistor in 1947 and its replacement of germanium with silicon. The integrated circuit was developed in 1958, combining multiple transistors on a single silicon chip. Gordon Moore observed that the number of transistors on integrated circuits doubled every year (Moore's Law). Key developments included the microprocessor in 1971 and dynamic random access memory. The document also describes the various levels of interconnection and packaging used to produce electronic devices and circuits.
Unit – 1 discusses the historical background of low power requirements in integrated circuits. As technology has scaled down, power dissipation has become a major concern due to higher integration densities and leakage currents. There are two main sources of power dissipation - dynamic power which is consumed during switching activity, and static power which is consumed even when no switching is occurring. Dynamic power has three components - switching power due to charging/discharging of capacitances, short circuit power due to direct paths between supply rails during switching, and glitching power due to unnecessary transitions. Low power design aims to reduce both dynamic and static power consumption.
The document discusses the history and development of VLSI (Very Large Scale Integration) technology and Moore's Law over time. It describes how transistors have gotten smaller through scaling, allowing more to fit on chips. This doubling of transistors every couple years is known as Moore's Law. 3D VLSI is presented as a potential solution to continue following Moore's Law by building chips in three dimensions rather than just two. Key challenges of 3D integration are also outlined.
The document discusses the history and development of nanotechnology. It begins with key early milestones like Richard Feynman's 1959 talk envisioning atomic engineering and the 1981 invention of the scanning tunneling microscope. The 1985 discovery of buckyballs also represented an important early discovery. The document then discusses topics like integrated circuits, nanotechnology funding levels, applications of nanotechnology, and how properties change at the nanoscale.
Roberto Siagri presented at Eurotech's 45th Annual Meeting on accelerating technological change. He discussed how Moore's Law and human ingenuity have led to exponential increases in computing power over decades. Eurotech's strategy is to provide platforms that reduce customers' total cost of ownership through scalable software over scalable hardware. Siagri argued that emerging technologies like pervasive computing and the Internet of Things will continue advancing and becoming indistinguishable from everyday life through innovation.
Vacuum tubes were the first electronic switches used in computers but were large, inefficient, and prone to failures. The invention of the transistor in 1947 revolutionized digital electronics by providing a reliable and efficient alternative. Integrated circuits enabled even smaller, cheaper, and more powerful devices by incorporating multiple transistors onto a single chip. The first microprocessor in 1971 was a milestone that led to microprocessors now containing billions of transistors and powering most electronic devices today. Digital signal processors are specialized microprocessors designed to efficiently process digital audio, video, and other signals.
Nanotechnology involves engineering materials at the nanoscale, around 1 to 100 nanometers. Richard Feynman is considered the father of nanotechnology. Nanotechnology has applications in many fields including electronics, computing, medicine, cosmetics, foods, the military, and energy. By 2020, products with nanotechnology components could be worth $1 trillion. Materials behave differently at the nanoscale compared to larger scales due to statistical mechanics and quantum effects. Nanotechnology is approached through top-down methods like lithography or bottom-up methods like self-assembly.
This document provides an overview of VLSI (Very Large Scale Integration) and its applications. It discusses the history of integrated circuits from their inception in the late 1940s to today's advanced nanoscale technologies. Key topics covered include Moore's law of transistor scaling, digital circuit design challenges, CMOS fabrication processes, and examples of how VLSI is used in various electronic systems and devices.
This document provides an overview of nanotechnology, including its definition, history, current applications, and future potential. It defines nanotechnology as the manipulation of matter at the nanoscale (1 billionth of a meter) to create new materials and devices. Some key points:
1) Nanotechnology is inspired by structures found in nature and was pioneered in the 1950s. Current applications include graphene for electronics, organic solar cells, printed electronic displays, and molecular robots for medical applications.
2) Future applications could include ultra-strong lightweight materials for construction, self-cleaning adaptive buildings, highly efficient solar energy, early disease detection chips, artificial organs produced with nanomedicine, and technologies to reverse climate change
This document provides an introduction to the VLSI Design course EEL3320. It discusses the evolution of integrated circuits from early computers to modern microprocessors containing billions of transistors. The course will cover CMOS device operation, circuit design, sequential elements, and design methodologies. Students will learn how to design and optimize digital circuits for cost, speed, power, and reliability. The document outlines the course content, textbooks, evaluation criteria, and instructor details.
The document discusses the history and structure of MOSFET transistors. It begins with the conception of transistors in 1947 and integrated circuits in 1958. MOSFETs became important building blocks as they require almost zero control current when idle. The document outlines the development of MOSFET logic gates in 1963, and the scaling of integration from SSI to VLSI. It describes full custom and gate array ASIC design approaches using prefabricated cells.
This PDF tells the basic Concept of ICs (Integrated Circuit) in Embedded System . This pdf also contain some examples including application of ICs in Solar Panel .
Nanotechnology involves manipulating materials at the atomic or molecular scale to create structures between 1 to 100 nanometers in size and endow them with new properties. It has applications in electronics, medicine, energy, and other fields. Some key points are that it allows integrating biology, chemistry and engineering to build devices with unprecedented properties. Challenges include developing new research tools and cleanroom infrastructure since nanoscale work requires precision. The future of nanotechnology may include flexible electronics, molecular devices, and applications like enhanced health monitoring and energy storage.
Nanotechnology involves manipulating materials at the atomic or molecular scale to create structures between 1 to 100 nanometers in size and endow them with new properties. It has applications in electronics, medicine, energy, and other fields. Some key points are that it allows integrating biology, chemistry and engineering to build devices with unprecedented properties. Challenges include developing new research tools and cleanroom infrastructure since nanoscale work requires precision. The future of nanotechnology may include flexible electronics, molecular devices, and applications like enhanced medical diagnostics and treatment.
This document discusses the evolution of transistor technology, focusing on the development of tri-gate transistors. It describes how tri-gate transistors allow Moore's Law to continue by addressing issues like short channel effects that arise at smaller scales. The document outlines different types of multigate transistors and explains the advantages of tri-gate transistors, such as reduced power dissipation and better control over leakage current. It also provides some examples of how small 22nm transistors are in comparison to everyday objects.
This document discusses the five generations of computers from 1940 to present. The first generation used vacuum tubes and were very large, expensive, and unreliable. The second generation used transistors, which made computers smaller, cheaper, and more reliable. The third generation used integrated circuits, making computers even smaller and faster. The fourth generation used microprocessors and VLSI circuits, leading to the development of personal computers. The fifth generation, the present time, uses ULSI technology and artificial intelligence for parallel processing and true artificial intelligence.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Infrastructure Challenges in Scaling RAG with Custom AI modelsZilliz
Building Retrieval-Augmented Generation (RAG) systems with open-source and custom AI models is a complex task. This talk explores the challenges in productionizing RAG systems, including retrieval performance, response synthesis, and evaluation. We’ll discuss how to leverage open-source models like text embeddings, language models, and custom fine-tuned models to enhance RAG performance. Additionally, we’ll cover how BentoML can help orchestrate and scale these AI components efficiently, ensuring seamless deployment and management of RAG systems in the cloud.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
AI 101: An Introduction to the Basics and Impact of Artificial IntelligenceIndexBug
Imagine a world where machines not only perform tasks but also learn, adapt, and make decisions. This is the promise of Artificial Intelligence (AI), a technology that's not just enhancing our lives but revolutionizing entire industries.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
2. The Telecom Infrastructure
• Telecommunications is the most critical
infrastructure of the 21st century.
• Investment in telecom is more productive than
investment in other kinds of infrastructure.
• World Bank research* shows a large productivity
benefit to investment in telecom—larger than
investments in roads, electricity or even education!
*Canning, David. Infrastructure’s Contribution to Aggregate Output, The World
Bank Policy Research Working Papers Series # 2246, 30 Nov 1999.
http://ideas.repec.org/p/wbk/wbrwps/2246.html 2
3. The Telecom Infrastructure
• And of course, the impact is exceptionally
noticeable in developing nations.
• More recent research* based on the spread
of mobile phones finds telecommunication’s
impact on economic growth is twice as large
in developing nations as in developed
nations.
*Waverman, Leonard; Meschi, Meloria and Fuss, Melvyn. The
Impact of Telecoms on Economic Growth in Developing
Countries, Vodafone Policy Paper Series, 2, March 2005,
pp. 10-23. 3
4. The Telecom Economy
• According to the new industry market study,
telecommunications services revenue
worldwide will grow from $2.1 trillion in 2014
to $2.4 trillion in 2019 at a combined average
growth rate of 2.1 percent.
4
6. Measurements of the Digital World
• Processors
– Processing power is measured in number of transistors
and operations per second
– The size of the features (the elements that make up the
structures on a chip) is today measured in nanometers.
• Digital storage
– Measured in Bytes (B)
• 1 bit = a single “1” or “0”
• 1 Byte = 8 bits = single character (letter “A”, number “2”, etc)
• Digital transmission rate
– Measured in bits per second (bps) 6
7. 7
Key Measurements of the
Digital Universe
• Kilo 103
Thousand 1,000
• Mega 106
Million 1,000 Kilo
• Giga 109
Billion 1,000 Mega
• Tera 1012
Trillion 1,000 Giga
• Peta 1015
Million Billion 1,000 Tera
• Exa 1018
Billion Billion 1,000 Peta
• Zetta 1021
Billion Trillion 1,000 Exa
• Yotta 1024
Trillion Trillion 1,000 Zetta
7
8. How Much Data Is That?*
• 2KB = one typewritten page
• 100KB = a small, low-resolution image
• 1MB = one digital chest x-ray, or one minute of
high-fidelity sound
• 5 MB = the complete works of Shakespeare, or
30 seconds of broadcast quality video
• 50 MB = one digital mammogram
• 100 MB = two encyclopedia volumes, or 3 feet
of books on a shelf
• 500 MB = one CD-ROM 8
9. How Much Data Is That?*
• 1 GB = one pickup truck full of books, or
symphony in high-fidelity sound, or one
broadcast movie
• 1 TB = all the X-ray films in a large
technological hospital, or 50,000 trees made
into paper and printed, or daily rate of EOS
(Earth Orbiting System) data (as of 1998)
• 20 TB = Printed collection of the U. S. Library
of Congress (LOC)
9
10. 10
How Much Data Is That?*
• 2PB = all U. S. academic research libraries
• 200PB = all printed material
• 1EB = one billion pickup trucks full of books, or one
trillion 400 page books
• 5EB = 37,000 Libraries of Congress, or about 30 feet
of books for each of the 6.3 billion people on earth
• 161EB = 12 stacks of books reaching from the earth
to the sun
• 1ZB = 250 million DVDs
Source: Source of examples includes “James S. Huggins' Refrigerator Door” at
http://www.jamesshuggins.com/h/tek1/how_big.htm and Seattle Discover Institute at
http://www.discovery.org/ 10
11. The Power of the Processor
• The current environment is generally defined by
Moore’s Law, realizing a doubling of processing power
every 18 months
• Moore’s Law is not infinite – experts now say silicon
chips have no more than four years of further
miniaturization left
• What solutions are being considered?
– Silicon photonics
– Carbon nanotubes
– Superconductors, or quantum computing
– Graphene
– DNA computing 11
12. The Power of the Processor
• The original transistor built by Bell Labs in
1947 could be held in your hand, while
hundreds of Intel’s new 45nm transistors can
fit on the surface of a single red blood cell.
– Transistors are the tiny switches that process the
ones and zeroes that make up our digital world.
• This very success is bringing chipmakers to
the brink of a new, steep obstacle to further
gains in performance.
– The problem lies in the tiny metal wires that
weave the transistors into integrated circuits. 12
13. Transistors Then And Today
13
Bell Labs Transistor - 1947
Intel Microchip – 2008
800 Million transistors
14. The Power of the Processor
• The Intel 130-nm
Pentium 4 processor
has
– 55 million transistors and
– uses roughly 4.8 km of
interconnects to support
each square centimeter
of its circuitry
• The Intel 90-nm
processor has
– 110 million transistors
and 6.9 km of
interconnects per square
centimeter
14
• A 65-nm processor has
- 1 billion transistors
and 11 km of interconnects per
square centimeter, for a total
Interconnect length of 34.6k m!
15. The Awful Truth About Moore’s Law
• Intel’s new generation processors are 45nm and 32nm.
– Intel's 45nm technology packs more than 400 million transistors
for dual-core processors and more than 800 million for quad-
core into the same silicon space.
– You could fit more than 2,000 45nm transistor gates, laid side
by side, across the width of a human hair.
– More than 2 million 45nm transistors could fit on a period at the
end of a sentence (estimated to be 1/10 square millimeter in
area).
– If a house shrunk at the same pace transistors have, you would
not be able to see a house without a microscope. To see the
45nm transistor, you need a very advanced microscope.
15
16. The Awful Truth About Moore’s Law
• 45-nm technology is used in the Intel® Core™2
processor family code named "Penryn,“.
– The price of a transistor in Penryn processors is about 1
millionth the average price of a transistor in 1968.
– If car prices had fallen at the same rate, a new car today
would cost about 1 cent.
• Next, Intel's 32nm chips and next-generation
“Nehalem” microprocessor architecture in 2009.
• Houses more than 1.9 billion transistors.
• 4 million 32nm transistors would fit on a period at the end
of a sentence. (1/10 square millimeter in area)
• 22nm chips by 2011, followed by 17nm . 1616
17. The Awful Truth About Moore’s Law
Informative Videos
• What is Moore’s Law
– http://www.youtube.com/watch?v=bLSMn0cNWAw
• Moore’s Law No More (with Gordon Moore)
– http://news.cnet.com/1606-2_3-29244.html
• Moore’s Law Got Me
– http://www.metacafe.com/watch/1151291/moores_law_got_
me/
• Moore’s Law Ends in 2020
– http://www.youtube.com/watch?v=5oLsMs5_19Y
• The Dark Secret Of Hendrik Schon: Moore's law
– http://science.discovery.com/videos/the-dark-secret-of-
hendrik-schon-moores-law.html 17
18. The Awful Truth About Moore’s Law
• Each new generation of chips only makes the
situation worse.
• The narrower the wire, the longer it takes a
signal to propagate along it.
• Today’s most advanced ICs, switch up to 10
billion times a second, and their metal
interconnects can barely keep up.
• The good news is that the industry is working
on solutions
• Two main approaches are evolving
– New low-K and Hi-K dielectrics
– Optical connections
18
19. Low-K Dielectrics
• Low-k dielectrics changes the propagation
characteristics of the tiny on-chip
transmission lines, or metal interconnects.
• Low-k reduces the interconnect capacitance
and wiring delay by changing the material
that insulates it from the surrounding silicon
chip as a well as from neighboring wire.
• The application for low-K is in the metal
interconnects connecting all the transistors
19
20. Low-K Dielectrics
• The bad news is that the low-k films are
extremely difficult to integrate into the
manufacturing – they are soft, weak, and adhere
poorly to both the silicon and the metal wire.
They also crack and delaminate easily.
• Today’s 65-nm and the next generation of 45-nm
chips require even lower-k materials.
– The addition of ultra-low-k interconnect provides a 15
per cent reduction in wiring-related delay as
compared to conventional low-k dielectrics. 20
21. Hi-K Dielectrics
• As elements in the chip were being reduced
to 45 nanometers, the gate dielectric began
to lose its insulating (dielectric) quality and
exhibited too much leakage.
• The new, higher-K dielectrics introduce a
more power-saving integrated circuit,
allowing even cooler and faster processors at
45 nm and beyond, as well as more circuits
per wafer area. 21
22. High-K Dielectrics
• Higher-K dielectrics are combined with metal
gates for increased performance gains.
• The new Hi-K dielectrics are used as the
transistor gate dielectric material.
• Intel is heralding this technological
breakthrough as the “Biggest Change to
Computer Chips In 40 Years!”
– http://www.intel.com/technology/architecture-silicon/45nm-
core2/demo/index.htm?iid=tech_arch_45nm+body_demo
22
23. Silicon Photonics
• The more significant development comes
from Intel’s successful demonstration of the
first continuous all-silicon laser.
– http://download.intel.com/technology/silicon/sp/download/Analyst_028.htm
• Silicon photonics have the potential to greatly
extend the lifetime of Moore’s Law.
– Optical connections can carry thousands of times
more data per second than copper wires can.
• Unfortunately, existing optical components
are far too expensive for use in individual
computers or even local networks. 2323
25. Other Alternatives To Extend
Moore’s Law
• Carbon nanotubes
– Tubes of pure carbon, about the width of a typical protein
molecule, also happen to conduct electricity, and could be
used as tiny molecular-scale wires for making electronic
circuitry. Unfortunately, they also cost about $500 a
gram.
• Superconductors, or materials that conduct
electricity with zero electrical resistance.
– New ways to harness the power of quantum qubits to
boost computing power, but there are very practical
difficulties in building a quantum computer
• Graphene - a new material exactly one-atom thick has
been used to create a one nanometer transistor. 25
26. Moore’s Law and
Telecommunications
• On the positive side, thanks to Moore’s Law,
network endpoints today are small, powerful,
inexpensive devices.
• With such power in the endpoints, the need to
embed the functions of a network in the network’s
core shrink.
• In addition, smart end devices can set up and
manage calls far better than a centralized network.
– In fact, when voice is implemented in end devices, the
ability to mix it into other kinds of interactions emerges
• online game play, collaboration, mutual web surfing and many
more yet to be discovered
• the idea of a “call” as a special, discrete event could well
disappear. 26
27. 2727
Processors and Applications
• The latest processors support fast visualization of
large data sets, and intensive math for real-time
simulations
• Applications include digital entertainment, 3-D
games, graphics, astronomy, biosciences, and
predictive modeling
• Simulation of blood flow in the human body
• Space weather modeling
• Virtual tests for therapeutic cancer drugs
• Global modeling of the Earth’s magnetosphere
• Simulations of shock waves and eddies in turbulent fluids
• Large-scale structure of galaxies and galaxy clusters
• Modeling the interaction of proteins within individual cells
• Studying instability and turbulence in plasmas
• Testing models of the formation of cosmological structures
28. Storage Dimensions
• 2 KB (16,000 bits)
– One typewritten page
• 1 MB (Megabyte) can store
– One long novel, stored as text
– One full-page black-and-white image
– One 3 x 5 inch color picture
– 2 minutes of telephone-quality sound
– 7 seconds of CD-quality sound
– 0.04 seconds of broadcast-quality video
• 4 GB (Gigabyte)
– One feature-length film
– High-definition movies use 6x the storage space of traditional
movies.
• 1TB (Terabyte) 28
29. Optical Storage Media
• 1st
Generation
– CD-ROM capacity = 650 MB
• 2nd
Generation
– DVD capacity = 4.6 GB to 17 GB
– Transfer rates range from 600 Kbps to 1.3 Mbps
• 3rd
Generation
– Blu-ray optical disc = 25GB (single-layer) or
50GB (dual-layer)
– Transfer rates range from 36Mbps to 432 Mbps 29
30. Optical Storage Media
• Alternative Disc Technologies
– Holographic Versatile Disc (HVD)
• Based on OptWare collinear holography technology
• Capacity of up to 3.9TB
• Transfer rates of up to 1Gbps possible
– 3D optical data storage
• Call/Recall Inc has unveiled the first optical storage
solutions where information can be recorded and/or read
in three dimensional resolution
• Capacity of 5-10 TB
• Transfer rates of 100Mbps to 500 Mbps 30
31. Emerging Storage Technologies
• Bacterial Protein Memory
– Bacteriorhodopsin (BR)
• can store TBs per cuvette
• Molecular Memory
– Rotaxanes
• Can store 100 Gbits per inch
• Likely to remain a laboratory curiosity
• Magnetic Sensors
– high degree of sensitivity means terabits of data
can be fit into a square inch of disk space 31
32. Bacterial Protein Memory
• Bacteriorhodopsin (BR) is one of the first forms of life
on our planet - a protein grown by salt marsh bacteria
at least 2.3 billion years ago – and it is is likely to
become the wave of the future in computer data
storage and manipulation.
• BR is a tiny, rugged protein that has improved through
billions of years of evolution to become extremely
efficient at converting light into energy.
• As a biological substance, the protein also enables
data to be stored in three dimensions, just like the
human brain.
• Expected to reach TBs 32
34. Rotaxanes
• A memory subsystem that uses molecules to
store digital bits.
• A vital piece of nanoelectronic circuitry has
been produced which could bring molecular
computers a step closer.
• The device is the size of a human white
blood cell - dimensions that its solid-state
equivalent are not expected to attain before
the year 2020.
34
35. Rotaxanes
• Teams of scientists from Caltech and UCLA have
made a molecular electronic device that mimics
dynamic random access memory (DRAM) circuits
on today's computer microchips.
• The complete electronic memory circuit contained
160,000 bits at a density of 1011
bits cm-2
.
• A single bit is only 15 nanometres wide, or about
one ten-thousandth the diameter of a human hair.
• By contrast, the most dense memory devices
currently available are approximately 140
nanometres in width.
3535
36. Rotaxanes
• New advancements have been scored by
Chinese Academy of Sciences (CAS)
scientists on ultrahigh-density information
storage as they successfully carried out the
reversible, erasable and rewritable
nanorecording on H2 thin films of rotaxane, a
superamolecular structure of dumbbell-like
molecules trapped within the cavity of
macrocycles.
3636
37. Magnetic Sensors
• Tiny magnetic sensors, microscopic whiskers
of nickel only a few atoms wide are capable
of detecting extremely weak magnetic fields.
• The high degree of sensitivity means terabits
of data -- or trillions of bits -- could be
crammed into a square inch of disk space.
37
38. 38
Storage vs Content
• In 2007, the amount of information created
surpassed the storage capacity available.
• By 2011, almost 50% of the digital universe
will not have a “permanent” home.
• The information created in 2011 will be
contained in more than 20 quadrillion (20
million billion) electronic information
containers – files, images, packets, tags, etc.
38
39. 39
Bandwidth Definition
• The term itself comes from the radio realm,
and the visualization of the electromagnetic
spectrum, where the spectrum is divided into
“bands”.
• The bands, and the channels within them,
have a “width” expressed in Hertz (cycles per
second).
• The wider the band, the more information it
can .
• Information transfer rate is expressed in bits
per second (bps).
41. Transfer Rate Examples
Document 2400 bps 56 Kbps 1.5Mbps 1.7Gbps
Page 8 sec 0.34 sec 0.013 sec 1.13x10-5
sec
Report 4 min 10.3 sec 0.38 sec 3.39X10-4
sec
Book 0.67 hr 1.7 min 3.84 sec .0034 sec
Dictionary 2.3 days 2.38 hrs 5.3 min 0.28 sec
Encyclopedia 5 days 5.15 hrs 11.6 min 0.61 sec
Local library 7.4 yrs 116 days 4.32 days 5.49 min
College
library
74 yrs 3.17 yrs 43.2 days 0.92 hrs
Library of
Congress**
1,900 yrs 81.5 yrs 3 yrs 23.5 hrs
41
42. 42
Transfer Rate Examples
Document 10
Gbps
100
Gbps
1 Tbps 1 Pbps 1 Ebps
Library of
Congress (LCO)**
**LCO (at 20TB)
is used as a key
measure to
denote the
immensity of data
we have today
2.35 hrs 14.1 min 1.41 min 8.26 sec .826 sec
43. Bandwidth Hungry Applications
• By 2015, estimates of annual traffic in the U.S. alone
are projected to equal over 1 Zettabyte!
– Movie downloads and P2P file sharing 100EB
– Video calling and virtual windows 400EB
– Cloud computing and remote backup 50EB
– Internet video gaming and virtual worlds 200EB
– Non-Internet IPTV >100EB
– Business IP traffic 100EB
– Phone, web, email, photos, music 50EB
• The result will be an U.S. Internet that is 50x larger
than it was in 2006! 43
44. How Much is an Exabyte?
• 5EB = amount of new information created and
stored in 2002
–equal to 37,000 Libraries of Congress (LOC)
• 161EB = the amount of digital information
created and copied in 2006
–equal to 3 million times all the books ever
written!
44
45. How Much is an Exabyte?
• 988EB = amount of digital information
predicted to be created and copied in
2010
• By 2015 it is expected the traffic on the
Internet will be equal to the information
contained in 50 million Libraries of
Congress (LOC)
45
46. From Exacosm to the Zettabyte Era
• Annual global IP traffic is over 2/3 of a ZB (667
EB) in 2013.
– the economic downturn has had only the slightest
of impacts on traffic growth
• By 2013, the Internet was 4x larger than in
2009
– each month, the equivalent of 10 trillion DVDs will
flow across the Internet
46Source: Cisco Visual Networking Index – Forecast and Methodology, 2008-2013, June 9, 2009
47. Traffic in the Zettabyte Era
• Traffic growth will be driven by
–visual networking
• usage of video increases with social networking
–the widgetization of Internet and TV
• network traffic grows beyond the boundaries of PC
browsers and TV screens
–hyperconnectivity
• all things that can or should communicate through the
network will communicate through the network
47
48. Visual Traffic in the Zettabyte Era
• Cisco Telepresence 15 Mbps,symmetrical, per session 2008
• MSN Msgr Video 4 PB per month 2008
• All Radio and TV 100 PB per year 2008
• YouTube 600 PB per year 2008
• HD YouTube 12 EB per year projected
• Amateur Video 5 EB per year 2008
• HiDef Video 50 EB per year projected
• HD Movie 1 GB per movie 2008
• HD NetFlix 5.5 EB per year projected
• Massive Parallel Game 100 PB per month 1 million players
• 3D HD Video 100 EB per experience projected
• Ultra-HD 1 ZB per experience projected
• 7,680x4,320 pixels, 33 megapixels per frame, 60 fps,
• Uncompressed 2 hour movie=25 TB, w/MPEG4 compression=360 GB 48
49. 49
Visual Communications and
Embedded Networked Intelligence
• Global Telephone and Videotelephony
Traffic
– 2008 30 EB per year
– w/video 300 EB per year (estimated)
– w/hi-def video 3 ZB per year (estimated)
• Devices Connected to the Internet (Global)
– 2000 100 million
– 2015 15 billion
Source: “ Estimating the Exaflood”, Discovery Institute, January 29, 2008 ,and Intel
50. The Hyperconnected Zettacosm
• Foundation of the hyperconnected universe
– Multitasking
– Passive Networking
• Enablers of hyperconnectivity
– Digitization of content
– Growing availability of broadband access
– Expanding screen surface area and resolution
– Growth in number of network-enabled devices
– Increase in power and speed of computing
devices 50
51. The Expansion of the “Network Day”
51Cisco Visual Networking Index – Forecast and Methodology, 2008-2013 June 9, 2009
54. Growth of Traffic Generating Units
54
Cisco Visual Networking Index – Forecast and Methodology, 2008-2013 June 9, 2009
Sample household today:
2 PCs, each w/ 11 apps
2 T Vs
2 DVRs
1 gaming console
1 Internet media device
1 portable gaming device
1 MP3 player
1 Smartphone w/3 apps
1 eBook reader
This household = 35 TGUs
55. Zettacosm Enablers
• Platforms are converging - PCs, digital TVs,
game consoles, mobile devices, and
intelligent consumer devices share four
things in common
–Broadband access
–High-performance processors
–Large display screens and resolution
–A variety of human-centered input-output
devices and accessories
55
56. Zettacosm Devices
• Increasingly PCs will represent a shrinking
percentage of all broadband-enabled devices , new
entrants will include
– Set-top boxes
– IP phone screens
– Gaming devices and handheld gaming consoles
– e-book readers
– Large-screen mobile devices
– In-vehicle-GPS displays
– TelePresence screens
– Digital advertising and sales displays 56
57. Application Trends
• The changing traffic patterns are ushering in a new
genre of applications requiring next generation
networks.
– Digital entertainment
– 3D virtual reality
– Streaming media
– Visualization
– Tele-presence
– Mobile alternatives
– Sensory networks 57
58. Applications Evolution
• Industries being revolutionized by the new era of
advanced applications and enabling telecoms
infrastructure include
– Entertainment
– Advertising
– Healthcare
– Education
– Transportation
– Government
– Warfare
– along with just about every industry you can think of!
58
59. Key ICT Trends
• The Age of Intelligence
–Things that think
–Intelligent wearables
–Man-machine interactions
–Virtual reality
–Robot squads
–Sensor networks
59
60. Key ICT Trends
• The Age of Intelligence (continued)
–Transformation and the 2.0 ecosystem
–Visual Reality
–Cloud computing
–Web Mashups
–Social Networking
–Green IT
60
61. The Age of Intelligence
• The devices used to communicate with the
Internet today, including PCs, organizers,
telephones, and mobiles, present two
problems
– They are at odds with human behavior
– They are often the bottlenecks impeding the
process and progress of information exchange.
61
67. The Age of Intelligence
• Ubiquitous computing (Ubicomp)
– Also known as ambient, calm, or pervasive
computing
– Takes computers out of boxes and puts them into
ordinary everyday things around you
• The emergence of things that think, and
communicate!
– smart appliances, smart furniture, smart homes
– smart cloths, smart food, smart wrappers, smart
needles
– smart cars, smart highways
– smart materials, smart structures, smart places 67
75. Wearable Evolution
• Business professionals, general
consumers, and youths worldwide are
carrying an increasing number of
portable electronic information and
communications gadgets
• E-textiles are emerging as the more
versatile, and elegant alternative.
75
81. Future Plans for Wearables
• Flexible electronic computer displays that will
result in outfits that change images,
projections, and patterns.
• Temperature-sensitive fibers could be woven
into mood fabrics
• The military is financing research into the
ultimate camouflage – “chameleon fabrics”
with colors and patterns that would change in
response to electrical commands.
• Smart cloths will likely be powered by
photovoltaic fibers, converting light or heat
into various functions. 81
82. Wearable Sensor Networks
• The importance of intelligent wearables has to do
with shifting traffic patterns.
• The projection is that by 2017, 95% of the traffic on
networks would come from machine to machine
communications.
• Embedded devices and intelligent wearables will
require access to communications networks in
order to be of value to their users.
82
83. Man-Machine Interactions
• The realm of man-machine interactions
covers a wide range of activities, including
– affective computing
– brain-computing interfaces
– software agents
– augmented reality
– virtual reality
– the growing presence of robots
83
84. 84
Man-Machine Interactions
Communications Channels
5 Direct
Input Channels
Sight
Hearing
Touch
Smell
Taste
2 Direct
Output Channels
Language
Motion
Future Indirect
Channels
Gaze Tracking
Brain Waves
Thought
Emotion
85. Affective Computing
• Affective computing
– gives computers the capability of recording
human responses and identifying behavior
patterns.
• Wearable computers refer to
– sensors embedded in clothing to register
biological and physiological parameters,and
communicate them if appropriate.
85
89. Augmented Reality
• Augmented Reality is the field of superimposing
computer data on real images.
• With this approach, hidden information about all
types of objects can be made visible.
• Applications exist in numerous areas, including
– architecture
– building engineering
– maintenance operations
– surgical procedures
– warfare 89
93. Virtual Reality – The Future
• In a February 2008 speech, inventor and futurologist
Ray Kurzweil predicted…..
– Ordinary machines will achieve human-like intelligence in the
next 20 years.
– Computers the size of blood cells will create fully immersive
virtual realities by 2033.
– "Today you can put a pea-sized computer inside your brain, if
you have Parkinson's disease and want to replace the
biological neurons that were destroyed by the disease."
– A billion-fold increase in computing performance and
capability over the next 25 years coupled with the 100,000
fold shrinking, would lead to "blood cell-size devices. 93
94. Virtual Reality – The Future
• “These "blood cell-size devices will be able to go
inside our bodies and keep us healthy and inside
our brain and expand our intelligence".
• He said the blood cell computers would be able to
"produce full immersion virtual reality from inside
the nervous system".
• He said the games industry had to be thinking
about the future development of computing now.
• "The games industry fits in well with the
acceleration of progress; in no other industry do you
feel that more than games." 94
95. Virtual Reality – The Future
• Mr Kurzweil said "In virtual worlds we do real
romance, real learning, real business. Virtual reality
is real reality."
• "Games are the cutting edge of what is happening -
we are going to spend more of our time in virtual
reality environments.”
• "Fully emergent games is really where we want to
go. We will do most of our learning through these
massively parallel interactions."
• "Play is how we principally learn and principally
create.” 95
96. Tele-immersion Applications
• Tele-immersion – the combination of real and
virtual environments for purposes of display
or interaction.
– Tele-meetings
– Tele-training
– Collaborative engineering and design
– Medical applications
– Entertainment services
96
99. Robots – The Next Frontier
• Intelligent robot squads
–self-organizing groups of robots
–under the control of neural
networks
–eliminating the need for humans
99
100. Bomb Squad Robots
• Meet Andros-Wolverine - he is a
six-wheeled, one-armed robotic
vehicle responsible for assisting
bomb squads in defusing of all
types of explosive devices.
• Today, he can only defuse simple
pipe and letter bombs, but Sandia
National Laboratories' Intelligent
Systems and Robotics Center
(ISRC) hopes to extend these
capabilities to car bombs and, one
day, even nuclear devices.
100
105. Micromechanical Flying Insect
105
…..Our country is at war in an
unfamiliar territory, and a battle is
about to begin...However, the enemy
doesn't know that its every move is
being monitored by robotic insects
equipped with tiny cameras, flying
overhead...called micro air vehicles
(MAVs)...dime-sized flying robots..
111. Sensor Networks Defined
• A sensor network refers to ……….
a group of specialized transducers with a
communications infrastructure intended to monitor
and record conditions at diverse locations.
• A transducer is defined as
an electronic device that converts energy from one
form to another, for example thermometers, position
and pressure sensors, microphones, and antenna.
111
112. What do Sensor Networks Sense?
• Commonly monitored parameters include…
– temperature - humidity
– pressure - wind direction and speed
– illumination intensity - vibration intensity
– sound intensity - power-line voltage
– pollutant levels - chemical concentrations
– vital body functions
112
113. How Do Sensor Networks Work?
• A sensor network consists of multiple
detection stations called sensor nodes, each
of which is small, lightweight and portable.
• Every sensor node is equipped with a
– transducer
– microcomputer
– transceiver
– power source
• The power for each sensor node is derived
from the electric utility or from a battery.
113
114. Sensor Node Functions
• Each of the sensor node’s components has a
unique function…..
– The transducer generates electrical signals
based on sensed physical effects and
phenomena.
– The microcomputer processes and stores the
sensor output.
– The transceiver, either hard-wired or wireless,
receives commands from a central computer and
transmits data to that computer.
114
115. 115
Sensor Network Diagram
* ***
* *
*
*
* *
*
* *
*
* *Mesh Node (router)
Gateway Node
Database
PC
Remote Base
Station
Sensor Nodes
*
Mobile
Tablet
Wireless Links
116. Grid Computing
• While not a new concept, by moving into
commercial markets, grid computing is
becoming the key to the future of e-business,
representing the next step in the
development of the Internet as a real-time
computing platform.
• Some 80% – 90% of processing capacity is
unused, regardless of whether it is a PC,
workstation or mainframe.
116
117. Grid Computing Categories
• Grid computing has three main
application areas
–On-demand computing grids
–Data storage grids
–Collaboration grids
117
118. Grid Computing Markets
• Life sciences
• Energy
• Manufacturing
• Financial
• Government
• Research and Development
118
119. Grid Computing Example
• Europe’s CERN nuclear research center planned to
start testing the Large Hadron Collider (LHC) in
2007. CERN has built a data grid to accomplish
this.
– This experiment involves 40 TB (terabytes) of
data per second
– Even with a reduction of data, via compression
and such, it will still generate 8 PB (petabytes) of
data per year.
– In addition to CERN, over 1,000 institutions plan
to provide storage capacity.
119
120. Computing and the Network
120
1990 – 1999
The Network Is The Computer
2000-2010
The Network Is Computing
121. Cloud Computing Defined
• It is a style of computing where IT-related
capabilities are provided “as a service”,
allowing users to access technology-enabled
services "in the cloud“,
without knowledge of,
expertise with, or control over the technology
infrastructure that supports them.
121
123. Cloud Computing Categories
• The concept generally incorporates
combinations of the following:
– Infrastructure as a Service (IaaS)
– Platform as a Service (PaaS)
– Software as a Service (SaaS)
– Other recent (ca. 2007–09) technologies that rely
on the Internet to satisfy the computing needs of
users.
123
124. Cloud Computing
• Cloud computing services often provide
common business applications online that
are accessed from a web browser, while the
software and data are stored on the servers.
• The term “cloud” is used as a metaphor for
the Internet, based on how the Internet is
depicted in computer network diagrams and
is an abstraction for the complex
infrastructure it conceals. 124
126. Web Mashups
• In web development, a mashup is a web
page or application that combines data or
functionality from two or more external
sources to create a new service.
• The term “mashup” implies easy, fast
integration, frequently using open APIs and
data sources to produce results that were not
the original reason for producing the raw
source data. 126
127. Web Mashup Categories
• There are many types of mashups, such as
– consumer mashups
– enterprise mashups
– data mashups
– business mashups
• The most common type of mashup is the
consumer mashup, aimed at the general
public.
127
128. Realtime Communications
• Realtime communications will generate
added value by reengineering and
differentiating business processes.
• Realtime communications will affect business
processes by substantially increasing the
– speed
– efficiency
– security
128
129. Realtime Communications
1st
and 2nd
Generations
• The core element of realtime communications is the
convergence of voice and data communications
based on IP (Internet Protocol).
• The first IP communication generation (1gIP) is
focused on using existing network infrastructure for
converged applications in order to cut costs.
• The second IP communication generation (2gIP)
will be primarily focused on reengineering and
differentiating business processes.
129
130. Realtime Communications Applications
• By integrating realtime communications into
IT, enterprises will enable their business
processes, creating the possibility of realtime
business.
• New productivity gains are expected from two
key applications
– Process-supporting communications
– Ad-hoc communications
130
131. Realtime Communications Benefits
• Productivity gains of these magnitudes
translate into substantial top-line revenue
enhancements
• At the same time, gains in responsiveness
can help raise brand awareness and loyalty,
reduce customer churn and extend market
reach.
131
135. 135
The New Public Network
• End-to-end digitalization
• End-to-end optical networking
• Intelligent, programmable networks
– PSTN
• distributed logic and databases
• high-speed common channel signaling, SS7
• open application program interfaces (APIs)
–IP networks
• IP Multimedia Subsystem (IMS)
• Convergence
• Networks, devices, applications
136. The New Public Network
• Network integration describes a major trend
in communication technology development.
• Driven by the market and new technologies,
PSTN is evolving to NGN and, with
transformation, the two jointly deliver high
quality, rich VoIP services, enhanced data
services, and video and multimedia services.
• Moreover, voice, data and multimedia
integration will bring users an unprecedented
digital experience 136
137. 137
The New Public Network
• Broadband infrastructure
–high bandwidth, multichannel transmission
lines
–high-speed fiber and broadband wireless
media
–low latencies
–multiservice agnostic platforms
–next generation telephony
–quality of service guarantees
–encryption and security services
138. Key Technology Shifts
• From Narrowband to Broadband
– from single channel to multichannel
– from low bandwidth to high bandwidth
• From Circuit switched to Packet switched
– from exclusive channel to shared channel
• From Data over Voice to Voice over Data
– from data over circuit-switched analog voice network to
voice over digital data packet network
• From Electronic to Optical
– shift from electronic networks to optical networking
138
139. Key Technology Shifts
• From Singlemedia to Multimedia
– from voice to multimodal communications
• From Fixed to Mobile
– from fixed wireline connections to mobile wireless
communications
• From Portable to Wearable
– from unresponsive standalone devices to affective
wearable computers
139
140. What is Converging?
• Networks Infrastructures
– PSTN, Internet, Wireless, Broadcast, Cable TV,
Corporate Back Office
• Network Services
– Local, Long Distance, Wireless, Internet, Hosting,
Applications Partnering, Security, Firewalls,
Legacy Systems Conversion, Settlement
• Devices
– Television, Telephone, Computer, Appliances,
Clothing & Jewelry, Tattoos, Neural Implants
140
141. What is Converging?
• Applications
– Communications, Information Services, Entertainment, E-
Commerce, Affective Computing, Location-based
Services
• Industries
– Biotechnology, Computing, Consumer Electronics,
Entertainment, Publishing, Power Utilities,
Telecommunications
• Man & Machine
– Artificial Limbs and Organs, Intelligent Implants, Neural
Interfaces, Artificial Life
141
Editor's Notes
[120] For example, imagine a horse ranch with sensors, each the size of a quarter, dispersed over the land, measuring environmental variables such as temperature and humidity. If each sensor were connected to a base station, the cost would be prohibitive, but in a micromesh network, they are all connected to each other and then to the base station. They synchronize with each other, collect data, and sleep until the next iteration. As this diagram illustrates, a sensor mesh network must bridge to processing elements that deal with the data collected by sensor nodes. It must also provide monitoring and command and control of the mesh in response to changing conditions. This functionality is best implemented in a gateway. As the size of a micromesh network grows, there is increasing need for a gateway-based systems-level architecture. It is predicted that gateways will become a core element of sensor-mesh networks in the coming years.