This presentation, given at AWS Michigan Meetup on 10-09-2012 provides an overview of how we used Amazon Web Services to conduct a quantitative trading system simulation on Amazon Web Services (AWS). We demonstrate an improvement in processing time of an order of magnitude and cost savings of greater than 99% compared to a traditional, in-house physical infrastructure.
Quantlogic is a leading developer of algorithmic and automated trading strategies. They have invested millions developing their own intellectual property and database of trading strategies and concepts. They provide research services to banks, hedge funds, and proprietary trading desks, including quantitative modeling, statistical analysis, and programming complex trading models. Their research capabilities are powered by a vast library of trading indicators and concepts, as well as industry-leading technology and tools.
ONEMARKETDATA and ONETICK, a company and product overviewOneMarketData, LLC
A presentation on the company and OneTick product, the leading technology for complex event processing (CEP) and tick database for the financial industry.
OneTick and the R mathematical language, a presentation from R in FinanceOneMarketData, LLC
A presentation on the use of OneTick and the R mathematical and analytical language. This session presents the integration of OneTick and R in three use-case scenarios
Serverless microservices allow building scalable and resilient applications from small, isolated services using AWS Lambda and API Gateway. Each microservice owns its own data in a decentralized data store like DynamoDB. API Gateway acts as a front door and handles authentication, authorization, and throttling. Lambda provides immutable function versions and aliases for deployments. While this makes applications highly available and scalable, it introduces challenges around transactions and data consistency. The document proposes using techniques like correlation IDs, rollback functions, DynamoDB streams, and a transaction manager to handle errors and rollbacks in a serverless environment.
How Apache Spark and Apache Hadoop are being used to keep banking regulators ...DataWorks Summit
The global financial crisis showed that traditional IT systems at banks were ill equiped to monitor and manage the daily-changing risk landscape during the global financial crisis. The sheer amount of data that needed to be crunched meant that many of the banks were day(s) behind in calculating, understanding and reporting their risk positions. Post crisis, a review by banking regulator, led the regulators to introduce a new legislation BCBS 239: Principles for effective risk data aggregation and reporting, that requires banks to meet more stringent (timeliness) requirement, in their ability to aggregate and report on their quickly-changing risk positions or risk fines to the tune of $millions. To meet these new requirements, banks have been forced to re-think their traditional IT architectures, which are unable to cope with sheer volume of risk data, and are instead turning to Apache Hadoop and Apache Spark to build out next generation of risk systems. In this talk you will discover, how some of the leading banks in the world are leveraging Apache Hadoop and Apache Spark to meet BCBS 239 regulation.
Speaker
Kunal Taneja
WSO2Con USA 2015: The Needs of Next Generation GiantsWSO2
50% of the Fortune 500 companies from 2000 are gone! Facebook just became more valuable than Walmart! Next Generation Giants are being created every day and they are standing on the shoulders of technology and new business models to get there. Billions of connected people, apps and things are leading to the creation of new digital communities and ecosystems. Next Generation Giants harness API enabled businesses models to deliver hyper-relevant, orchestrated communications to customers. Smart things change how we experience the world. Deep insights power decisions for internal and external users. This session will answer questions such as
What’s the right alchemic formula of integrating, aggregating, mashing, and stacking to fuel the innovation necessary to meet the needs of these Next Generation Giants so that we may build one?
What expectations and experiences will our new digital neighbors demand?
How will we work dynamically in secure cyber communities?
How do we continue to disrupt how it has always been done?
Deploying Enterprise Scale Deep Learning in Actuarial Modeling at NationwideDatabricks
The traditional approach to insurance pricing involves fitting a generalized linear model (GLM) to data collected on historical claims payments and premiums received. The explosive growth in data availability and increasing competitiveness in the marketplace are challenging actuaries to find new insights in their data and make predictions with more granularity, improved speed and efficiency, and with tighter integration among business units to support strategic decisions.
In this session we will share our experience implementing deep hierarchical neural networks using TensorFlow and PySpark on Databricks. We will discuss the benefits of the ML Runtime, our experience using the goofys mount, our process for hyperparameter tuning, specific considerations for the large dataset size and extreme volatility present in insurance data, among other topics.
Authors: Bryn Clark, Krish Rajaram
MongoDB World 2018: A Journey to the Cloud with Fraud Detection, Transactions...MongoDB
This presentation discusses Fair Isaac Corporation's Falcon Assurance Navigator product and its migration to MongoDB Atlas. FAN is a fraud detection and compliance monitoring solution. The presentation summarizes FAN's architecture, challenges with its previous monolithic on-premises implementation, and goals for a new microservices-based architecture using MongoDB Atlas. It provides examples of complex data modeling, analytics pipelines, and strategies for distributed computing, fault tolerance, and ensuring data consistency in the new architecture.
Quantlogic is a leading developer of algorithmic and automated trading strategies. They have invested millions developing their own intellectual property and database of trading strategies and concepts. They provide research services to banks, hedge funds, and proprietary trading desks, including quantitative modeling, statistical analysis, and programming complex trading models. Their research capabilities are powered by a vast library of trading indicators and concepts, as well as industry-leading technology and tools.
ONEMARKETDATA and ONETICK, a company and product overviewOneMarketData, LLC
A presentation on the company and OneTick product, the leading technology for complex event processing (CEP) and tick database for the financial industry.
OneTick and the R mathematical language, a presentation from R in FinanceOneMarketData, LLC
A presentation on the use of OneTick and the R mathematical and analytical language. This session presents the integration of OneTick and R in three use-case scenarios
Serverless microservices allow building scalable and resilient applications from small, isolated services using AWS Lambda and API Gateway. Each microservice owns its own data in a decentralized data store like DynamoDB. API Gateway acts as a front door and handles authentication, authorization, and throttling. Lambda provides immutable function versions and aliases for deployments. While this makes applications highly available and scalable, it introduces challenges around transactions and data consistency. The document proposes using techniques like correlation IDs, rollback functions, DynamoDB streams, and a transaction manager to handle errors and rollbacks in a serverless environment.
How Apache Spark and Apache Hadoop are being used to keep banking regulators ...DataWorks Summit
The global financial crisis showed that traditional IT systems at banks were ill equiped to monitor and manage the daily-changing risk landscape during the global financial crisis. The sheer amount of data that needed to be crunched meant that many of the banks were day(s) behind in calculating, understanding and reporting their risk positions. Post crisis, a review by banking regulator, led the regulators to introduce a new legislation BCBS 239: Principles for effective risk data aggregation and reporting, that requires banks to meet more stringent (timeliness) requirement, in their ability to aggregate and report on their quickly-changing risk positions or risk fines to the tune of $millions. To meet these new requirements, banks have been forced to re-think their traditional IT architectures, which are unable to cope with sheer volume of risk data, and are instead turning to Apache Hadoop and Apache Spark to build out next generation of risk systems. In this talk you will discover, how some of the leading banks in the world are leveraging Apache Hadoop and Apache Spark to meet BCBS 239 regulation.
Speaker
Kunal Taneja
WSO2Con USA 2015: The Needs of Next Generation GiantsWSO2
50% of the Fortune 500 companies from 2000 are gone! Facebook just became more valuable than Walmart! Next Generation Giants are being created every day and they are standing on the shoulders of technology and new business models to get there. Billions of connected people, apps and things are leading to the creation of new digital communities and ecosystems. Next Generation Giants harness API enabled businesses models to deliver hyper-relevant, orchestrated communications to customers. Smart things change how we experience the world. Deep insights power decisions for internal and external users. This session will answer questions such as
What’s the right alchemic formula of integrating, aggregating, mashing, and stacking to fuel the innovation necessary to meet the needs of these Next Generation Giants so that we may build one?
What expectations and experiences will our new digital neighbors demand?
How will we work dynamically in secure cyber communities?
How do we continue to disrupt how it has always been done?
Deploying Enterprise Scale Deep Learning in Actuarial Modeling at NationwideDatabricks
The traditional approach to insurance pricing involves fitting a generalized linear model (GLM) to data collected on historical claims payments and premiums received. The explosive growth in data availability and increasing competitiveness in the marketplace are challenging actuaries to find new insights in their data and make predictions with more granularity, improved speed and efficiency, and with tighter integration among business units to support strategic decisions.
In this session we will share our experience implementing deep hierarchical neural networks using TensorFlow and PySpark on Databricks. We will discuss the benefits of the ML Runtime, our experience using the goofys mount, our process for hyperparameter tuning, specific considerations for the large dataset size and extreme volatility present in insurance data, among other topics.
Authors: Bryn Clark, Krish Rajaram
MongoDB World 2018: A Journey to the Cloud with Fraud Detection, Transactions...MongoDB
This presentation discusses Fair Isaac Corporation's Falcon Assurance Navigator product and its migration to MongoDB Atlas. FAN is a fraud detection and compliance monitoring solution. The presentation summarizes FAN's architecture, challenges with its previous monolithic on-premises implementation, and goals for a new microservices-based architecture using MongoDB Atlas. It provides examples of complex data modeling, analytics pipelines, and strategies for distributed computing, fault tolerance, and ensuring data consistency in the new architecture.
ICP for Data- Enterprise platform for AI, ML and Data ScienceKaran Sachdeva
IBM Cloud Private for Data, an ultimate platform for all AI, ML and Data Science workloads. Integrated analytics platform based on Containers and micro services. Works with Kubernetes and dockers, even with Redhat openshift. Delivers the variety of business use cases in all industries- FS, Telco, Retail, Manufacturing etc
Data Engineer's Lunch #60: Series - Developing Enterprise ConsciousnessAnant Corporation
In Data Engineer's Lunch #60, Rahul Singh, CEO here at Anant, will discuss modern data processing/pipeline approaches.
Want to learn about modern data engineering patterns & practices for global data platforms? A high-level overview of different types, frameworks, and workflows in data processing and pipeline design.
Big Data: InterConnect 2016 Session on Getting Started with Big Data AnalyticsCynthia Saracco
Learn how to get started with Big Data using a platform based on Apache Hadoop, Apache Spark, and IBM BigInsights technologies. The emphasis here is on free or low-cost options that require modest technical skills.
With the rapid growth in data and move towards data commercialisation there are multiple aspects to focus on and prioritize the steps being taken across an enterprise. Enterprises face many challenges when it comes to truly becoming a data driven organization and realize the full potential of data. Some of those challenges include data availability, capacity to process, store and analyze this data, sharing the models and data artefacts across different teams etc. Most of these challenges could be handled through a platform which is Cloud based, scalable, and offers different capabilities for Governance, security, reusability and their likes. In this talk, I will talk about how IBM Cloud Pak serves as a framework for implementing your AI Strategy and how it could be used to build different artefacts while adhering to above listed requirements and being future ready. We will further illustrate how Cloud Pak for Data fastens and shortens the route to data commercialisation?
ADV Slides: The Evolution of the Data Platform and What It Means to Enterpris...DATAVERSITY
Thirty years is a long time for a technology foundation to be as active as relational databases. Are their replacements here?
In this webinar, we look at this foundational technology for modern Data Management and show how it evolved to meet the workloads of today, as well as when other platforms make sense for enterprise data.
엔터프라이즈의 AI/ML 활용을 돕는 Paxata 지능형 데이터 전처리 플랫폼 (최문규 이사, PAXATA) :: AWS Techforum...Amazon Web Services Korea
This document discusses Paxata, an intelligent data preparation platform. It summarizes Paxata's history and products, and describes common data challenges that enterprises face. These include spending significant resources on manual data preparation in Excel, which can introduce errors and limit agility. The document then outlines how Paxata addresses these challenges through its self-service, visual, intelligent and collaborative data preparation capabilities. It provides examples of Paxata's use in machine learning pipelines and integration with AWS services. Customer use cases and industry analyst recognition of Paxata as a leader are also mentioned.
This document provides guidance on developing a cloud migration strategy for typical large enterprise customers. It recommends starting with a cohesive approach involving sales, partners, solutions architects, and support teams. Key steps include obtaining executive sponsorship, identifying cloud champions, presenting integrated solutions, and thinking big. It also provides tips on assessing applications and prioritizing migrations, including focusing first on underutilized assets and those needing immediate scaling. Proof of concepts are recommended to build support and validate the approach before full migrations. Success criteria should go beyond just costs to include factors like agility, time to market, and new opportunities.
All-Flash Versus Hybrid VMware Virtual SAN™: Performance vs. Price Western Digital
An expert panel discussed all-flash versus hybrid VMware Virtual SAN storage solutions. The discussion included:
- A lab report that found an all-flash four-node Virtual SAN cluster delivered 49% better performance and 26% better price/performance than a hybrid configuration for database workloads.
- Key features of Virtual SAN like policy-based management, high availability, and scalability from 2-64 nodes.
- Reference architectures using SanDisk flash products in Virtual SAN deployments that achieved high performance results.
- Certified ready nodes from Lenovo and Supermicro that integrate SanDisk flash for Virtual SAN.
ADV Slides: How to Improve Your Analytic Data Architecture MaturityDATAVERSITY
Many organizations are immature when it comes to data use. The answer lies in delivering a greater level of insight from data, straight to the point of need. Enter: machine learning.
In this webinar, William will look at categories of organizational response to the challenge across strategy, architecture, modeling, processes, and ethics. Machine learning maturity levels tend to move in harmony across these categories. As a general principle of maturity models, you can’t skip levels in any category, nor can you advance in one category well beyond the others.
Vis-à-vis ML, attaining and retaining momentum up the model is paramount for success. You will ascend the model through concerted efforts delivering business wins utilizing progressive elements of the model, and thereby increasing your machine learning maturity. The model will evolve. No plateaus are comfortable for long.
With ML maturity markers, sequencing, and tactics, this webinar provides a plan for how to build analytic Data Architecture maturity in your organization.
The document discusses Oracle's autonomous database technology. It summarizes that autonomous databases can self-drive, self-repair, and self-secure with reduced human labor. Machine learning is used to continuously optimize databases and adapt to changing workloads. This allows DBAs to focus on higher value tasks like innovation rather than maintenance operations. Oracle's autonomous database is presented as the world's first fully autonomous database.
Enterprise Blockchains – A Pragmatic & Realistic Guide for CIOs/CDOSIceventure
Enterprise Blockchains – A Pragmatic & Realistic Guide for CIOs/CDOS – an Iceventure briefing
Background for the Enterprise Blockchain briefing
The blockchain promises to alter the way business transactions, data storage, industry value chains and associated revenue models. And unlike other innovation first deployed by start-ups many Enterprises as well as SMEs investigate the various use cases.
For example, Banking & Financial Services is the leading sector in enterprise blockchain adoption with a high number of PoCs. Adoption of use cases around payments, trade finance, and wallets are advancing rapidly.
Challenges for Enterprises/SMEs adopting an Enterprise Blockchain
But the novelty of the Enterprise Blockchain technology brings also about a lot of questions asked by market participants and our customers alike. Some examples are:
How to evaluate the advantages/disadvantages
What are the value drivers
Fitting it into the existing IT landscape
Where is the market and what are my risks
In a webinar, we therefore discuss the challenges of a successful Enterprise Blockchain use-case to separate the hype from reality. We also introduce a framework for evaluation
Watch the slides for:
A picture of the possibilities and challenges of the Blockchain
The Blockchain technology in the Enterprise context
A different use case valuation approach
An overview of the current market with focus on Germany
All Iceventure studies: https://www.iceventure.de/studien.html
This document outlines an agenda for a meetup on Spark UDF performance. The meetup will include: an introduction of the speaker ([Guilherme Braccialli]); an overview of QuantumBlack, where the speaker works; a presentation on Spark UDF performance, including a live demo; conclusions about the performance of PySpark vs Scala UDFs; and time for questions. The speaker will share learnings from running Spark at scale and practical examples. He will conclude that PySpark Pandas UDFs can be faster than regular PySpark UDFs but not always, and that PySpark UDFs are slower than equivalent Scala UDFs. The speaker's approach is to use PySpark UDFs
Estrategias para explotar las tendencias de SaaS y Cloud ComputingSoftware Guru
The document discusses strategies for leveraging Software as a Service (SaaS) and cloud computing technologies. It begins by defining SaaS and cloud computing, then discusses their importance and growth. The remainder outlines strategies for software vendors, including building SaaS applications using platform as a service (PaaS) or combining various cloud services, as well as addressing technical considerations like multi-tenancy and billing when developing SaaS products.
Insights into Real-world Data Management ChallengesDataWorks Summit
Oracle began with the belief that the foundation of IT was managing information. The Oracle Cloud Platform for Big Data is a natural extension of our belief in the power of data. Oracle’s Integrated Cloud is one cloud for the entire business, meeting everyone’s needs. It’s about Connecting people to information through tools which help you combine and aggregate data from any source.
This session will explore how organizations can transition to the cloud by delivering fully managed and elastic Hadoop and Real-time Streaming cloud services to built robust offerings that provide measurable value to the business. We will explore key data management trends and dive deeper into pain points we are hearing about from our customer base.
Domino and AWS: collaborative analytics and model governance at financial ser...Domino Data Lab
The document discusses how financial services firms use analytics for tasks like predictive modeling, validation, pricing, and research. It notes the challenges of legacy systems, collaboration across teams, and reproducibility. It then provides an example of how DBRS, a credit rating agency, uses Domino and AWS for securitization analysis. Models are developed in Jupyter notebooks and governed via a GitHub repository, with analysts interacting through Excel/R Shiny frontends on Domino. This allows for an auditable, scalable, and collaborative workflow while developers maintain control. The document concludes that collaborative platforms like Domino enable subject matter experts to focus on models rather than infrastructure.
It about the technical session. I have given a talk so that local people know about the cloud and they feel motivated to work with the cloud. It was basically for newbies who are planning to start their career. I tried to show them who they would be a cloud engineer what's will be their future responsibility and more.
Software development is changing rapidly. Enterprises that want to capture value faster, have to deliver value faster. The way to do that is by delivering software in production fast. Think multiple x a day. How do you transform to a digital enterprise that enables that?
PayPal datalake journey | teradata - edge of next | san diego | 2017 october ...Deepak Chandramouli
PayPal Data Lake Journey | 2017-Oct | San Diego | Teradata Edge of Next
Gimel [http://www.gimel.io] is a Big Data Processing Library, open sourced by PayPal.
https://www.youtube.com/watch?v=52PdNno_9cU&t=3s
Gimel empowers analysts, scientists, data engineers alike to access a variety of Big Data / Traditional Data Stores - with just SQL or a single line of code (Unified Data API).
This is possible via the Catalog of Technical properties abstracted from users, along with a rich collection of Data Store Connectors available in Gimel Library.
A Catalog provider can be Hive or User Supplied (runtime) or UDC.
In addition, PayPal recently open sourced UDC [Unified Data Catalog], which can host and serve the Technical Metatada of the Data Stores & Objects. Visit http://www.unifieddatacatalog.io to experience first hand.
GlobalLogic Java Community Webinar #18 “How to Improve Web Application Perfor...GlobalLogic Ukraine
Під час доповіді відповімо на питання, навіщо потрібно підвищувати продуктивність аплікації і які є найефективніші способи для цього. А також поговоримо про те, що таке кеш, які його види бувають та, основне — як знайти performance bottleneck?
Відео та деталі заходу: https://bit.ly/45tILxj
This talk will cover ScyllaDB Architecture from the cluster-level view and zoom in on data distribution and internal node architecture. In the process, we will learn the secret sauce used to get ScyllaDB's high availability and superior performance. We will also touch on the upcoming changes to ScyllaDB architecture, moving to strongly consistent metadata and tablets.
More Related Content
Similar to Finance Trading in The Cloud - AWS Michigan Meetup
ICP for Data- Enterprise platform for AI, ML and Data ScienceKaran Sachdeva
IBM Cloud Private for Data, an ultimate platform for all AI, ML and Data Science workloads. Integrated analytics platform based on Containers and micro services. Works with Kubernetes and dockers, even with Redhat openshift. Delivers the variety of business use cases in all industries- FS, Telco, Retail, Manufacturing etc
Data Engineer's Lunch #60: Series - Developing Enterprise ConsciousnessAnant Corporation
In Data Engineer's Lunch #60, Rahul Singh, CEO here at Anant, will discuss modern data processing/pipeline approaches.
Want to learn about modern data engineering patterns & practices for global data platforms? A high-level overview of different types, frameworks, and workflows in data processing and pipeline design.
Big Data: InterConnect 2016 Session on Getting Started with Big Data AnalyticsCynthia Saracco
Learn how to get started with Big Data using a platform based on Apache Hadoop, Apache Spark, and IBM BigInsights technologies. The emphasis here is on free or low-cost options that require modest technical skills.
With the rapid growth in data and move towards data commercialisation there are multiple aspects to focus on and prioritize the steps being taken across an enterprise. Enterprises face many challenges when it comes to truly becoming a data driven organization and realize the full potential of data. Some of those challenges include data availability, capacity to process, store and analyze this data, sharing the models and data artefacts across different teams etc. Most of these challenges could be handled through a platform which is Cloud based, scalable, and offers different capabilities for Governance, security, reusability and their likes. In this talk, I will talk about how IBM Cloud Pak serves as a framework for implementing your AI Strategy and how it could be used to build different artefacts while adhering to above listed requirements and being future ready. We will further illustrate how Cloud Pak for Data fastens and shortens the route to data commercialisation?
ADV Slides: The Evolution of the Data Platform and What It Means to Enterpris...DATAVERSITY
Thirty years is a long time for a technology foundation to be as active as relational databases. Are their replacements here?
In this webinar, we look at this foundational technology for modern Data Management and show how it evolved to meet the workloads of today, as well as when other platforms make sense for enterprise data.
엔터프라이즈의 AI/ML 활용을 돕는 Paxata 지능형 데이터 전처리 플랫폼 (최문규 이사, PAXATA) :: AWS Techforum...Amazon Web Services Korea
This document discusses Paxata, an intelligent data preparation platform. It summarizes Paxata's history and products, and describes common data challenges that enterprises face. These include spending significant resources on manual data preparation in Excel, which can introduce errors and limit agility. The document then outlines how Paxata addresses these challenges through its self-service, visual, intelligent and collaborative data preparation capabilities. It provides examples of Paxata's use in machine learning pipelines and integration with AWS services. Customer use cases and industry analyst recognition of Paxata as a leader are also mentioned.
This document provides guidance on developing a cloud migration strategy for typical large enterprise customers. It recommends starting with a cohesive approach involving sales, partners, solutions architects, and support teams. Key steps include obtaining executive sponsorship, identifying cloud champions, presenting integrated solutions, and thinking big. It also provides tips on assessing applications and prioritizing migrations, including focusing first on underutilized assets and those needing immediate scaling. Proof of concepts are recommended to build support and validate the approach before full migrations. Success criteria should go beyond just costs to include factors like agility, time to market, and new opportunities.
All-Flash Versus Hybrid VMware Virtual SAN™: Performance vs. Price Western Digital
An expert panel discussed all-flash versus hybrid VMware Virtual SAN storage solutions. The discussion included:
- A lab report that found an all-flash four-node Virtual SAN cluster delivered 49% better performance and 26% better price/performance than a hybrid configuration for database workloads.
- Key features of Virtual SAN like policy-based management, high availability, and scalability from 2-64 nodes.
- Reference architectures using SanDisk flash products in Virtual SAN deployments that achieved high performance results.
- Certified ready nodes from Lenovo and Supermicro that integrate SanDisk flash for Virtual SAN.
ADV Slides: How to Improve Your Analytic Data Architecture MaturityDATAVERSITY
Many organizations are immature when it comes to data use. The answer lies in delivering a greater level of insight from data, straight to the point of need. Enter: machine learning.
In this webinar, William will look at categories of organizational response to the challenge across strategy, architecture, modeling, processes, and ethics. Machine learning maturity levels tend to move in harmony across these categories. As a general principle of maturity models, you can’t skip levels in any category, nor can you advance in one category well beyond the others.
Vis-à-vis ML, attaining and retaining momentum up the model is paramount for success. You will ascend the model through concerted efforts delivering business wins utilizing progressive elements of the model, and thereby increasing your machine learning maturity. The model will evolve. No plateaus are comfortable for long.
With ML maturity markers, sequencing, and tactics, this webinar provides a plan for how to build analytic Data Architecture maturity in your organization.
The document discusses Oracle's autonomous database technology. It summarizes that autonomous databases can self-drive, self-repair, and self-secure with reduced human labor. Machine learning is used to continuously optimize databases and adapt to changing workloads. This allows DBAs to focus on higher value tasks like innovation rather than maintenance operations. Oracle's autonomous database is presented as the world's first fully autonomous database.
Enterprise Blockchains – A Pragmatic & Realistic Guide for CIOs/CDOSIceventure
Enterprise Blockchains – A Pragmatic & Realistic Guide for CIOs/CDOS – an Iceventure briefing
Background for the Enterprise Blockchain briefing
The blockchain promises to alter the way business transactions, data storage, industry value chains and associated revenue models. And unlike other innovation first deployed by start-ups many Enterprises as well as SMEs investigate the various use cases.
For example, Banking & Financial Services is the leading sector in enterprise blockchain adoption with a high number of PoCs. Adoption of use cases around payments, trade finance, and wallets are advancing rapidly.
Challenges for Enterprises/SMEs adopting an Enterprise Blockchain
But the novelty of the Enterprise Blockchain technology brings also about a lot of questions asked by market participants and our customers alike. Some examples are:
How to evaluate the advantages/disadvantages
What are the value drivers
Fitting it into the existing IT landscape
Where is the market and what are my risks
In a webinar, we therefore discuss the challenges of a successful Enterprise Blockchain use-case to separate the hype from reality. We also introduce a framework for evaluation
Watch the slides for:
A picture of the possibilities and challenges of the Blockchain
The Blockchain technology in the Enterprise context
A different use case valuation approach
An overview of the current market with focus on Germany
All Iceventure studies: https://www.iceventure.de/studien.html
This document outlines an agenda for a meetup on Spark UDF performance. The meetup will include: an introduction of the speaker ([Guilherme Braccialli]); an overview of QuantumBlack, where the speaker works; a presentation on Spark UDF performance, including a live demo; conclusions about the performance of PySpark vs Scala UDFs; and time for questions. The speaker will share learnings from running Spark at scale and practical examples. He will conclude that PySpark Pandas UDFs can be faster than regular PySpark UDFs but not always, and that PySpark UDFs are slower than equivalent Scala UDFs. The speaker's approach is to use PySpark UDFs
Estrategias para explotar las tendencias de SaaS y Cloud ComputingSoftware Guru
The document discusses strategies for leveraging Software as a Service (SaaS) and cloud computing technologies. It begins by defining SaaS and cloud computing, then discusses their importance and growth. The remainder outlines strategies for software vendors, including building SaaS applications using platform as a service (PaaS) or combining various cloud services, as well as addressing technical considerations like multi-tenancy and billing when developing SaaS products.
Insights into Real-world Data Management ChallengesDataWorks Summit
Oracle began with the belief that the foundation of IT was managing information. The Oracle Cloud Platform for Big Data is a natural extension of our belief in the power of data. Oracle’s Integrated Cloud is one cloud for the entire business, meeting everyone’s needs. It’s about Connecting people to information through tools which help you combine and aggregate data from any source.
This session will explore how organizations can transition to the cloud by delivering fully managed and elastic Hadoop and Real-time Streaming cloud services to built robust offerings that provide measurable value to the business. We will explore key data management trends and dive deeper into pain points we are hearing about from our customer base.
Domino and AWS: collaborative analytics and model governance at financial ser...Domino Data Lab
The document discusses how financial services firms use analytics for tasks like predictive modeling, validation, pricing, and research. It notes the challenges of legacy systems, collaboration across teams, and reproducibility. It then provides an example of how DBRS, a credit rating agency, uses Domino and AWS for securitization analysis. Models are developed in Jupyter notebooks and governed via a GitHub repository, with analysts interacting through Excel/R Shiny frontends on Domino. This allows for an auditable, scalable, and collaborative workflow while developers maintain control. The document concludes that collaborative platforms like Domino enable subject matter experts to focus on models rather than infrastructure.
It about the technical session. I have given a talk so that local people know about the cloud and they feel motivated to work with the cloud. It was basically for newbies who are planning to start their career. I tried to show them who they would be a cloud engineer what's will be their future responsibility and more.
Software development is changing rapidly. Enterprises that want to capture value faster, have to deliver value faster. The way to do that is by delivering software in production fast. Think multiple x a day. How do you transform to a digital enterprise that enables that?
PayPal datalake journey | teradata - edge of next | san diego | 2017 october ...Deepak Chandramouli
PayPal Data Lake Journey | 2017-Oct | San Diego | Teradata Edge of Next
Gimel [http://www.gimel.io] is a Big Data Processing Library, open sourced by PayPal.
https://www.youtube.com/watch?v=52PdNno_9cU&t=3s
Gimel empowers analysts, scientists, data engineers alike to access a variety of Big Data / Traditional Data Stores - with just SQL or a single line of code (Unified Data API).
This is possible via the Catalog of Technical properties abstracted from users, along with a rich collection of Data Store Connectors available in Gimel Library.
A Catalog provider can be Hive or User Supplied (runtime) or UDC.
In addition, PayPal recently open sourced UDC [Unified Data Catalog], which can host and serve the Technical Metatada of the Data Stores & Objects. Visit http://www.unifieddatacatalog.io to experience first hand.
Similar to Finance Trading in The Cloud - AWS Michigan Meetup (20)
GlobalLogic Java Community Webinar #18 “How to Improve Web Application Perfor...GlobalLogic Ukraine
Під час доповіді відповімо на питання, навіщо потрібно підвищувати продуктивність аплікації і які є найефективніші способи для цього. А також поговоримо про те, що таке кеш, які його види бувають та, основне — як знайти performance bottleneck?
Відео та деталі заходу: https://bit.ly/45tILxj
This talk will cover ScyllaDB Architecture from the cluster-level view and zoom in on data distribution and internal node architecture. In the process, we will learn the secret sauce used to get ScyllaDB's high availability and superior performance. We will also touch on the upcoming changes to ScyllaDB architecture, moving to strongly consistent metadata and tablets.
What is an RPA CoE? Session 2 – CoE RolesDianaGray10
In this session, we will review the players involved in the CoE and how each role impacts opportunities.
Topics covered:
• What roles are essential?
• What place in the automation journey does each role play?
Speaker:
Chris Bolin, Senior Intelligent Automation Architect Anika Systems
The Department of Veteran Affairs (VA) invited Taylor Paschal, Knowledge & Information Management Consultant at Enterprise Knowledge, to speak at a Knowledge Management Lunch and Learn hosted on June 12, 2024. All Office of Administration staff were invited to attend and received professional development credit for participating in the voluntary event.
The objectives of the Lunch and Learn presentation were to:
- Review what KM ‘is’ and ‘isn’t’
- Understand the value of KM and the benefits of engaging
- Define and reflect on your “what’s in it for me?”
- Share actionable ways you can participate in Knowledge - - Capture & Transfer
Conversational agents, or chatbots, are increasingly used to access all sorts of services using natural language. While open-domain chatbots - like ChatGPT - can converse on any topic, task-oriented chatbots - the focus of this paper - are designed for specific tasks, like booking a flight, obtaining customer support, or setting an appointment. Like any other software, task-oriented chatbots need to be properly tested, usually by defining and executing test scenarios (i.e., sequences of user-chatbot interactions). However, there is currently a lack of methods to quantify the completeness and strength of such test scenarios, which can lead to low-quality tests, and hence to buggy chatbots.
To fill this gap, we propose adapting mutation testing (MuT) for task-oriented chatbots. To this end, we introduce a set of mutation operators that emulate faults in chatbot designs, an architecture that enables MuT on chatbots built using heterogeneous technologies, and a practical realisation as an Eclipse plugin. Moreover, we evaluate the applicability, effectiveness and efficiency of our approach on open-source chatbots, with promising results.
"NATO Hackathon Winner: AI-Powered Drug Search", Taras KlobaFwdays
This is a session that details how PostgreSQL's features and Azure AI Services can be effectively used to significantly enhance the search functionality in any application.
In this session, we'll share insights on how we used PostgreSQL to facilitate precise searches across multiple fields in our mobile application. The techniques include using LIKE and ILIKE operators and integrating a trigram-based search to handle potential misspellings, thereby increasing the search accuracy.
We'll also discuss how the azure_ai extension on PostgreSQL databases in Azure and Azure AI Services were utilized to create vectors from user input, a feature beneficial when users wish to find specific items based on text prompts. While our application's case study involves a drug search, the techniques and principles shared in this session can be adapted to improve search functionality in a wide range of applications. Join us to learn how PostgreSQL and Azure AI can be harnessed to enhance your application's search capability.
The Microsoft 365 Migration Tutorial For Beginner.pptxoperationspcvita
This presentation will help you understand the power of Microsoft 365. However, we have mentioned every productivity app included in Office 365. Additionally, we have suggested the migration situation related to Office 365 and how we can help you.
You can also read: https://www.systoolsgroup.com/updates/office-365-tenant-to-tenant-migration-step-by-step-complete-guide/
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving
Manufacturing custom quality metal nameplates and badges involves several standard operations. Processes include sheet prep, lithography, screening, coating, punch press and inspection. All decoration is completed in the flat sheet with adhesive and tooling operations following. The possibilities for creating unique durable nameplates are endless. How will you create your brand identity? We can help!
Dandelion Hashtable: beyond billion requests per second on a commodity serverAntonios Katsarakis
This slide deck presents DLHT, a concurrent in-memory hashtable. Despite efforts to optimize hashtables, that go as far as sacrificing core functionality, state-of-the-art designs still incur multiple memory accesses per request and block request processing in three cases. First, most hashtables block while waiting for data to be retrieved from memory. Second, open-addressing designs, which represent the current state-of-the-art, either cannot free index slots on deletes or must block all requests to do so. Third, index resizes block every request until all objects are copied to the new index. Defying folklore wisdom, DLHT forgoes open-addressing and adopts a fully-featured and memory-aware closed-addressing design based on bounded cache-line-chaining. This design offers lock-free index operations and deletes that free slots instantly, (2) completes most requests with a single memory access, (3) utilizes software prefetching to hide memory latencies, and (4) employs a novel non-blocking and parallel resizing. In a commodity server and a memory-resident workload, DLHT surpasses 1.6B requests per second and provides 3.5x (12x) the throughput of the state-of-the-art closed-addressing (open-addressing) resizable hashtable on Gets (Deletes).
LF Energy Webinar: Carbon Data Specifications: Mechanisms to Improve Data Acc...DanBrown980551
This LF Energy webinar took place June 20, 2024. It featured:
-Alex Thornton, LF Energy
-Hallie Cramer, Google
-Daniel Roesler, UtilityAPI
-Henry Richardson, WattTime
In response to the urgency and scale required to effectively address climate change, open source solutions offer significant potential for driving innovation and progress. Currently, there is a growing demand for standardization and interoperability in energy data and modeling. Open source standards and specifications within the energy sector can also alleviate challenges associated with data fragmentation, transparency, and accessibility. At the same time, it is crucial to consider privacy and security concerns throughout the development of open source platforms.
This webinar will delve into the motivations behind establishing LF Energy’s Carbon Data Specification Consortium. It will provide an overview of the draft specifications and the ongoing progress made by the respective working groups.
Three primary specifications will be discussed:
-Discovery and client registration, emphasizing transparent processes and secure and private access
-Customer data, centering around customer tariffs, bills, energy usage, and full consumption disclosure
-Power systems data, focusing on grid data, inclusive of transmission and distribution networks, generation, intergrid power flows, and market settlement data
"Scaling RAG Applications to serve millions of users", Kevin GoedeckeFwdays
How we managed to grow and scale a RAG application from zero to thousands of users in 7 months. Lessons from technical challenges around managing high load for LLMs, RAGs and Vector databases.
What is an RPA CoE? Session 1 – CoE VisionDianaGray10
In the first session, we will review the organization's vision and how this has an impact on the COE Structure.
Topics covered:
• The role of a steering committee
• How do the organization’s priorities determine CoE Structure?
Speaker:
Chris Bolin, Senior Intelligent Automation Architect Anika Systems
2. Objectives
‣ Introduce Quantitative Trading
‣ Present a case study on AWS usage in Quantitative Trading System
Evaluation.
‣ Discuss potential improvements upon our presented architecture.
http://www.solidlogic.com
2
4. Solid Logic Technology develops innovative custom
technology solutions across a variety of industries
using leading software, infrastructure and
business practices.
http://www.solidlogic.com
4
5. About us
Our expertise Industry experience
Infrastructure and cloud computing ‣ Financial and legal services
‣ Scalable, programmatic infrastructure ‣ Logistics
management
‣ Automotive
‣ Strategic data center design
‣ Defense and homeland security
‣ VMware architecture and management
‣ Consumer sales and service
‣ Multi-cloud development and deployment
‣ Scalable web infrastructure with CDN
‣ Academic and scientific research
‣ Security and compliance methods and
implementation
Software development Company Information
‣ Analytical solutions - simulation, optimization, big ‣ Founded in 2011
data, natural language processing, quant. finance
‣ Entirely mobile company
‣ Enterprise content management, workflow
solutions, system integration ‣ Develop both internal projects (IP) and
‣ Oracle Transportation Management client software solutions
‣ Database technology (Oracle, Vertica, Postgres,
Cassandra, etc.)
‣ Web application and website development
http://www.solidlogic.com
5
6. Solid Logic Management Team
‣ Eric Detterman, CEO and Co-Founder
• Professional Experience
- Legal IT Business Analyst, Lean Startup, Cloud Computing, Processing Engineering and Consulting
- Researched and developed core investment strategies for Birmingham, MI RIA
- Currently in production and managing > $20M, AUM growth > 50% annually
- Proprietary trading (equities, futures, options), web and software development
• Education: B.S. Economics – Oakland University
‣ Michael Bommarito, CIO and Partner
• Relevant Experience
- “Big data” consultant, Oracle ERP architect, Linux cluster administrator.
- Software developer - NYC-based quantitative hedge fund
- Consultant - multiple quantitative hedge funds
• Education : M.S.E Financial Engineering, M.S. Political Science, B.S. Mathematics –
University of Michigan
‣ Ronald Redmer, Board Member and Lead Technical Advisor
• Relevant Experience
- CIO, National Default Exchange (NDeX), a business unit of The Dolan Company (NYSE:DM)
- CEO defense supplier company, Airport systems software, CEO auto testing company, Affina –
software dev mgr, EDS - tech lead http://www.solidlogic.com
6
8. Case Study: Proprietary Trading Simulation
Quantitative Trading and Investment Systems:
‣ (Loose) Definition:
• Rules-based mathematical ‘model’ created by testing and validating a hypothesis
about how a tradable market acts or optimizing parameters to create an equation
to describe the market.
• The goal is to outperform the broad market (S&P 500) or some benchmark after
costs.
‣ Example Strategy:
• Investment universe = ~50 Fidelity Mutual Funds
• Strategy #1: Invest in the top six ranked mutual funds based on proprietary
momentum (p0 > p-1) based ranking algorithm. Analyze and rank fund universe
every 45 days and re-allocate.
• Strategy #2: Invest in the top six ranked mutual funds based on proprietary mean
reversion (p0 > p-1) based ranking algorithm. Analyze and rank fund universe
every 45 days and re-allocate. http://www.solidlogic.com
8
10. Case Study: Proprietary Trading Simulation
Challenge: Scope:
‣ Characterize the performance and ‣ Assets 62
sensitivity of an equity trading
‣ Tests/asset 96
system across input parameters and
market conditions ‣ Total tests 5,952
‣ Optimize parameters based on profit
and risk measures Test Information
‣ Estimated runtime is unacceptable
‣ Mean components/asset 395
on local workstation (>1 month)
‣ Primary bottlenecks are in dense ‣ Points/component 3,135
linear algebra operations ‣ Points/test 1,238,325
• Spectral decomposition (ARPACK)
Pairwise comparison of higher-order
Total elements 7,370,510,400
•
distribution moments (M-M arithmetic) ‣
http://www.solidlogic.com
10
11. Case Study: Proprietary Trading Simulation
Potential solutions:
‣ Run on existing hardware – wait for results
‣ Physical or virtualized servers with supporting job schedulers –
requires hardware, software, and specialized labor
‣ Setup cloud infrastructure to process work – requires software
and specialized labor
http://www.solidlogic.com
11
12. Trading Simulation: Architecture
This was our initial version – Not overly elegant, but works very well
with minimal effort to setup. Easy to improve upon.
Strategy Test
Trading System Results
Source Code and Custom (S3 Buckets)
Config Data Created
(Git Repo) AMIs (x16)
Availability Zone
US East Region
http://www.solidlogic.com
12
13. Trading Simulation: Test Process
Trading
System
Source Code
(Git Repo)
Strategy Test
Results (S3
Custom Buckets)
Created
AMIs (x16)
Availability Zone
US East Region
Local
Development
Environment
http://www.solidlogic.com
13
14. Trading Simulation: Overview
Technology Solution: Compute Instance (x16):
‣ Built an optimized simulation ‣ 88 Elastic Compute Units (ECU)
environment as virtual image ‣ 2x Xeon E5-2670s-16 cores
(AWS EC2 AMI)
‣ 60.5GB RAM
‣ Provisioned and configured
‣ 10GbE, dual NIC
centralized storage (AWS S3)
‣ 3+TB instance scratch
• Experiment configuration
• Simulation input Total Compute Resources:
• Simulation output ‣ 1408 ECUs
• Post-processed results ‣ 512 concurrent threads (HT)
‣ Fully automated deployment of ‣ 968GB RAM
simulation to instances through
master source control system (1 ECU~=5GFlops)
(git) http://www.solidlogic.com
14
15. Trading Simulation: AMI Creation Process
‣ Use standard Ubuntu Server 12.04.1 LTS for Cluster Instances
AMI x64 (ami-eb7bcf82)
• cc2.8xLarge – 88ECUs, 16 cores, 60.5GB RAM
‣ Install git, s3cmd, PostgreSQL JDBC drivers
‣ Install and configure test environment and all dependencies
‣ Create new AMI based on the above
http://www.solidlogic.com
15
16. Trading Simulation: Test Execution
‣ For each test instance…..
• ssh -X -i /home/ericd/.aws/first/name.pem ubuntu@IP
• cd /home/ubuntu/testcode/tradingsystemsales
• git pull
• cd /usr/local/testcode//bin
• sudo ./testcode -nodesktop
• parameterSweepSingleNode('Yes','Yes',
'homeubuntutestcodetradingsystemsalesmodelsdailyAdaptiveStateSpaceSPYdatamasterlist.mat',
'/home/ubuntu/testcode/tradingsystemsales/models/daily/AdaptiveStateSpaceSPY/data/etfs', 'csv', '',
'/home/ubuntu/testcode/tradingsystemsales/models/daily/AdaptiveStateSpaceSPY/data/etfs', 1,
'homeubuntutestcodetradingsystemsalesmodelsdailyAdaptiveStateSpaceSPYdataETFsToTest.csv',
'/home/ubuntu/testcode/tradingsystemsales/models/daily/AdaptiveStateSpaceSPY/data/etfs', 'mat',
'/home/ubuntu/testcode/tradingsystemsales/models/daily/AdaptiveStateSpaceSPY/src', 'No', 'No')
• parameterSweepSingleNode('No','Yes',
'homeubuntutestcodetradingsystemsalesmodelsdailyAdaptiveStateSpaceSPYdatamasterlist.mat',
'/home/ubuntu/testcode/tradingsystemsales/models/daily/AdaptiveStateSpaceSPY/data/etfs', 'csv', '',
'/home/ubuntu/testcode/tradingsystemsales/models/daily/AdaptiveStateSpaceSPY/data/etfs', 1,
'homeubuntutestcodetradingsystemsalesmodelsdailyAdaptiveStateSpaceSPYdataETFsToTest.csv',
'/home/ubuntu/testcode/tradingsystemsales/models/daily/AdaptiveStateSpaceSPY/data/etfs', 'mat',
'/home/ubuntu/testcode/tradingsystemsales/models/daily/AdaptiveStateSpaceSPY/src', 'No', 'No')
http://www.solidlogic.com
16
17. Trading Simulation: Initial Test Results
‣ Result sets saved to S3 buckets using S3cmd
• Approximately 6000 result sets
http://www.solidlogic.com
17
18. Trading Simulation: Output
‣ Run Time:
• Cloud: 45 hours
• Single-seat: 1-2 months
• Order of magnitude improvement in time!
http://www.solidlogic.com
18
19. Trading Simulation: Economics
Co- On- Spot
On-Site
Location Demand Pricing Cost Model Details
Server Hardware / ‣ Cost estimates using assumptions
$69,045 $69,045 $1,647 $193
Instance Usage
and calculations in Cost Comparison
Network Hardware 13,809 13,809 - -
Worksheet
Hardware Maint. 24,856 24,856 - - ‣ Costs represent one year annualized
costs. Assumes a useful life of three
Operating System - - - -
years for purchased equipment
Power and Cooling 9,907 - - - ‣ 1= Cost savings using On-Site as
Data Center baseline
Construction / Co- 8,618 65,136 - -
Location Expense
‣ 2= On-Site and Co-Location assume
Admin. / Remote
100% usage
105,000 240 - -
Hands Support ‣ 3= Based on actual 686 machine
Data Transfer 1 4 1 1 hours used
Total $231,237 $173,091 $1,647 $193
Cost Savings1 N/A 25.12% 99.29% 99.92%
$ / Compute Hr.2,3 $26.40 $19.76 $2.40 $0.28
http://www.solidlogic.com
19
20. Trading Simulation: Next Steps
Potential Improvements:
‣ Develop improved cloud infrastructure management tools
• Allocation of work across instances
• Allow user defined completion time and programmatically
scale compute resources to work towards goal
• Spread work across unused internal and available external
compute resources
http://www.solidlogic.com
20
21. Thank you
Eric Detterman Michael Bommarito
CEO, Co-Founder CIO, Partner
Eric.Detterman@solidlogic.com Michael.Bommarito@solidlogic.com
Direct: (248) 792 – 8001 Direct: (646) 450 – 3387
(248) 792 – 8000
www.solidlogic.com
330 East Maple Rd. #231
Birmingham, MI 48009 21