As customer data grows massively, you need the tools to process data with the goal to answer important questions related to the success of your business. Traditional data processing tools have been effective in the past, but don't scale to grapple with the massive volume, velocity, and variety of data that's available to drive these decisions today. In addition, these tools required Salesforce customers to move data off-platform for processing. Salesforce provides a new tool - Data Pipeline - to help you process trillions of customer interactions on our trusted platform. Join us as we deep-dive and demo the Data Pipeline solution and cover interesting customer use-cases around Big Data Processing.
Kaizentric is a Data Analytics firm, based in Chennai, India. Statistical Analysis is performed on a well-built client specific data warehouse, supported by Data Mining.
Considerations for Data Migration D365 Finance & OperationsGina Pabalan
Harvesting enterprise data is central to how organizations compete, and even survive, as industries transform digitally. Yet, as companies merge and technologies shift, managing data has become an extremely complex but critical task, especially handled alongside of an enterprise ERP implementation.
For companies moving from an on-premise legacy ERP system to Microsoft’s cloud-based Dynamics 365 for Finance and Operations (“D365”), there are some unique challenges and new tools to leverage when considering the data migration activity.
Microsoft delivers the Data Management Framework (“DMF”) tool to assist customers with data migration for D365. Data migration itself consists of three distinct activities, as illustrated below: Data extraction (from legacy systems), data transformation and data import into D365. DMF will assist in the import into the new D365 application, but what is the best way to extract and transform, to “ready” the data for the import?
What you need to know about Data Migration for D365 Finance & OperationsGina Pabalan
This is a "nuts & bolts" whitepaper discussing the capabilities and challenges of migrating data to Microsoft Dynamics365 for Finance and Operations (D365).
Case Study - Ibotta Builds A Self-Service Data Lake To Enable Business Growth...Vasu S
Read a case study that how Ibotta cut costs thanks to Qubole’s autoscaling and downscaling capabilities, and the ability to isolate workloads to separate clusters
https://www.qubole.com/resources/case-study/ibotta
Kaizentric is a Data Analytics firm, based in Chennai, India. Statistical Analysis is performed on a well-built client specific data warehouse, supported by Data Mining.
Considerations for Data Migration D365 Finance & OperationsGina Pabalan
Harvesting enterprise data is central to how organizations compete, and even survive, as industries transform digitally. Yet, as companies merge and technologies shift, managing data has become an extremely complex but critical task, especially handled alongside of an enterprise ERP implementation.
For companies moving from an on-premise legacy ERP system to Microsoft’s cloud-based Dynamics 365 for Finance and Operations (“D365”), there are some unique challenges and new tools to leverage when considering the data migration activity.
Microsoft delivers the Data Management Framework (“DMF”) tool to assist customers with data migration for D365. Data migration itself consists of three distinct activities, as illustrated below: Data extraction (from legacy systems), data transformation and data import into D365. DMF will assist in the import into the new D365 application, but what is the best way to extract and transform, to “ready” the data for the import?
What you need to know about Data Migration for D365 Finance & OperationsGina Pabalan
This is a "nuts & bolts" whitepaper discussing the capabilities and challenges of migrating data to Microsoft Dynamics365 for Finance and Operations (D365).
Case Study - Ibotta Builds A Self-Service Data Lake To Enable Business Growth...Vasu S
Read a case study that how Ibotta cut costs thanks to Qubole’s autoscaling and downscaling capabilities, and the ability to isolate workloads to separate clusters
https://www.qubole.com/resources/case-study/ibotta
Apache Hadoop and Spark are best-of-breed technologies for distributed processing and storage of very large data sets: Big Data. Join us as we explain how to integrate Salesforce with off-the-shelf big data tools to build flexible applications. You'll also learn how Force.com is evolving in this area and how Big Objects and Data Pipelines will provide Big Data capability within the platform.
Customer-Centric Data Management for Better Customer ExperiencesInformatica
With consumer and business buyer expectations growing exponentially, more businesses are competing on the basis of customer experience. But executing preferred customer experiences requires data about who your customers are today and what will they likely need in the future. Every business can benefit from an AI-powered master data management platform to supply this information to line-of-business owners so they can execute great experiences at scale. This same need is true from an internal business process perspective as well. For example, many businesses require better data management practices to deliver preferred employee experiences. Informatica provides an MDM platform to solve for these examples and more.
Informatica Solution for SWIFT IntegrationKim Loughead
Overview of Informatica's solution for financial services organizations who need to exchange payment data including SWIFT, NACHA, SEPA, FIX, etc. messages with other financial institutions
Slides from a session at GigaCon Big Data conference in Warsaw, Poland on Jan 27, 2014. Updated with the content presented at Big Data Budapest on Aug 18, 2014.
MongoDB World 2019: Managing a Heterogeneous Data Stack with Informatica and ...MongoDB
Over the past few months, MongoDB and Informatica have worked together to extend the functionality and performance of connectivity to MongoDB. These connectivity improvements enhance overall user experience and utilize MongoDB's native drivers to connect to MongoDB to achieve great performance while managing data across systems.
This session will focus on managing a data stack including SQL Server, Oracle, and MongoDB Atlas using Informatica’s Intelligent Cloud Services (IICS)/iPaaS suite. We’ll discuss several real-world use cases, and demonstrate how to track data lineage, and develop complex data integration flows with Informatica iPaaS tooling.
Apache Hadoop and Spark are best-of-breed technologies for distributed processing and storage of very large data sets: Big Data. Join us as we explain how to integrate Salesforce with off-the-shelf big data tools to build flexible applications. You'll also learn how Force.com is evolving in this area and how Big Objects and Data Pipelines will provide Big Data capability within the platform.
Customer-Centric Data Management for Better Customer ExperiencesInformatica
With consumer and business buyer expectations growing exponentially, more businesses are competing on the basis of customer experience. But executing preferred customer experiences requires data about who your customers are today and what will they likely need in the future. Every business can benefit from an AI-powered master data management platform to supply this information to line-of-business owners so they can execute great experiences at scale. This same need is true from an internal business process perspective as well. For example, many businesses require better data management practices to deliver preferred employee experiences. Informatica provides an MDM platform to solve for these examples and more.
Informatica Solution for SWIFT IntegrationKim Loughead
Overview of Informatica's solution for financial services organizations who need to exchange payment data including SWIFT, NACHA, SEPA, FIX, etc. messages with other financial institutions
Slides from a session at GigaCon Big Data conference in Warsaw, Poland on Jan 27, 2014. Updated with the content presented at Big Data Budapest on Aug 18, 2014.
MongoDB World 2019: Managing a Heterogeneous Data Stack with Informatica and ...MongoDB
Over the past few months, MongoDB and Informatica have worked together to extend the functionality and performance of connectivity to MongoDB. These connectivity improvements enhance overall user experience and utilize MongoDB's native drivers to connect to MongoDB to achieve great performance while managing data across systems.
This session will focus on managing a data stack including SQL Server, Oracle, and MongoDB Atlas using Informatica’s Intelligent Cloud Services (IICS)/iPaaS suite. We’ll discuss several real-world use cases, and demonstrate how to track data lineage, and develop complex data integration flows with Informatica iPaaS tooling.
Unified Big Data Processing with Apache SparkC4Media
Video and slides synchronized, mp3 and slide download available at URL http://bit.ly/1yNuLGF.
Matei Zaharia talks about the latest developments in Spark and shows examples of how it can combine processing algorithms to build rich data pipelines in just a few lines of code. Filmed at qconsf.com.
Matei Zaharia is an assistant professor of computer science at MIT, and CTO of Databricks, the company commercializing Apache Spark.
Unleash the Potential of Big Data on SalesforceDreamforce
Salesforce hosts billions of customer records on Salesforce App Cloud. Making timely decisions on this invaluable data demands a new set of capabilities. From interacting with data real-time to leveraging a fluid integration with Salesforce Analytics, these capabilities are just around the corner. Join us in this roadmap session to see what the near-future of Big Data on Salesforce App Cloud looks like and how you can benefit from it. Watch the video now: https://www.youtube.com/watch?v=a-wFfdfGgvM
Analyze billions of records on Salesforce App Cloud with BigObjectSalesforce Developers
Salesforce hosts billions of customer records on Salesforce App Cloud. Making timely decisions on this invaluable data demands a new set of capabilities. From interacting with data in real-time to leveraging a fluid integration with Salesforce Analytics, these capabilities are just around the corner. Join us in this roadmap session to see what the near-future of Big Data on Salesforce App Cloud looks like and how you can benefit from it.
Key Takeaways
- Learn what 100 billion+ records on the Salesforce App Cloud could actually mean to you.
- Understand new services such as AsyncSOQL that can can deliver reliable, resilient query capabilities over your sObjects and BigObjects.
-Gain insights for large scale federated data filtering and aggregation.
-Transform data movement so all your customer records are available across their life cycle.
Intended Audience
This session is for Salesforce Administrators, Developers, Architects and just about anyone who wants to learn more about BigObjects!
Unlock SAP - Release the potential of your existing backend systems with Sale...Salesforce Deutschland
When you unlock SAP with the Salesforce Platform, you can get more out of your back office data. Quickly deliver value to your company with new apps that help every department and employee be more productive, and move at the speed of the business. Learn in this session from our customer Koenig & Bauer and us how easy this is, also for your organisation.
When you unlock SAP with the Salesforce App Cloud, you can get more out of your back office data. You can quickly deliver value with new apps that help every department and employee be more productive, and move at the speed of the business. Join this session to learn about the opportunities our customers unlocked by connecting their SAP data to Salesforce. Watch the video now: https://www.youtube.com/watch?v=aOXflyCKfLI
Understanding the Salesforce Architecture: How We Do the Magic We DoSalesforce Developers
Join us for a deep dive into the architecture of the Salesforce1 Platform. We'll explain how multitenancy actually works and how it affects you as a Salesforce customer. By understanding the technology we use and the design principles we adhere to, you'll see how our platform teams manage three major upgrades a year without causing any issues to existing development. We'll cover the performance and security implications around the platform to give you an understanding of how limits have evolved. By the end of the session, you'll have a better grasp of the architecture underpinning Force.com and understand how to get the most out of it.
One of the core advantages of the Salesforce Analytics Cloud is the ability to process extremely large data sets without sacrificing performance. Building those data sets is not always easy. Vendor tools do a great job but can be expensive and, in some cases, organizations need complete control over their data transformations. In this session, we will explore the different native options available for loading data into the Analytics Cloud, and review the pros and cons of using Apex. We will demonstrate a simple data loader app built with the Analytics Cloud simple API.
Understanding Multitenancy and the Architecture of the Salesforce PlatformSalesforce Developers
Join us as we take a deep dive into the architecture of the Salesforce platform, explain how multitenancy actually works, and how it affects you as a developer. Showing the technology we use and the design principles we adhere to, you'll see how our platform teams manage three major upgrades a year without causing any issues to existing development. We'll cover the performance and security implications around the platform to give you an understanding of how limits have evolved. By the end of the session you'll have a better grasp of the architecture underpinning Force.com and understand how to get the most out of it.
Join us as we take a deep dive into the architecture of the Salesforce1 Platform, explain how multitenancy actually works, and how it affects you as a developer. Showing the technology we use and the design principles we adhere to, you'll see how our platform teams manage three major upgrades a year without causing any issues to existing development. We'll cover the performance and security implications around the platform to give you an understanding of how limits have evolved. By the end of the session, you'll have a better grasp of the architecture underpinning Force.com and understand how to get the most out of it.
Boxcars and Cabooses: When One More XHR Is Too MuchPeter Chittum
RESTful APIs have simplified backend access providing clean URL-based resource representations using standard HTTP methods such as GET and POST. But growth in the number of these APIs can lead to inefficiencies: if an app needs to access many of these resources at once, performance can bog down and user experience can suffer, especially for mobile devices.
'Boxcarring' is the bundling together of multiple HTTP requests into a single request. For example, a client framework might abstract the transport layer and bundle multiple requests into a single call. On the server, a RESTful API must be surfaced to accept a POST to multiple resource representations in one request. This talk will demonstrate both client and server side examples of boxcar requests.
Sample Gallery: Reference Code and Best Practices for Salesforce DevelopersSalesforce Developers
Exploring the code within sample applications is a great way to learn new languages, frameworks, and platforms. That’s why we built the Sample Gallery (https://trailhead.salesforce.com/sample-gallery), a collection of Salesforce Customer 360 Platform reference applications that demonstrate examples of what you can create and how to build it. In this interactive webinar, we introduce you to a few Sample Apps and show you how to make the best use of them in your day-to-day development projects.
Maximizing Salesforce Lightning Experience and Lightning Component PerformanceSalesforce Developers
We all want the Salesforce Lightning Experience to be fast- but how do we define fast, and how do we make it even faster? When you’re building a UI, everything you add to the page affects performance, and to make load times faster and perform the way users expect, we need to treat speed as an essential design feature. In this presentation, you’ll learn how to measure performance, learn a few tips on how to maximize performance, and take responsibility for your feature’s performance from design to production.
Last year was eventful for Salesforce Developers - we started with the launch of Lightning Web Components (LWC), open-sourced it, enabled local development, and ended the year by open-sourcing Base Lightning Components. In this webinar, we will explore exciting new developments within Base Components and we will show you how to use open-source Base Components to build engaging applications faster with local development.
In this session we will,
- Spin up a local development environment to build Lightning web components
- Use and customize the base components and recipes to build pages and apps quickly
- Explore the latest features of VS Code developer tooling while coding for a use case
Over the past two months, we’ve announced many new resources for developers at Dreamforce and TrailheaDX India. To learn all about them, watch this video, where we'll explore live demos showcasing the latest updates for Lightning Web Components (LWC), Einstein, Heroku, and a lot more on the Customer 360 Platform.
In this session we,
- Explore key highlights from TrailheaDX India
- Show live demos of generally available features
- Explain how you can benefit from these features
TrailheaDX (TDX) is coming to ‘namma’ Bengaluru in India on Dec 19th and 20th! TrailheaDX India is the conference for everyone who builds on and customises Salesforce — including admins, developers, architects, and partners. This event will have sessions, demos and fun for those just getting started with the platform, as well as for advanced admins, architects and developers.
You might have questions about the event - and to answer your questions Kavindra Patel, known as the father of the Indian Salesforce Community, joins Shashank Srivatsavaya, Head of APAC Developer Relations. Register for our exclusive webinar to:
- Get a sneak peek into exclusive sessions and activities
- Find out who you shouldn’t miss at TrailheaDX India
- Understand what comes with your #TDX19 registration
CodeLive: Build Lightning Web Components faster with Local DevelopmentSalesforce Developers
GitHub repo: https://github.com/satyasekharcvb/lwc-local-dev.git
With the release of a new beta version of Local Development, you can now build Lightning web components faster than ever before! You can now render changes, iterate rapidly, troubleshoot errors, and even connect with data from your org by spinning up a local development server on your machine.
In this session, we build Lightning web components in real time. The exciting new capabilities we showcase will enable you to be an even more productive developer.
In this CodeLive session we:
- Spin up a local development server from the CLI to rapidly edit and view components
- Observe how a rich error handling experience simplifies testing and debugging
- Learn how to proxy data from an org for more context and fine-tuned development
CodeLive: Converting Aura Components to Lightning Web ComponentsSalesforce Developers
GitHub repo: https://github.com/adityanaag3/aura2lwc
Lightning Web Components (LWC) give any JavaScript developer a modern, web standards-based path to building apps and experiences on the Salesforce Platform.
In this live coding session, you’ll learn how to first evaluate if an existing Aura Component needs to be converted, then convert it to LWC using the latest features of Salesforce Extensions for VS Code. We demonstrate conversion to LWC by walking through various components of a real world Aura Component - including input and output, Tables, Forms, and more.
In this live coding session we:
- Convert Aura Components to LWC
- Leverage the latest IDE features
- Share implementation best practices
Earlier this year, we released Lightning Web Components (LWC), a new UI framework based on web standards and optimized for performance and developer productivity. We have now open sourced the Lightning Web Components framework so that anyone can build applications on any platform.
Join our webinar where we'll explore how this framework, based on standard HTML, modern JavaScript (ES6+), and the best of native Web Components, helps you create web components and apps using the stack and tools you prefer.
We recently announced over 300 new features and enhancements at TrailheaDX '19 and for the Summer '19 release. In our Developer Highlights webinar we explore the top features with in-depth demos, including the latest updates for Lightning Web Components (LWC), LWC Open Source, CLI updates, Change Data Capture for external objects, Asynchronous Apex Triggers, Notification Builder and more.
Watch the recording here: https://youtu.be/kmOkk74QiCo
You can watch a recording of the live coding session at https://sforce.co/2IbU3yJ
In the fifth and final webinar of our five part series, we learn how to build Lightning web components for different use cases. We explore how to leverage all of the concepts you’ve learned in this webinar series to create modern apps with Lightning Web Components. We also demonstrate various ways to migrate Aura components to Lightning web components for a seamless transition.
In the fourth episode of our five part series on Lightning Web Components, we show you how static resources and custom JavaScript are used with Lightning Web Components. You’ll learn how to use external APIs in conjunction with Lightning Locker to secure your JavaScript code. And finally, we’ll demonstrate how you can test your Lightning Web Components using Jest.
LWC Episode 3- Component Communication and Aura InteroperabilitySalesforce Developers
Lightning Web Components gives JavaScript developers a web standards-based path to building apps on the Salesforce Lightning Platform.
In the third episode of our five part series on Lightning Web Components, we cover design considerations and compositions of Lightning Web Components.
Lightning Web Components give any JavaScript developer a modern, web standards-based path to building apps and experiences on the Lightning Platform.
In the second episode of our five part webinar series on Lightning Web Components, we will show you how to access data from the Salesforce org, get record data, create records and handle errors using Lightning Web Components. In this webinar you’ll learn all about Lightning Data Service, reactive wire service and Lightning App Builder.
Lightning Web Components are a JavaScript programming model for building web applications and interfaces that is built on the best of web standards.
Any front-end web developer can get started with LWC and leverage custom elements, and JavaScript APIs to create modern apps and experiences that are fully aware of Salesforce data and processes.
Migrate legacy Salesforce CPQ to Advanced Calculator with the help of JSQCP. With Advanced Calculator, you can boost the runtime of quote creation. Join this webinar to learn what is required for migration - we’ll also walk through JSQCP.
Replicate Salesforce Data in Real Time with Change Data CaptureSalesforce Developers
Migrate your batch processing, scheduled ETL, and nightly workloads to event-driven, real-time integrations using Change Data Capture. CDC means data change events are published to an event stream, allowing businesses to have up-to-date information across systems and applications. Join us to learn how to configure Change Data Capture and subscribe to the stream of change events, streamlining your architectures and processes.
Using Salesforce DX and its suite of offerings can significantly increase your development productivity.
Join this webinar to learn more about source control, scratch orgs, CLI, Metadata Coverage, VS Code, and Unlocked Packages. We will even do a live demo on continuous delivery using Salesforce DX.
Lightning Flow makes it easier for developers to build dynamic process-driven apps with Process Builder and the new Flow Builder. Join us and learn more about how you can get in the Flow!
Integrate CMS Content Into Lightning Communities with CMS ConnectSalesforce Developers
In this webinar we will show you how to display content from external Content Management Systems into Lightning Communities. This lets you take advantage of the Communities framework and help you centralize content in whatever CMS system you have chosen.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
Processing Big Data At-Scale in the App Cloud
1. Processing Big Data At Scale
Naren Chawla
Senior Director, Product Management (nchawla@salesforce.com)
Prashant Kommireddi @prashant1784
Leverage platform-native Data Pipelines for ETL
2. Safe Harbor
Safe harbor statement under the Private Securities Litigation Reform Act of 1995:
This presentation may contain forward-looking statements that involve risks, uncertainties, and assumptions. If any such uncertainties materialize or if
any of the assumptions proves incorrect, the results of salesforce.com, inc. could differ materially from the results expressed or implied by the forward-
looking statements we make. All statements other than statements of historical fact could be deemed forward-looking, including any projections of
product or service availability, subscriber growth, earnings, revenues, or other financial items and any statements regarding strategies or plans of
management for future operations, statements of belief, any statements concerning new, planned, or upgraded services or technology developments and
customer contracts or use of our services.
The risks and uncertainties referred to above include – but are not limited to – risks associated with developing and delivering new functionality for our
service, new products and services, our new business model, our past operating losses, possible fluctuations in our operating results and rate of growth,
interruptions or delays in our Web hosting, breach of our security measures, the outcome of any litigation, risks associated with completed and any
possible mergers and acquisitions, the immature market in which we operate, our relatively limited operating history, our ability to expand, retain, and
motivate our employees and manage our growth, new releases of our service and successful customer deployment, our limited history reselling non-
salesforce.com products, and utilization and selling to larger enterprise customers. Further information on potential factors that could affect the financial
results of salesforce.com, inc. is included in our annual report on Form 10-K for the most recent fiscal year and in our quarterly report on Form 10-Q for
the most recent fiscal quarter. These documents and others containing important disclosures are available on the SEC Filings section of the Investor
Information section of our Web site.
Any unreleased services or features referenced in this or other presentations, press releases or public statements are not currently available and may not
be delivered on time or at all. Customers who purchase our services should make the purchase decisions based upon features that are currently
available. Salesforce.com, inc. assumes no obligation and does not intend to update these forward-looking statements.
3. Topics
Big Data Processing Problem and Proposed Solution
Data Pipeline Deep-dive
Demo
Key Use-cases
Customer Stories
Summary
Q&A
4. Problem
ERP
HCM SCM Logs
1. Acquire & Store Data
2. Prepare Data
(Cleanse, Augment, Transform, Join)
Data Lake /
EDW
4. Take Action
Customer Success Platform
3. Analyze
Wave
Firewall
• Cost and complexity of
managing external data
platforms
• Slow time-to-value, poor
support for ad-hoc
analysis
• Inability to deliver high-
value packaged analytic
solutions
5. Solution
ERP
HCM SCM Logs/Machine Data
4. Take Action
Salesforce Apps
3. Analyze
Wave
Firewall
• Greater ease-of-use,
consistent end-to-end
experience
• Greater flexibility and
faster time-to-value
• Packaged Analytic
Solutions
2. Prepare Data
Data Pipelines / Async Query
1. Acquire & Store Data
BigObjects
6. Data Pipelines Overview
Currently in Pilot
Data Pipelines
Programmatic language based on
Apache Pig plus whitelisted UDF
libraries (Piggybank, DataFu)
Multi-tenancy resource management,
scheduling, job monitoring and
management
Data Sources Data Targets
SObjects
BigObjects
Wave Data Sets
External Objects
Files
Archive Objects
SObjects
BigObjects
Wave Data Sets
External Objects
Files
Archive Objects
Generate mapReduce Jobs
Hadoop
Big Data Processing
9. BigObjects vs. SObjects
SObjects BigObjects
Use cases CRM transactional data Read-only immutable data
Data volumes <50m Rows Billions of Rows
Field types All Types Strings, numbers, dates, json
Query Real Time Query Response Blend of real time and asynchronous query response
determine by size of result set
Transactions ACID transactions Record Level Consistency
Access Management Full Sharing User Permissions and Field-level Security
APIs Full Support SOQL, Async Query, Data Pipelines
Triggers Full Support None
Reports Full Support Limited CRTs
Search Full Support None
12. Key Use Cases
Big Object
Ext Object
Files
sObject
Wave
sObject
sObject
Native Big Data Processing
Data Prep for Descriptive
Analytics
Data Enrichment to turn “Insight into
Actions”
Big Object
Ext Object
Files
sObject
Wave
sObject
Handling Semi-structured Data
JSON, HTML, XML and
other complex semi-
structured data...
13. Customer Stories
Gamification - based on
experience points update user
levels
Computing Partner
Scorecards
Asset Management Analytics Analytics
Large volume data processing
(250M + records). Trawl the
rewards and update user-
objects. Later, will like to use
analytics.
Scorecard determines
status which in turn
determines pricing,
resources that partners
have access to assist in
sales.
Calculated multiple times
every week for Partner
Accounts (70h+).
Account assignment at account/
office/contact levels.
Will like to run daily
Correlate game-play data
with customer interaction
to improve customer
retention, loyalty, etc.
Multi-org consolidation;
White-space analysis.
18. Summary & Next-Steps
Why Data Pipeline?
● Massive Parallelism (10-40X performance improvement)
● Overcome governor limits
● Work towards Data Lake Architecture
● Reduce complexity/cost - 100% Platform-Native
Resources
● Implementation Guide - http://docs.releasenotes.salesforce.com/en-us/summer15/release-notes/
rn_forcecom_data_pipelines.htm
Join the Pilot Program
Any questions: nchawla@salesforce.com
20. And make any adjustments needed before loading.
FUTURE
21. BigObjects
External
SObjects
• New object type optimized for extremely large row-count
• Use cases: read-only data from external systems, point-of-sale
data, connected product event data, clickstream data, etc.
• Backed by HBase as a System of Record
• Integrated into platform via External sObject framework,
Phoenix, Pliny
HBase
Phoenix
SQL
Pliny
SOQL
Platform
22. Data Pipelines Overview
Data Pipelines
Programmatic language based on
Apache Pig plus whitelisted UDF
libraries (Piggybank, DataFu)
Declarative tooling
for admins and
analysts
WaveDevConsoleSetup
Multi-tenancy Hadoop, resource
management, scheduling, job monitoring
and management
Data Sources
Data
Targets
SObjects
BigObjects
Wave Data Sets
External Objects
Files
Archive Objects
SObjects
BigObjects
Wave Data Sets
External Objects
Files
Archive Objects
Data Set
Objects
Snapshot for
provenance
tracking
Generate Data Pipelines
Generate mapReduce Jobs
Data Processing
Data Set
Objects
Snapshot for
provenance
tracking
Remove Data Sets Object
Declarative Tooling - bring it
later
23. Customer Name Brief Description Use-cases
Cloud App CloudApps increases
organisational performance by
enabling, encouraging,
enhancing and measuring
behavioural change using
gamification
Large volume data processing (250M + records). Trawl the rewards and update user-objects. Later,
will like to use analytics.
EMC Computing Partner Scorecards Business Partner scorecards help partners track whether they qualify for a particular Partner Tier
status (Gold, Silver, Platinum). Tier status determines pricing, resources that partners have access
to assist in sales. Scorecards are calculated multiple times every week for Partner Accounts. This
takes 70h+ to calculate. When being processed Scorecards are zero'ed out and a Partner cannot
not see the details of why they are in a certain status. In order to process them in a shorter window
(~10h), they've reduced the total number of Partner Accounts that qualify for the Business Partner
program from 22K to 780.
Legg Mason Asset Management Legg Mason has built an internal process to updates account assignment at account/office/contact
levels. They will like to do this more frequently, but async batch apex process is causing them to hit
several limits and preventing them to run this process daily.
Activision Video Game Developer Activision want’s to correlate game-play data with customer interaction to improve customer
retention, loyalty, etc. Currently, they load game-play data every 2 weeks, they will like to do that
daily. Plus, use Pipeline to join game play data with Case records and use Analytics to drive insight
(for example, impact of service issue on gaming behaviour)
Financial Force ERP on Platform FF gets files in emails and they have to do manual downstream processing to generate invoices, etc
based on this incoming files. They want to leverage Pipelines to scale and automate some steps
USPS Business Transformation USPS wants to combine CRM data with external data (from Equifax) to marry physical address with
digital identity for a user. They expect 500 million external records. And they will build
transformational applications based on this data (For ex, twitter handle on envelopes, Uber for
Customer Stories
24. Data Pipelines Roadmap (WORK ON THIS SLIDE)
- Spark for internal customers
- Wave connectors
- Better error handling
- Monitoring improvements
- Basic limits
198
Winter ’16 / DF15
- Resource management
- Scheduler
- Performance / optimization
- Hardening
200
Spring 16
- Metadata API
- Simple Monitoring
- Dev Console integration
- Logging improvements
- Deployment to HBase servers
196
Summer ‘15
Pilot II Pilot III
GA
(stretch goal)
25. Salesforce.com
Confidential
External
SObjects
BigObjects
• New object type optimized for extremely large row-count
• Targeted functionality
• Use cases: read-only data from external systems, point-of-sale
data, connected product event data, clickstream data, etc.
• Backed by HBase as a System of Record
• Integrated into platform via External sObject framework,
Phoenix, Pliny
HBase
Phoenix
SQL
Pliny
SOQL
Platform
18
6
2.4
26. Salesforce.com
Confidential
BigObjects vs. SObjects
SObjects BigObjects
Use cases CRM transactional data Write-once / Read-only data from external systems, point-of-
sale data, connected product event data, clickstream data,
etc
Data volumes <50m Rows Billions of Rows
Filed types All Types Strings, numbers, dates
Query Realtime query response Blend of real time and asynchronous query response
determine by size of result set
Transactions ACID transactions Eventually consistent
Access Management Full Sharing Object Perm Based, Sharing Descriptors in future
APIs Full Support REST, SOQL, Bulk
Triggers Full Support None
Reports Full Support Limited CRTs
Search Full Support None