The document discusses spend classification and provides definitions and challenges related to spend classification including categorization at source, inconsistent categorization, multiple disparate taxonomies, and classifying spend into miscellaneous categories. It also discusses taxonomy, standard vs custom taxonomies, and how machine learning such as named entity recognition can help with spend classification.
Part of the MaRS BioEntrepreneurship event series
Speaker: Wendy Hill, Gap Strategies
This event is available as an audio file:
http://www.marsdd.com/bioent/feb12
Efficient Point Cloud Pre-processing using The Point Cloud LibraryCSCJournals
Robotics, video games, environmental mapping and medical are some of the fields that use 3D data processing. In this paper we propose a novel optimization approach for the open source Point Cloud Library (PCL) that is frequently used for processing 3D data. Three main aspects of the PCL are discussed: point cloud creation from disparity of color image pairs; voxel grid downsample filtering to simplify point clouds; and passthrough filtering to adjust the size of the point cloud. Additionally, OpenGL shader based rendering is examined. An optimization technique based on CPU cycle measurement is proposed and applied in order to optimize those parts of the pre-processing chain where measured performance is slowest. Results show that with optimized modules the performance of the pre-processing chain has increased 69 fold.
What is data mining? The process of analyzing data to discover hidden patterns and relationships that can help you manage and improve your business.
Check out: www.eleaderstochange.com
Follow #eleaders2change
Who Does What, When, and How for a Divestiture?eprentise
Businesses that are going through a divestiture need to consider several key questions: What are we selling, and why? What data belongs to me, and how do I find it? As my business changes by the rules, how do I achieve agility? What are my options for divesting? This webinar will address potential business and IT issues affecting organizations that go through the divestiture process. The presentation will touch on key insights and outline approaches that can be used by both business and IT organizations to ensure success.
Part of the MaRS BioEntrepreneurship event series
Speaker: Wendy Hill, Gap Strategies
This event is available as an audio file:
http://www.marsdd.com/bioent/feb12
Efficient Point Cloud Pre-processing using The Point Cloud LibraryCSCJournals
Robotics, video games, environmental mapping and medical are some of the fields that use 3D data processing. In this paper we propose a novel optimization approach for the open source Point Cloud Library (PCL) that is frequently used for processing 3D data. Three main aspects of the PCL are discussed: point cloud creation from disparity of color image pairs; voxel grid downsample filtering to simplify point clouds; and passthrough filtering to adjust the size of the point cloud. Additionally, OpenGL shader based rendering is examined. An optimization technique based on CPU cycle measurement is proposed and applied in order to optimize those parts of the pre-processing chain where measured performance is slowest. Results show that with optimized modules the performance of the pre-processing chain has increased 69 fold.
What is data mining? The process of analyzing data to discover hidden patterns and relationships that can help you manage and improve your business.
Check out: www.eleaderstochange.com
Follow #eleaders2change
Who Does What, When, and How for a Divestiture?eprentise
Businesses that are going through a divestiture need to consider several key questions: What are we selling, and why? What data belongs to me, and how do I find it? As my business changes by the rules, how do I achieve agility? What are my options for divesting? This webinar will address potential business and IT issues affecting organizations that go through the divestiture process. The presentation will touch on key insights and outline approaches that can be used by both business and IT organizations to ensure success.
enVista’s Supply Chain Solutions practice provides a unique perspective on your supply chain execution performance through an Operations and Technology Performance Assessment.
Our expert consultants can assess and quickly benchmark your operations against best in class performers. We have the experience and objectivity to spot quick hit tactical changes that deliver immediate productivity increases.
Oracle PIM: Phantasmal Item Descriptions in your OrganizationAXIA Consulting Inc.
One of the top US-based manufacturers in the material handling space had item descriptions were inconsistent, vague, duplicate, incomplete and many times incomprehensible. This presentation accentuates AXIA’s journey on item description standardization by methodical creation of parts taxonomy & classification, schema structure and enrichment of user defined attributes to generate item descriptions by leveraging the attribute values and functions in Oracle Product Information Management. It also includes business considerations and best practices that were employed by AXIA to achieve success.
What is Internal Analysis?
The process of identifying and evaluating an organization’s specific characteristics
Resources, capabilities, and core competencies
Looks at organization’s
Current vision
Mission(s)
Strategic & financial objectives
Strategies
Moyez Dreamforce 2017 presentation on Large Data Volumes in SalesforceMoyez Thanawalla
As enterprises continue to push more or their data to the cloud, Salesforce has seen data volumes in its tenant orgs grow at an exponential rate. How do you manage such volumes efficiently? How do you build queries and reports that respond in a timely manner?
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
2. Acknowledgements
Sincere thanks to
Keshava Rangarajan,
Chief Architect, Halliburton Corporation
for all the contribution and guidance, without which this research would not
have been possible.
3. What is Spend Classification ?
•Definition: Process of determining a purchase code for each spend record
(Requisitions, Purchase Orders, Receipts, Invoices, etc.) from a hierarchical
structure (Taxonomy).
Requisitions, POs, Receipts, Invoices, etc.
4. Why to classify spend ?
•Once all spend transactions are classified with a standard code from a
taxonomy – simple queries can be answered like
•What are my top 10 spend categories ?
•What is my travel spend ?
•What is my spend for a given Supplier ?
•What is my spend for a given Part ?
•What is my spend for a given Business Unit ?
•If your classification is done on a consolidated data across all systems in your
organization, you get visibility across all systems with classification.
5. What is Taxonomy ?
•A simple hierarchical level of coding structure used to classify spend at
different levels.
Segment
Family
Class
Commodity
6. What is the Spend Classification challenge ?
•Categorization at source
•Categorization itself is inconsistent or missing completely
•Multiple disparate Taxonomies may exist in a company
•Classifying into “MISCELLANEOUS” category
•No standardization of Taxonomies
7. What is the “Categorization at source” challenge ?
Exercise: Buying a work laptop and expensing via procurement
X Category: Facility. Building.Hardware
Category: IT.Hardware.Laptop
Characteristics:
•User entered, hence error-prone
•No standardization across the supply chain – business units, customers, or
suppliers.
8. What is the “inconsistent/missing
Categorization” challenge ?
• Category: IT.Hardware.Laptop
• Category: IT.Hardware.Computers.Laptop
9. What is the “multiple disparate Taxonomies”
challenge ?
•Multiple (and disparate) taxonomies may also exist in the organization
where classification could be carried out business unit-wise without regard
to, or referring to, the taxonomies used in other business units.
Business Unit 3
Business Unit 2
Business Unit 1
Taxonomy 3
Taxonomy 2
Taxonomy 1
10. What is the “MISCELLANEOUS category” challenge ?
•Spend transactions are classified into the 'Miscellaneous' category, making it
very difficult for business analysts to figure out which category the item
should actually belong to.
•Spend analytics data will then show a weighted 'Miscellaneous' category,
which is incorrect and thus does not reflect a true picture of spend by
categories for the organization.
•Similar popular categories: OTHERS, UNCATEGORIZED
11. What is the standardization of Taxonomies need ?
•An enterprise may have multiple taxonomies at different levels – corporate,
strategic, business unit and regional center.
•Multiple taxonomies at various levels creates a number of issues when
analyzing spend, therefore it is important to create or use standard
taxonomies across the enterprise.
12. What are the types of Spend Classification
Taxonomies ?
SPEND
CLASSIFICATION
TAXONOMY
Standard
Custom
13. Standard Taxonomies
•UNSPSC: United Nations Standard Products and Services Code. It is 5 level
hierarchy coded as an 8-digit number.
Example:
•Segment 44. Office Equipment and Accessories and Supplies.
• Family 10. Office machines and their supplies and accessories.
• Class 15. Duplicating machines.
• Commodity 01. Photocopiers.
• Business Function 14. Retail.
14. Custom Taxonomies
•If your own coding structure is strong enough for your business, or you think
your business is more acquainted with your own structure
15. 1) Requisitions ERP Category
2) Purchase Orders
3) Receipts
4) Invoices
Procurement & Spend Analysis
Item Invoice Categories Supplier
Description Description Description
And Attribute And Attribute And Attribute
ERP Taxonomy UNSPSC Code Custom
Taxonomies
Data
Mining
Spend Classification
16. What is Spend Analysis?
•Process of collecting, cleansing, classifying and analyzing expenditure data
with the purpose of reducing procurement.
•Process of aggregating, classifying, and leveraging spend data for the purpose
of gaining visibility into cost reduction, performance improvement, and contract
compliance opportunities.
•Enables to answer the following questions:
•Who is buying ?
•What ?
•From whom ?
•When ?
•(optionally) Where ?
•At what price ?
17. Who needs Spend Analysis?
•It is the process of organizing a company’s spend in such a way that one
understand it, slice it, dice it and uncover hidden savings opportunities.
•Impacts more than just the sourcing team
•Spend analysis/ visibility serves three internal user community groups:
•Leadership and CxOs: who need up-to-date reports to drive strategic direction
•Managers, accountants: who need to drill down into a spend data set to explore specific areas
of interest or track down payment specifics
•Sourcing power users: who need to locate, drive, and monitor the next set of savings initiatives
18. What is Spend Management?
•Process in which companies control and optimize the money they spend.
•Involves cutting operating and other costs associated with doing business.
•Includes spend analysis, sourcing, procurement, receiving, payment settlement and
management of accounts payable and general ledger accounts.
•In an enterprise, spend management is managing how to spend money to best
effect in order to build products and services.
•Encompasses processes such as outsourcing, procurement, e-procurement, and supply chain
management.
19. Benefits of Spend Management
•Decreasing "maverick" spend
•Increase of spend economies of scale
•Strategic sourcing (also called "supplier rationalization")
•Sourcing optimization
•Co-operative sourcing
•Increase process efficiencies
•Increase procurement efficiency
20. Life cycle of a PO
Create PO
1
Add items to PO
2
Add PO to Cart *
3
Create Document for the PO in the Cart
4
Create Requisition for the Document
5
Note: PO needs to be classified before it hits the Cart. After the Order
hits the Cart, then it is too late for classification.
21. Classifying Spend
• We have a set of pre-defined fields chosen for classification from a Purchase
Order. All these fields are concatenated to form one giant string. (Note:
This textual string could have multi-lingual strings.)
• Lexers can be used for detecting languages. (eg: Auto lexers, World lexers)
• SVM could be used for Textual mining.
22. Where does Machine Learning fit in?
(Spend Auto-Classification)
Ontology (including Spend
Descriptions + other textual
attributes) Taxonomies
Spend transaction
Spend
Auto-classifier
Linguistics (UIMA) +
Neural Net Engine/
Text SVM
Auto-Classified
Spend
23. Training data set
• To begin with, customers provide a Training data set. This is from their
historic data. They take some well known data set from their most common
use cases. This would constitute a good representation of their problem.
• We run our logic against this training set and get the results. The results are
verified. We iterate this for some cycles to tune the logic.
• Repeat the same over other use cases.
24. Data Mining Model
Create a Model
Model created
Enrich/Re-train
Cleanse incorrect classification
Support new categories (if needed)
25. What is Named Entity Recognition ?
•“Named-entity recognition (NER) (also known as entity identification and
entity extraction) is a subtask of information extraction that seeks to locate
and classify atomic elements in text into predefined categories such as the
names of persons, organizations, locations, expressions of times, quantities,
monetary values, percentages, etc.” -- Wikipedia
•Most research on NER systems has been structured as taking an
unannotated block of text, such as this one
• Jim bought 300 shares of Acme Corp. in 2006.
•And producing an annotated block of text, such as this one:
• <ENAMEX TYPE="PERSON">Jim</ENAMEX>bought<NUMEX
TYPE="QUANTITY">300</NUMEX>shares of<ENAMEX TYPE="ORGANIZATION">Acme
Corp.</ENAMEX> in <TIMEX TYPE="DATE">2006</TIMEX>.
26. Anatomy of a query …
Query = “Find Approved Status POs with High
Amount”
27. Stemmed Entity Recognition & Linguistic
Parsing yields…
Search Verb:
“Find”
Find Approved Status POs with High Amount
28. Stemmed Entity Recognition & Linguistic
Parsing yields…
Search Verb:
“Find”
Attribute:Status= “Approved”
Find Approved Status POs with High Amount
29. Stemmed Entity Recognition & Linguistic
Parsing yields…
Search Verb: Entity:
“Find” Attribute:Type=“PO”
Attribute:Status= “Approved”
Find Approved Status POs with High Amount
30. Stemmed Entity Recognition & Linguistic
Parsing yields…
Search Verb: Entity:
“Find” Attribute:Type=“PO”
Attribute:Amount= “High”
Attribute:Status= “Approved”
Find Approved Status POs with High Amount
31. Stemmed Entity Recognition & Linguistic
Parsing yields…
Search Verb: Target Entity:
“Find” Attribute:Type=“PO”
Attribute:Amount= “High”
Attribute:Status= “Approved”
Find Approved Status POs with High Amount
32. Stemmed Entity Recognition & Linguistic
Parsing yields…
Search Verb: Target Entity:
“Find” Attribute:Type=“PO”
Having
Attribute
Attribute:Amount= “High”
Attribute:Status= “Approved”
Find Approved Status POs with High Amount
33. Stemmed Entity Recognition & Linguistic
Parsing yields…
Search Verb: Target Entity:
“Find” Attribute:Type=“PO”
Having Having
Attribute Attribute
Attribute:Amount= “High”
Attribute:Status= “Approved”
Find Approved Status POs with High Amount
36. OWL:
attribute: string
Transaction Party
has a Code
OWL:class has OWL:class
many
Role OWL:class
has an plays OWL:
Is A
Bank
attribute: string
is related
ID to
Person Corporation
OWL:class Is A
OWL:class
OWL:attribute: Finance
number Corporation
OWL:class
has has
First has Name ID
has many many
Name OWL:class OWL:
Address attribute: string
Last has an OWL:attribute:
Account number
Name
OWL:attribute: has an in
ID
string
OWL:attribute: has
number
Door Street City State Zip Country
Number Name
OWL: OWL: OWL: OWL: attribute:
OWL: OWL: attribute:string
attribute: string attribute: string attribute: string attribute:string string
37. Transaction
ID:200911071234
has Party
has ID: SBK
has Role: S? Bank Role
played by
Bank
has Name: Bank Of Congo
has
many Address
has Street Name: Afrique Au Congo
has Country: RDC
38. Transaction
ID:200911071235
has Party
has ID: ORP
has Role: Ordering Party Role
played by
Person
has First Name: John
has Last Name: Doe
has
many Address
Account has City: Kinshasa
has Account Id: 123456 has Country: CD
in Bank
has Name: Bank Of Congo
39. Transaction
Transaction
ID:200911071234
ID:200911071235 has
is related
Party to
Party
has has ID: ORP has ID: SBK
has Role: Ordering Party Role has Role: S? Bank Role
played by
Person
has First Name: John played by
has Last Name: Doe Address
has City: Kinshasa Address
has Country: CD has Street Name:
has Afrique Au Congo
many has Country: RDC
Account
has Account Id: 123456
in Bank has
has Name: Bank Of Congo many
40. A possible solution: Pipelining approach
•Flow 1:
•Machine learning Pipeline: Input data is directly fed to the Machine Learning piece.
•Flow 2:
•Domain Ontology Pipeline: Input data is fed to a Domain Ontology.
•Standardize the output from the Domain Ontology.
•Machine learning Pipeline: Feed it into the Machine Learning piece.
•Flow 3:
•NER Pipeline: Input data is fed to a NER.
•Domain Ontology Pipeline: Output from the NER is fed to the Domain Ontology.
•Standardize the output from the Domain Ontology.
•Machine learning Pipeline: Feed it into the Machine Learning piece.
•Note:
•Domain Ontology and NER Pipelines can be optionally turned on or off
49. SVM Steps
1.Identify taxonomy (hierarchical or flat) to be classified against
2.Identity representative training data that has been classified to this taxonomy
3.Run training data against blank SVM model and the given taxonomy
4.Classify training data as per required taxonomy
5.Classify the data
6.Increase training population and enrich classification model
7.Recognize and realign impact of original model against fresh training data
8.Classify (manually) misclassifications into proper taxonomy nodes
9.Run step 6 through 8 until all the variations for a given domain have been recognized
10.Introduce live data
11.Repeat steps 4 and 5 for misclassifications
12.Store the result in a relational database
13.Insert data in an Ontology
14.Enable analysis using RQL or SPARQL
50. Open source software
1.Jena
2.Pentaho http://www.pentaho.com/
3. Stanford NER, http://nlp.stanford.edu/software/CRF-NER.shtml
4.Annie NER
5.GATE
6.UIMA
7.SVM, http://en.wikipedia.org/wiki/Support_vector_machine
We will talk about Auto-Classification and the place for Machine Learning . When a Spend transaction is added, what needs to happen is, the positioning of a spend in terms of a formal taxonomy might have to be dynamically changed. And that is not something that a person can manually do it in real time. We need an automated way of doing that. The spend transactions themselves have descriptions . When a tagging activity happens, when a review is written up , there is textual information. We could use UIMA, to pick out all the textual tokens – break them out into attributes and do Named Entity recognition. And then bring out a trained SVM engine which works on a model, that is able to pick up all the spend descriptions, and all its attributes from the Classification model, and tag it, and then position it appropriately in the Taxonomy. There are two flavors available: Neural Net Engine SVM They both have comparable performance. The bottom line is, we took in the spend Taxonomy , we took in the spend Ontology that describes the entire Spend model as well as the description of the spend - you can run it into a Neural Net Engine and then you can tag things, so that, as and when a new spend transaction is introduced, it is appropriately positioned in the Taxonomy, dynamically .