This document describes Charles Taylor InsureTech's Tide platform for handling delegated authority data. Tide standardizes and validates data from various sources like brokers, coverholders, and TPAs. It processes bordereaux and XML files, applying rules to map data fields and resolve issues. Tide aims to give all parties real-time access to risk data while reducing the reporting burden through features like an intuitive bordereau submission portal and automated validation checks.
Do you need a startup advisory board?
If you are a founder who has been asking yourself this question, check out this presentation from Adam Quinton, Founder/CEO of Lucas Point Ventures and Early Growth Financial Services (EGFS).
Topics include:
- How to build your startup advisory board
- Finding the right key advisors
- Board compensation best practices
- Structuring an effective board
- and more!
Collaborating with Direct Spend Suppliers in the Life Sciences Industry - 56556SAP Ariba Live 2018
Join this session to learn how a medical devices OEM collaborates with suppliers and contract manufacturing partners on key supply chain processes, including planning and product quality. Learn how to improve productivity and visibility. Hear about the collaboration and onboarding journey, wins, and lessons learned.
Do you need a startup advisory board?
If you are a founder who has been asking yourself this question, check out this presentation from Adam Quinton, Founder/CEO of Lucas Point Ventures and Early Growth Financial Services (EGFS).
Topics include:
- How to build your startup advisory board
- Finding the right key advisors
- Board compensation best practices
- Structuring an effective board
- and more!
Collaborating with Direct Spend Suppliers in the Life Sciences Industry - 56556SAP Ariba Live 2018
Join this session to learn how a medical devices OEM collaborates with suppliers and contract manufacturing partners on key supply chain processes, including planning and product quality. Learn how to improve productivity and visibility. Hear about the collaboration and onboarding journey, wins, and lessons learned.
VC Bootcamp By DFJ Gotham Ventures and Wilson Sonsini Goodrich & RosatiMark Davis
Slides from the Venture Capital Bootcamp event hosted by DFJ Gotham Ventures and Wilson Sonsini Goodrich & Rosati at Columbia University on June 3, 2009. A video of the 3 hour event is available at www.dfjgotham.com.
Spring Cloud: API gateway upgrade & configuration in the cloudOrkhan Gasimov
In this presentation we walk through features of Spring Cloud Gateway and Spring Cloud Config projects, overview new features provided by Spring Cloud Gateway including advanced routing options for API services supporting parallel APIs in several versions, discuss code examples and configuration options. Once API gateway is deployed, we don’t want to redeploy it on configuration changes as well as redeploy other services upon configuration updates. And this is where Spring Cloud Config enters the game. It allows us to keep configurations in the cloud, for example in a Git repository, and once paired with tools necessary, enables almost zero-down-time configuration updates, audit of changes and parallel configurations for different environments.
•Gain a deeper understanding of SAP S/4HANA and its benefits
•Understand how SAP S/4HANA is organized into core, products, and suite
•Learn the new capability planned for the October update
•Hear about the Gartner report on vendors in the SAP HANA solution space
•Learn about the infrastructure required to run SAP HANA
Change Data Streaming Patterns For Microservices With Debezium (Gunnar Morlin...confluent
Debezium (noun | de·be·zi·um | /dɪ:ˈbɪ:ziːəm/) – Secret Sauce for Change Data Capture Apache Kafka has become the defacto standard for asynchronous event propagation between microservices. Things get challenging though when adding a service’s database to the picture: How can you avoid inconsistencies between Kafka and the database? Enter change data capture (CDC) and Debezium. By capturing changes from the log files of the database, Debezium gives you both reliable and consistent inter-service messaging via Kafka as well as instant read-your-own-write semantics for services themselves. Join this session to learn how to leverage CDC for reliable microservices integration and solving typical challenges such as gradually extracting microservices from existing monoliths, maintaining different read models in CQRS-style architectures, and updating caches as well as full-text indexes. You’ll find out how Debezium streams all the changes from datastores such as MySQL, PostgreSQL, SQL Server and MongoDB into Kafka, and how Debezium is designed to not compromise on data correctness and completeness also if things go wrong. In a live demo we’ll show how to use Debezium to set up a change data stream out of your application’s database, without any code changes needed. You’ll see how to consume change events in other services, how to gain real-time insight into your changing data using Kafka Streams and much more.
Microservices - Death of the Enterprise Service Bus (ESB)? (Update 2016)Kai Wähner
Microservices are the next step after SOA: Services implement a limited set of functions. Services are developed, deployed and scaled independently.
Continuous Integration and Continuous Delivery control deployments. This way you get shorter time to results and increased flexibility. Microservices have to be independent regarding build, deployment, data management and business domains. A solid Microservices design requires single responsibility, loose coupling and a decentralized architecture. A Microservice can to be closed or open to partners and public via APIs. This session discusses the requirements, best practices and challenges for creating a good Microservices architecture, and if this spells the end of the Enterprise Service Bus (ESB). A live demo will show how middleware and Microservices complement each other using containers, continuous integration, REST services, and open source frameworks such as Cloud Foundry.
A live demo showed a "Microservices Middleware Architecture" using Cloud Integration (with Cloud Foundry PaaS), Integration and Services (with TIBCO BusinessWorks Container Edition), API Management / Open API (with Mashery) amd Log Management / IT Operations Analytics (ITOA, with Papertrail and LogLogic / Unity).
SAP S4HANA : Learn From Our Implementation JourneyAnup Lakra
Are you looking to streamline data models, optimize daily business transactions, accelerate strategic decision making and simplify everything in between – whether you use it on the cloud or on premise. Know about how you can prepare and simplify the SAP S4HANA adoption journey through our successful implementation of SAP S4HANA
Implementation of usb transceiver macrocell interfaceeSAT Journals
Abstract The universal serial bus (USB) transceiver macro cell interface (UTMI) is a two wire, bi-directional serial bus interface. UTMI consists of transmitting and receiving sections, in which the transmitter of the UTMI sends data to different USB devices through D+ and D- lines whereas the receiver gets data on the same lines. UTMI is one of the important functional blocks of USB controller, which can transmit and receive data to or from USB devices. There are three functional blocks present in USB controller; those are Serial Interface Engine (SIE), UTMI and Device Specific Logic (DSL). The parallel data from SIE is taken into the transmit hold register and this data is sent to transmit shift register from where the data is converted serially. This serial data is bit stuffed to perform data transitions for clock recovery and NRZI (1) encoding. Then the encoded data is sent on to the serial bus. When the data is received on the serial bus, it is decoded, bit unstuffed and is sent to receive shift register. After the shift register is full, the Data is sent to receive hold register. This presentation reveals the FPGA implementation of UTMI transmission rate providing with USB 2.0 specifications. Further UTMI has been designed by using VHDL code and simulated, synthesized and programmed to the targeted Spartan family of FPGA in the Xilinx environment. Index Terms: INTRODUCTION, USB TRANCIEVER MACROCELL INTERFACE AND DESIGN OF HARDWARE MODEL
WIRELESS HOME AUTOMATION USING PIC MICROCONTROLLER BASED ON RF-MODULEEng.Manfred Kibona
This paper presents a project-based final year project course on completion of Bachelor degree in Telecommunication engineering. The goal of this project was to provide engineering students a hands-on experience involving actual engineering design on communication circuits, printed circuit board (PCB) design, layout, fabrication, assembly, and testing.
The scope of the project was been projected in develop a RF remote control system that will be capable for controlling various electrical appliances in the vicinity of 100 meters. The control signals has to traverse wirelessly by means of modulation with radio frequency carrier signal ranging from frequency band of 385MHz to 480MHz.
Taming the regulatory tiger with jwg and smartlogicAnn Kelly
From CEOs to board members to operational managers, regulatory compliance is an ongoing concern. In a rapidly changing marketplace where complex regulations come from multiple regulatory bodies, the consequences of non-compliance can be costly to the enterprise in time, money and damage to their reputation.
JWG, a London think tank, has created RegDelta – a state-of-the-art regulatory change management platform - that allows individual stakeholders to quickly understand the impact of regulations and maintain a single source of truth for their regulatory obligations.
Hear Elliot Burgess, Head of Product and Client Services at JWG and Paul Gunstone, Sales Director at Smartlogic discuss the challenges organizations face identifying and complying with relevant regulations, JWG’s approach to taming the regulatory tiger with semantics and see a demo of the JWG RegDelta platform.
CTRM in the Cloud – Research and ReportCTRM Center
The data generated by our survey of the industry suggests that, in general, Energy and/or Commodity Trading and Risk Management (E/CTRM) buyers are increasingly open to considering alternatives to traditional “on- premises” implementation models including both SaaS and hosted in the cloud delivery. While a small, but committed, minority continue to resist anything but the traditional on-premises implementation approach, the overwhelming majority of respondents will consider SaaS/hosted in the cloud for a variety of vertical application areas in and around commodity trading.
Despite that finding, only 16% of those who responded to the survey actually utilize a SaaS or hosted in the cloud E/CTRM solution, and while the data strongly suggests a great deal of interest in the cloud for E/CTRM, it does indicate that the final procurement decision isn’t necessarily a slam-dunk in favor of the cloud. Though 54% of our respondents would consider a SaaS/hosted in the cloud alternative, there are indications that the final decision is still more likely to lean toward a traditional installation on-premises – at least for now. ComTech’s forecast growth rates of 15% per year for SaaS/hosted in the cloud solutions do seem to be reasonable but may accelerate in the future if a sufficient numbers of trading firms adopt the model, are successful with it and are willing to advocate the approach to their peers in the industry. Overall, this finding is in agreement with broader studies such as those conducted by Gartner that found that interest in cloud-based solutions is primarily in horizontal applications such as accounting, HR or billing; and that as a result of buyer concerns around integration and ability to customize, the uptake of cloud-based vertical applications like CTRM lags somewhat.
VC Bootcamp By DFJ Gotham Ventures and Wilson Sonsini Goodrich & RosatiMark Davis
Slides from the Venture Capital Bootcamp event hosted by DFJ Gotham Ventures and Wilson Sonsini Goodrich & Rosati at Columbia University on June 3, 2009. A video of the 3 hour event is available at www.dfjgotham.com.
Spring Cloud: API gateway upgrade & configuration in the cloudOrkhan Gasimov
In this presentation we walk through features of Spring Cloud Gateway and Spring Cloud Config projects, overview new features provided by Spring Cloud Gateway including advanced routing options for API services supporting parallel APIs in several versions, discuss code examples and configuration options. Once API gateway is deployed, we don’t want to redeploy it on configuration changes as well as redeploy other services upon configuration updates. And this is where Spring Cloud Config enters the game. It allows us to keep configurations in the cloud, for example in a Git repository, and once paired with tools necessary, enables almost zero-down-time configuration updates, audit of changes and parallel configurations for different environments.
•Gain a deeper understanding of SAP S/4HANA and its benefits
•Understand how SAP S/4HANA is organized into core, products, and suite
•Learn the new capability planned for the October update
•Hear about the Gartner report on vendors in the SAP HANA solution space
•Learn about the infrastructure required to run SAP HANA
Change Data Streaming Patterns For Microservices With Debezium (Gunnar Morlin...confluent
Debezium (noun | de·be·zi·um | /dɪ:ˈbɪ:ziːəm/) – Secret Sauce for Change Data Capture Apache Kafka has become the defacto standard for asynchronous event propagation between microservices. Things get challenging though when adding a service’s database to the picture: How can you avoid inconsistencies between Kafka and the database? Enter change data capture (CDC) and Debezium. By capturing changes from the log files of the database, Debezium gives you both reliable and consistent inter-service messaging via Kafka as well as instant read-your-own-write semantics for services themselves. Join this session to learn how to leverage CDC for reliable microservices integration and solving typical challenges such as gradually extracting microservices from existing monoliths, maintaining different read models in CQRS-style architectures, and updating caches as well as full-text indexes. You’ll find out how Debezium streams all the changes from datastores such as MySQL, PostgreSQL, SQL Server and MongoDB into Kafka, and how Debezium is designed to not compromise on data correctness and completeness also if things go wrong. In a live demo we’ll show how to use Debezium to set up a change data stream out of your application’s database, without any code changes needed. You’ll see how to consume change events in other services, how to gain real-time insight into your changing data using Kafka Streams and much more.
Microservices - Death of the Enterprise Service Bus (ESB)? (Update 2016)Kai Wähner
Microservices are the next step after SOA: Services implement a limited set of functions. Services are developed, deployed and scaled independently.
Continuous Integration and Continuous Delivery control deployments. This way you get shorter time to results and increased flexibility. Microservices have to be independent regarding build, deployment, data management and business domains. A solid Microservices design requires single responsibility, loose coupling and a decentralized architecture. A Microservice can to be closed or open to partners and public via APIs. This session discusses the requirements, best practices and challenges for creating a good Microservices architecture, and if this spells the end of the Enterprise Service Bus (ESB). A live demo will show how middleware and Microservices complement each other using containers, continuous integration, REST services, and open source frameworks such as Cloud Foundry.
A live demo showed a "Microservices Middleware Architecture" using Cloud Integration (with Cloud Foundry PaaS), Integration and Services (with TIBCO BusinessWorks Container Edition), API Management / Open API (with Mashery) amd Log Management / IT Operations Analytics (ITOA, with Papertrail and LogLogic / Unity).
SAP S4HANA : Learn From Our Implementation JourneyAnup Lakra
Are you looking to streamline data models, optimize daily business transactions, accelerate strategic decision making and simplify everything in between – whether you use it on the cloud or on premise. Know about how you can prepare and simplify the SAP S4HANA adoption journey through our successful implementation of SAP S4HANA
Implementation of usb transceiver macrocell interfaceeSAT Journals
Abstract The universal serial bus (USB) transceiver macro cell interface (UTMI) is a two wire, bi-directional serial bus interface. UTMI consists of transmitting and receiving sections, in which the transmitter of the UTMI sends data to different USB devices through D+ and D- lines whereas the receiver gets data on the same lines. UTMI is one of the important functional blocks of USB controller, which can transmit and receive data to or from USB devices. There are three functional blocks present in USB controller; those are Serial Interface Engine (SIE), UTMI and Device Specific Logic (DSL). The parallel data from SIE is taken into the transmit hold register and this data is sent to transmit shift register from where the data is converted serially. This serial data is bit stuffed to perform data transitions for clock recovery and NRZI (1) encoding. Then the encoded data is sent on to the serial bus. When the data is received on the serial bus, it is decoded, bit unstuffed and is sent to receive shift register. After the shift register is full, the Data is sent to receive hold register. This presentation reveals the FPGA implementation of UTMI transmission rate providing with USB 2.0 specifications. Further UTMI has been designed by using VHDL code and simulated, synthesized and programmed to the targeted Spartan family of FPGA in the Xilinx environment. Index Terms: INTRODUCTION, USB TRANCIEVER MACROCELL INTERFACE AND DESIGN OF HARDWARE MODEL
WIRELESS HOME AUTOMATION USING PIC MICROCONTROLLER BASED ON RF-MODULEEng.Manfred Kibona
This paper presents a project-based final year project course on completion of Bachelor degree in Telecommunication engineering. The goal of this project was to provide engineering students a hands-on experience involving actual engineering design on communication circuits, printed circuit board (PCB) design, layout, fabrication, assembly, and testing.
The scope of the project was been projected in develop a RF remote control system that will be capable for controlling various electrical appliances in the vicinity of 100 meters. The control signals has to traverse wirelessly by means of modulation with radio frequency carrier signal ranging from frequency band of 385MHz to 480MHz.
Taming the regulatory tiger with jwg and smartlogicAnn Kelly
From CEOs to board members to operational managers, regulatory compliance is an ongoing concern. In a rapidly changing marketplace where complex regulations come from multiple regulatory bodies, the consequences of non-compliance can be costly to the enterprise in time, money and damage to their reputation.
JWG, a London think tank, has created RegDelta – a state-of-the-art regulatory change management platform - that allows individual stakeholders to quickly understand the impact of regulations and maintain a single source of truth for their regulatory obligations.
Hear Elliot Burgess, Head of Product and Client Services at JWG and Paul Gunstone, Sales Director at Smartlogic discuss the challenges organizations face identifying and complying with relevant regulations, JWG’s approach to taming the regulatory tiger with semantics and see a demo of the JWG RegDelta platform.
CTRM in the Cloud – Research and ReportCTRM Center
The data generated by our survey of the industry suggests that, in general, Energy and/or Commodity Trading and Risk Management (E/CTRM) buyers are increasingly open to considering alternatives to traditional “on- premises” implementation models including both SaaS and hosted in the cloud delivery. While a small, but committed, minority continue to resist anything but the traditional on-premises implementation approach, the overwhelming majority of respondents will consider SaaS/hosted in the cloud for a variety of vertical application areas in and around commodity trading.
Despite that finding, only 16% of those who responded to the survey actually utilize a SaaS or hosted in the cloud E/CTRM solution, and while the data strongly suggests a great deal of interest in the cloud for E/CTRM, it does indicate that the final procurement decision isn’t necessarily a slam-dunk in favor of the cloud. Though 54% of our respondents would consider a SaaS/hosted in the cloud alternative, there are indications that the final decision is still more likely to lean toward a traditional installation on-premises – at least for now. ComTech’s forecast growth rates of 15% per year for SaaS/hosted in the cloud solutions do seem to be reasonable but may accelerate in the future if a sufficient numbers of trading firms adopt the model, are successful with it and are willing to advocate the approach to their peers in the industry. Overall, this finding is in agreement with broader studies such as those conducted by Gartner that found that interest in cloud-based solutions is primarily in horizontal applications such as accounting, HR or billing; and that as a result of buyer concerns around integration and ability to customize, the uptake of cloud-based vertical applications like CTRM lags somewhat.
Cloud computing is a model for enabling convenient, on-demand network access to a shared pool of configurable
computing resources (e.g., networks, servers, storage, applications, and services) that can be rapidly provisioned
and released with minimal management effort or service provider interaction.
Cloud computing is a model for enabling convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications, and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction.
The environment of physical energy and non-energy commodity trading and marketing has grown increasingly complex, marked by globalization bringing about rapid changes in supply and demand patterns, increased regulatory scrutiny and evolving trading and reporting rules, volatility along the entirety of the physical supply chain, and increasing uncertainty as to future price movements. In order to react to these changes quickly and appropriately, participants in these markets must increasingly rely on a sophisticated infrastructure of software and technologies to ensure a complete view of their trading positions and external market conditions that can quickly and severely impact their values. The core component of these now requisite trading and marketing technologies are energy and commodity trading and risk management (CTRM) systems.
IT Requirements (Stage 3)In addition to the functional requireme.docxpriestmanmable
IT Requirements (Stage 3)
In addition to the functional requirements for a new system, the IT requirements must also be developed. We will do this by evaluating the applicability and importance of a list of IT characteristics. The requirements that are identified as relevant to the business process will form the IT (technical) requirements for a system. In evaluating and selecting an IT system, both the functional (business) requirements and the IT requirements need to be considered. The areas that need to be considered in developing the IT requirements are listed below in the Table of IT Requirements, which is similar to the one you will complete in the Stage 3 assignment. Only a few of the rankings and explanations are filled in here, but your table will rank and explain them all. (Be sure to use the table and instructions provided in the assignment.)
Requirement
High/Medium/Low Importance or Relevance or Not Applicable (N/A)
Explanation for Ranking
Usability
Maintainability
Scalability
Medium
The system must be able to handle an increase in the number of users and devices being used for Returns during peak hours. The same system is used for inventory and in the Checkout process, so the entire system needs to be able to handle a large volume of customers if necessary. Walmart expects its business to continue to grow, and returns will grow along with the rest of the business.
Reliability/ Availability
High
In order for the expedited process to work, the system must be up and running and be available for use. It must also process the return correctly. The system must be accurate in the information it provides.
Extensibility
Portability
N/A
There is no requirement for any or all of the system to be able to operate on multiple platforms. It does not need to work with different hardware or software. It only needs to operate in the environment as designed.
Security
Information Quality
Authentication
Business Continuity Plan
Cloud Computing
Enterprise Systems (ERP, CRM, SCM)
Communications
High
The Communication Network in the Walmart store must be operational in order for the devices to interface with the system. If the network is inoperable, the Greeter must revert to the old process. The internet connection must be operational for the data to be passed to the Corporate Headquarters system.
Database, Data Warehouse, Data Mining
Business Intelligence
Transaction Processing
Decision Support
Executive Information
Business-to-Business eCommerce
Business-to-Consumer eCommerce
In this particular case, only a few of the IT considerations evaluated above are unique to the customer return process. This is because there is a large transaction processing system (POS) and inventory system already in use. This solution is merely expanding the capability by making the return process portable and increasing the number of returns that can be processed at one time with no increase in staff. It also is important to note that so ...
Arrayworks was founded with an ideology that the
only constant is change and technology should
empower organizations to continuously adapt their
solutions to match changing business demands.
Five Priorities for Quality Engineering When Taking Banking to the CloudCognizant
As banks move to cloud-based banking platforms for lower costs and greater agility, they must seamlessly integrate technologies and workflows while ensuring security, performance and an enhanced user experience. Here are five ways cloud-focused quality assurance helps banks maximize the benefits.
Blockchain with Machine Learning Powered by Big Data: Trimble Transportation ...Hortonworks
Trimble Transportation Enterprise is a leading provider of enterprise software to over 2,000 transportation and logistics companies. They have designed an architecture that leverages Hortonworks Big Data solutions and Machine Learning models to power up multiple Blockchains, which improves operational efficiency, cuts down costs and enables building strategic partnerships.
https://hortonworks.com/webinar/blockchain-with-machine-learning-powered-by-big-data-trimble-transportation-enterprise/
The environment of physical energy and non-energy commodity trading and marketing has grown increasingly complex, marked by globalization bringing about rapid changes in supply and demand patterns, increased regulatory scrutiny and evolving trading and reporting rules, volatility along the entirety of the physical supply chain, and increasing uncertainty as to future price movements. In order to react to these changes quickly and appropriately, participants in these markets must increasingly rely on a sophisticated infrastructure of software and technologies to ensure a complete view of their trading positions and external market conditions that can quickly and severely impact their values. The core component of these now requisite trading and marketing technologies are energy and commodity trading and risk management (CTRM) systems. As market complexity has increased and multi-commodity trading has become more common, CTRM solutions have had to become more sophisticated and provide a greater depth of capability in order to capture and value the unique characteristics of the multitude of physical commodities being transacted along the physical supply chain, from source to market. Given the capabilities of these CTRM systems, they do represent a significant investment for any trading or marketing organization, generally trailing only the large scale ERP solutions, like SAP, in terms of costs to purchase and implement. Allegro Development, one of the world’s largest CTRM solutions providers, engaged Commodity Technology Advisory to conduct a survey of a number of their clients to determine their views as to the value of their investment and the operational and financial impacts of deploying Allegro’s CTRM solution. This report summarizes the results of that survey and discusses the key considerations for any company seeking to develop their own assessment of the value of their CTRM technology investment via a Return on Investment (ROI) calculation.
Life and Annuity Insurance | Life Insurance Outsourcing | Life Insurance BPO ...RNayak3
WNS provides Life and Annuity insurance BPO services that enable clients to transform their processes while lowering the cost of operations. Visit- https://www.wns.com/industries/insurance/life-and-annuities
Capital market firms are making decisions on which business lines, asset classes and services to keep and operate and which ones to exit. Regulatory reform and the
clearing mandate are driving the firms to consolidate their traditional exchangetraded derivatives (Futures and Options) and OTC derivatives into a single clearing
business, even while bi-lateral, uncleared derivatives will continue to co-exist with cleared products.
BASC presentation on security and application architecturewbjwilliams3
Boston Application Security Consortium presentation on the Lattice of Application and Security Architecture seen through the unique requirements of distributed applications.
In today’s complex, competitive landscape, it is no longer enough to have a basic M2M solution in place. For more details visit https://www.calamp.com/
2. Charles Taylor InsureTech
2
Charles Taylor InsureTech
3
Policy Admin System
MOCHA TMEL
DA Systems
ATLASBARS
Lineage
Bordereau Portal Tide Platform
Questionnaire
Rules
Cleansing
Credit Control
Transformation
XML
Gateway
API
Gateway
Tide XML Adapter
Underwriting Performance
Validation
Bordereau Management
Store
Data Warehouse
Difference EngineData Framework
BrokersInsurersBrokersTPAsCoverholders Service ProvidersAuditors
tide
Analytics
Reporting
Coverholder
Management
Contact
Management
Audit
Management
Google Maps
Geo-Coding
Built from deep knowledge of Delegated
Authority processes and data, Charles
Taylor InsureTech’s Tide platform entirely
rethinks the way binder data is handled.
Introducing Tide
We take information from offline systems into a centralised, cloud-
based platform which processes, standardises and validates
data against binder rules at source – without the need for insurer
intervention. We ease the reporting burden for Coverholders
through an intuitive bordereau submission portal or real-time
data feeds from their Coverholders’ systems. Insurers have an
unprecedented level of access to data, a rich set of reporting tools
and for the first time ever, can see risks written in real-time.
The feature diagram highlights the core features, processes and
interactions within Tide which together provide the complete end-
to-end platform that the market requires. All parties involved with
Delegated Authority have access and have an important role to
play. This ranges from the Coverholders and TPAs submitting data
(using either bordereaux or XLM from an administration system),
to Brokers reviewing data and providing narrative to Insurers,
to Insurers analysing, reporting and experting data.
50k
Bordereau Processed
Knowledge that went into
the development of Tide
45years
Experience of DA went into
the development of Tide
Coverholder Benefits
–– Single submission of data to the London Market
–– Reduced time spent preparing and submitting
bordereaux to the London Market
–– Reduced time dealing with bordereau
queries, as these are dealt with at the time
of submission
–– Reduced chance of writing risk outside of
binding agreement, as XML can be validated in
real-time at the point of binding
–– Access to extensive management dashboards,
especially useful for validating that bordereaux
have been interpreted correctly by the Broker
and Insurer
Insurer Benefits
–– Coverholders do not need to change or adapt
–– Real-time data visibility
–– Efficient and effective management of the entire
DA portfolio
–– Ability to quickly identify and intervene when
things go wrong
–– Reduced claim disputes
–– Reduced cost of processing bordereau data
through the elimination of time-consuming query
loops and the need for technical resources
–– Meet the data reporting requirements of today
and future proof for those of tomorrow
3. Anyone Uploads
Bordereau
Processing
Complete
Questionnaire
Complete
Process
Bordereau
Begin
Questionnaire
Begin
Mapping
Process
Bordereau
Map Reference
Data
Deploy to
Production
Assign to Section
Risk Codes
Store in
Data Warehouse
Review Correct
Data Type Issues
Review Correct
Contract Issues
Review
Summary
Review Correct
Lifecycle Issues
Approve
Processing
Create
Formulas
Mapping
Complete
Determine
MI Req.
Determine
Regulatory Req.
PRE TRANSFORMATION STEPS ARE AUTOMATED
TRANSFORMATIONS ARE AUTOMATICALLY APPLIED BASED
UPON RESPONSES TO QUESTIONS, NO CODING REQUIRED
QUERY LOOP
MANUALLY TRIGGERED
Insurer Uploads
Bordereau
Review
Bordereau
Determine
COB Req.
Create
Template
Map Basic
Fields
Add
Validation
Hard-Core
Missing Data
Respond with
Resolution
Action
Query
Review
Query
Processing
Complete
Extract
Data
Resolve Data
Type Errors
Request Updated
Bordereau
Review
Lifecycle Errors
Resolve
Lifecycle Errors
Review
Summary
Review Data
Type Errors
Submit for
Review
Request Data
Type Corrections
Review
Contract Errors
Resolve
Contract Errors
Request Updated
Bordereau
Manually
Reconcile
System Processing
Complete
Add Section
Risk Codes Details
Tide Process
Tranditional Process
Business Input Required
Charles Taylor InsureTech
4
Charles Taylor InsureTech
5
Love them or hate them, the market is awash with bordereaux. Attempts to standardise, collaborate, simplify and consolidate have gained
little traction, leading to a situation where perhaps there are more formats than ever before. Commercial relationships are sensitive and new
bordereau format requests have been met with a hostile reception.
Tide has been designed with this problem in mind, processing data from whatever it is presented with. There is no requirement for
Coverholders and TPAs to change. Tide will do the hard work by searching for data and reporting back what it finds, along with
assessments against the relevant market and regulatory requirements. No mapping is required.
The bordereau express
Tide fundamentally changes how bordereau data is handled. Gone is the need for skilled technical resources. With only a knowledge
of the bordereau required. Only a knowledge of bordereaux is required. Individuals uploading a bordereau are guided through a series
of contextual questions to enable it to be translated, standardised and stored. Data quality issues are resolved en route to the finish,
where bordereau data, associated to contracts at a risk code level, is ready to be consumed via reports and dashboards. The result
is an intuitive and simple experience that anyone, including Coverholders and TPAs, can directly connect with, confident that once
complete, all the data requirements of the London Market will have been met.
Data transformation the right way
Creating mappings from one structure to another is easy.
It’s effectively copying and pasting performed over and
over again. But mapping is a minor aspect of the bigger
transformation process…
For example, a column called ‘Gross Premium’ may at first appear
to be a simple item to map, but is it? Does it include Tax? Is it for
the whole policy or just for a particular part of it (e.g. property vs
liability)? Does it contain currency symbols? How is it reported
for endorsements? Is it the sum of all sub-locations or for just
one location? Will the same value be reported for the declaration
in each bordereau or only when the value changes?
This is a lot to think about that is only in relation to one column. To
make matters worse, in any other system, the user would then need
to implement the resulting rules into a traditional mapping tool using
formulae and expressions. Even if you outsourced this aspect,
you would be required to validate the end result.
Tide approaches the transformation process in the right way.
It understands Delegated Authority data and the way that it is
reported, having been built from the ground up with this data as
its focus. By using Tide, users are not required to answer complex
questions. Tide does this for them, after it has applied automatic
cleansing, based upon the type of data.
• Reference data is cleansed using 50+ standard
pre‑applied lookups
• Monetary values are automatically stripped of any characters
• Dates are standardised without any user input with
100% success
Where there is ambiguity regarding a bordereau column, Tide
will apply knowledge from previously processed bordereaux to
pre‑answer plain English questions about it. The user can then
verify or correct Tide’s answer. Not a formula or expression in sight.
Asking the right questions…
When a bordereau is presented to Tide
for the first time, it’s structure is analysed
to ascertain if a bordereau with a similar
structure has been seen before. If not,
a questionnaire process is automatically
triggered, asking a series of questions
about the bordereaux, solely relevant
to the class of business or product it is
associated with. The questions remove the
need for the individual to have knowledge of
traditional bordereau mapping/processing,
regulatory requirements, Class of Business
requirements, MI requirements, formulae,
validation rules or Excel itself; all of which
are automatically applied by Tide.
To further simplify the process, the questions
are presented in plain English, with further
guidance information available at every stage.
Answers to questions leads dynamically to
relevant follow up questions, ensuring that
only relevant questions are asked.
Knowledge from previous bordereau is
used to pre-answer questions meaning
these only have to be reviewed rather than
answered again. As a result, and with
an ever expanding dataset, the majority
of bordereau fields are automatically
translated to Tide fields.
It’s as easy as 1, 2, 3…
Tide is a central portal for bordereau submission and
transformation. The submission and transformation processes
are separate, but interconnected, and either can be performed
by coverholders, brokers and/or insurers.
The performer of each process is defined against the Contracts
in Tide. Coverholders can submit bordereaux to their broker(s)
and insurer(s) simply by uploading them; and in the ideal scenario
the Coverholder would also transform each bordereau using the
powerful Tide questionnaire. The latter process however could
be assigned to the broker or Lead insurer if necessary; it’s up
to you to decide.
Whatever the case, when a bordereau has been submitted it can
be accessed by any party associated with the Contract against
which it was uploaded.
Coverholders can access the secure portal to submit bordereaux
to the London Market in accordance with the terms of the Contract.
Submitting a bordereau involves just three clicks of the mouse:
1. Select the Entry against which the bordereau will be submitted
2. Supply the bordereau file when prompted
3. Click ‘Submit’
That’s it – there is nothing else to do.
The questionnaire includes:
üü Relevant regulatory data requirements
üü Relevant MI/BI data requirements
üü Relevant data standards
The questionnaire does not require:
ûû Bespoke data lookups
ûû Formulae to be created by the user
ûû Any technical knowledge
The questionnaire automatically applies
relevant validation and verification checks,
using previously processed data
The ‘B’ word
4. Insurers
Brokers
Coverholders
TPAs
Brokers
Bordereau
Bordereau
Bordereau
tide
Workflow
Charles Taylor InsureTech
6
Charles Taylor InsureTech
7
Coverholder/
TPA system
MOCHALineage
Reference
Data Validation
Structure
Validation
Tide XML
Adapter
Rating
Validation
To Standard
Transformation
Process
Tide XML Gateway
XML has long been presented as the solution to a world of
bordereau confusion, but salvation has yet to materialise.
Tide has been developed with XML input just as important as
bordereaux. The same customisable schemas that are used for
bordereaux are available for XML. Changes to schemas trigger
immediate changes to the XML gateway and the validation
rules applied to bordereaux are the same as those applied
to XML messages.
Tide can process XML messages on a real-time, scheduled or
monthly basis and the contents of each message can represent
one or many transactions. Reporting methods and frequency for
each and every contract can be set and different methods can be
combined (e.g. risk data by XML and premium data by bordereau
for a single Contract).
As would be expected, Tide acts as the central repository for all
delegated authority data submitted by Coverholders, TPAs and
Brokers. Uploaded bordereaux, are associated with Contracts and
Periods in accordance with the conditions on the slip. Every file
uploaded is virus scanned so that Brokers and Insurers accessing
the files can feel safe in the knowledge that there will be no
nasty surprises.
But Tide isn’t just any repository. As well as storing the files, Tide will
provide operational reports detailing those that have and have not
been uploaded. Such reports allow the performance of bordereau
submissions to be monitored. Exceptions can be highlighted and
tracked in more detail. Tide also provides basic data quality checks
on bordereau files that are simply stored and not transformed.
A future feature will be the ability to create notifications which are
triggered when a specified bordereau is received. The notifications
will be sent by email to any specified recipient with a direct hyperlink
to the new file.
The powerful workflow functionality used for the transformation,
validation and approval processes also extends to the submission
process and the repository. It ensures that actions and status are
clearly visible to all interested parties, and that they know the correct
course of action to take next.
XML adapter and rating validation
In addition to the comprehensive XML gateway,
Tide also employs an XML adapter that can be
installed alongside the Coverholder systems,
extracting all data captured during the quote/
bind process via XML directly to the Tide
platform. This allows Coverholders to continue
to write and account as they usually would
with the added benefit that they would now
meet any current and future UK regulatory
and reporting requirements.
The adapter has been developed to simplify
and expedite the ability to deliver XML from
Coverholders and TPAs to the London market.
One-off configuration is required to adapt it to
the Coverholder/TPAs system but, once that is
complete, data can be picked up in real-time
by the adapter and delivered to Tide.
As well as providing real-time data, the adapter
also enables real-time rating validation. The
details of the Policy being written are verified
using contractual terms stored in Tide against
the associated Contract(s) in real-time. If there
is a problem, the risk cannot be bound in the
Coverholder’s system until it has been resolved.
To date the adapter has been integrated with
Coverholder systems that account for 80% of the
US Property Surplus Lines market. It is currently
in live use by over 30 Coverholders generating
over 15,000 transactions a month.
Nothing left behind
When time is of the essence, it is sometimes necessary to “just get
the job done”. This may mean only the core data (perhaps required
for a regulatory return) is transformed from a bordereau to save time
up-front. This would impact the ability to produce MI or BI, unless
it is Tide into which the data is being transformed…
Bordereau columns that are not translated during processing
are still imported and retained in Tide. At any stage in the future,
these columns can be re-visited and meaning can be applied.
Tide will then transform the data automatically without the need
to re‑process any bordereaux or to delete any existing data. The
newly understood column(s) are simply imported in the same way
that previous columns were and they will be presented for reporting
and analysis. It is as simple as that.
15,000Transactions processed
a month in 2016
80%
Of US Coverholders could
use the Adapter today
without customisation
Experience the future All in one place
5. ER3001
Regulatory
Requirements
tide
TOM Data Initiatives
Project Tomorrow
Lloyd’s
Coverholder Standards
Class of
Business Standards
Charles Taylor InsureTech
8
Charles Taylor InsureTech
9
Dataset
Rules
Tide Data Directory
(3 x Datasets)
Reference
Data
Schema
Rules
Tide
Schema A
Tide
Schema B
Tide
Schema C
Contract
Rules (inc Limits)
MI/BI Data
Standards Data
Regulatory Data
Tide uses the market standards as its base dataset. Using rules
and pre-defined reference data, Tide implements the requirements
of the standards, with appropriate conditions, directly into the
Schemas. This results in the situation that, when bordereaux are
presented to Tide, the system will respond with only the relevant
data requirements from these standards, not every data item
across all standards.
But what does this mean in practice… If a bordereau is being
processed against a European Property Contract which has a High
Conduct MI rating, Tide will raise questions regarding property data,
Conduct MI and FloodRe. When Tide is processing the bordereau,
it will check that all required fields have been provided. In the case
of Conduct MI, all of the risks in the bordereau must have all of
the Conduct MI data; if any do not Tide will raise an exception.
In contrast, the FloodRe data requirements are only relevant for the
UK based risks. But this is not a problem for Tide; it knows about
this condition, so when it is checking the FloodRe data it will only
report issues in respect of missing or invalid FloodRe data when
the Risk is located in the UK.
Other examples include: SII Pillar 3, ELTO, Surplus Lines Broker
details and US, UK and Australian Terrorism requirements. None
of these are hard-coded, all are customisable via the Tide UI which
means that any rule change to reflect a regulatory change can be
implemented quickly.
Data framework
The data framework in Tide governs the translation of all incoming
Delegated Authority data into standardised data, ready for use and
reporting. All components of the framework are user-configurable,
enabling immediate deployment without the wait for a future
version release. The dynamic rules module enables the immediate
deployment of dataset, compliance data, schema, and contract
rules. Rules directly connect with dataset fields and schemas so
they can be coherently applied to incoming Delegated Authority
data. Rules can be used by a central body to set mandatory
compliance data requirements that are validated during processing.
Data Directory
Tide’s data directory comprises
three datasets: risks, premiums
and claims. Each dataset
contains all available fields
against which incoming
delegated authority data is
matched. For each data field
information such as definition,
data type, field grouping and
other attributes are recorded to
direct field use across various
system functions. Fields are
created for either regulatory
reasons, based on market
standards, or for business
reasons.
Data Schema
Distinct logical collections of
fields from datasets are used to
create schemas. This flexibility
means that any number of
different schemas can be
created, typically for particular
classes of business or product.
All incoming delegated authority
data is translated to a single
schema, irrespective of whether
it was received in Excel or XML
format, meaning that individuals
are only presented with the
fields that are relevant to the
data they are submitting.
Reference Data
Where a field in a dataset
is defined as a Reference
Data Type it is then linked to
a Reference Dataset. Any
incoming delegated authority
data is then cleansed to the
single set of values contained
in the Reference Dataset.
This cleansing process is
informed by the previous
cleansing of all values across
market participants. The
result is a highly efficient
process of cleansing and a
fully standardised dataset ready
for reporting and analysis.
Rules
User configurable rules can be
added to datasets, schemas
and individual contracts,
allowing incoming data to be
validated against any criteria.
Breaches are clearly reported
and if necessary can be
configure to prevent import.
Applying the standards
6. Charles Taylor InsureTech
10
Charles Taylor InsureTech
11
Tide Data Warehouse
Tide provides a dedicated and scalable Microsoft Azure Data
Warehouse instance for each and every client. The data it
contains is yours and only yours. Additionally, data analysis and
reporting tools are provided to you.
Coverholder
Management
Contract
Management
Risk, Premium
Claim Data
Audit
Management
Rules
Output
Microsoft Excel Analysis
Simply connect Tide as a data
source in Excel. The entire Excel
toolkit is at your disposal to
analyse data.
Your preferred tool
Tide will allow you to connect
any data analysis tool to the data
warehouse via SQL, so you
can use the tool you are most
familiar with.
Tide Dashboards
Tide possesses powerful
interactive dashboards built
upon Microsoft Power BI
infrastructure.
Centred around themes,
such as Underwriting,
Claims Operations.
Tide Reports Extracts
Tide has a wide range of pre
-built reports and extracts
straight ‘out of the box’.
These include operational
and MI reports,
schema-centred and
regulatory extracts.
Risk/Written
Data
Contractural
Terms
Credit
Control
Premium/Paid
Data
Payment Status Gross Premium
Unpaid 1,000.00
Unpaid 33,262.55
Partially Unpaid 50,131.05
Unpaid 52,980.86
Unpaid 55,609.97
Unpaid 68,755.52
Unpaid 75,328.29
Tide does not restrict in any way how the data it stores is used,
other than controlling access to valid stakeholders. In fact, Tide
provides advanced data warehouse technology and the data is
updated in real-time as bordereaux and XML transactions are
And coming soon...
Balancing the books
Tide transparently presents the credit control position between
risk and premium datasets to all parties, at the most granular level.
Once risk and premium Delegated Authority data is processed,
Tide immediately presents the credit control position. Based on
the contractual terms recorded in Tide, screens and reports show
risks which are overdue for payment at portfolio and contract
Contract management
Knowing the specifics of the Binders and Lineslips against which
Delegated Authority data is received is crucial. Tide can collect
every detail and use it to verify, validate and enrich the incoming
data. Limits, restrictions, shares, EPI and other details can be
recorded and monitored. The result is underwriting and claim
performance reporting like never before.
Regulatory returns
As part of the standard offering, Tide includes the following
reports to meet regulatory requirements:
• Solvency II Pillar 3—ASR249
• Solvency II Pillar 3—ASR250
• Solvency II Pillar 3—ASR252
• Conduct MI reporting suite
Cat modelling
Tide has pre-built exposure management risk data outputs for
both RiskLink (RMS) and Touchstone (AIR). These are produced
using contractual details such as Perils and Interests, entered into
Tide, in conjunction, with the bordereau data that is received. The
outputs can be produced using any criteria that you wish, such
as one or many specific Contract(s), a particular class of business
or perhaps on a portfolio basis.
Tide will pro-actively monitor the artefacts that coverholders
must possess to be compliant such as EO, Reports and
Accounts, Licences, Individual’s credentials.
Any issues are raised and reported to the associated
Insurers. To assist the process, Coverholders can submit
updated artefacts when they process their next bordereau.
Coverholder
Management
Audit
Management
An audit module within Tide will provide a framework for
the scoping, execution (including costs) and management
of findings that result. Every finding recorded in Tide can
have one or many actions assigned to it, allowing the Audit
to be managed until completion.
During the scoping phase, details that Tide stores about
the in-scope Coverholder, and the Contracts it writes, can
be exported and shared with the Auditor, greatly increasing
the focus of each Audit.
approved. The data structure is based upon the powerful Data
Framework within Tide, however it is greatly simplified in terms of its
hierarchy, to facilitate quick and efficient analysis. And the best part
is that it’s included within the standard Tide service.
level by amount and age, thus giving the earliest possible view
to all contract participants. Insurers can gain confidence that all
premiums due for written risks have been settled. Coverholders
view the same credit control position, alongside automated alerts
and emails highlighting non-payments, as part of their single
gateway to the London market.
Every detail captured by Tide regarding the contract can be used
within rules, allowing you to set any conditions, alerts or warnings
that you require which are applied to all incoming data.
It’s your data, come and get it
7. Submission
Schedule
Bordereau/XML
Submission
Transformation
Cleansing
Validation
Issue
Management
Review
Approve
Report
Charles Taylor InsureTech
12
Charles Taylor InsureTech
13
Data Enrichment
Approval
LPAN Production
Bodereau
Submission/Monitoring
Analytics
Reporting
Broker
Knowing the status of your delegated book is becoming ever more
important. In order to achieve this, the individual components need
to be closely tracked and managed.
At all stages in the process, Tide provides real-time information on
the status of individual Bordereaux and XML messages. The data
is visible at a granular level or at any other aggregated level. Actions
...and that includes brokers
Be it Coverholders, TPAs, Brokers or Insurers, Tide provides
a bespoke service to each.
For Brokers, Tide provides a platform that can be used to manage
the collection of data from Coverholders and its delivery to Insurers.
Brokers can Transform the data themselves or assign this task to
the Coverholder. Whichever is the case, the Broker can choose
to review data prior to its submission to Insurers. Notation can be
added, analysis included and messages communicated along
with the cleansed bordereau data.
Tide is an inclusive solution with a role for everyone. It achieves this through a customisable framework, putting control in to the hands of
the practitioners. Every contract requires the following roles to be defined:
Although Tide provides the ability for roles to be performed by different companies, it is not necessary for this to be the case in all
instances. If desired, all roles can be performed by the same company, as would be the case if an insurer chose to Upload, Process and
Approve bordereau data itself, with no involvement from the Coverholder.
All around the world
Tide provides functionality for its clients to upload Rates of
Exchange, using an intuitive user interface, on a monthly, quarterly
or an annual basis. Tide supports multiple exchange rate “types”
per client. For example, annual reinsurance rates can be uploaded
and managed in parallel to monthly PIM rates.
In addition, Tide stores central rates on behalf of the market.
can be created against items and assigned to individuals or teams
for input and completion. As would be expected, notifications are
raised for any exceptions encountered and the whole process
is tracked with a full audit trail.
In addition to the management of bordereau delivery, Tide also
provides market reporting functionality so that LPANs and Tax
submissions can be produced without any manual intervention.
Bordereau data is assigned to Sections, Risk Codes, Countries
and, where applicable, to States. This is ideally done using data
in the bordereau. If such data is not available, logic defined by the
Brokers/Insurers themselves can be used. This enhanced data
can then be aggregated for reporting in the re-defined formats.
During the transformation process, exchange rate periods are
applied to submitted delegated underwriting authority data, from
all Risk, Premium and Claim Bordereaux and all XML transactions.
This provides the ability for reports, analyses and analytics to
be produced in any currency required, using any exchange rate
type that has been uploaded into Tide, without additional offline
processing or calculation.
The company responsible for
administering the contract, which
includes capturing the relevant
contract and entering it into Tide.
Contract
Administrator
Bordereau
Administrator
Data
Owner
The company responsible for
processing each bordereau into Tide.
In the case of system-to-system
XML submissions, there are no
specific manual activities to perform
in this role.
The company responsible for
approving the processed data prior
to it being usable by other companies
associated with the contract.
Working for you A role for everyone
8. tide
Audit Trail
Data
Submission
Data
Transformation
Data
Enrichment
Error
Handling
Actions, Notes
Notifications
Contract
Management
Charles Taylor InsureTech
14
Charles Taylor InsureTech
15
A trail of evidence
When traditional bordereau processing encounters data issues
the process stops, the user resolves the issues off-line and
then re-starts the process. Tide breaks free from this disjointed
approach, providing screens on which issues are resolved as
they are identified. Examples of this include correcting data types,
identifying and resolving new Reference Data values which need
to be translated and handling out-of-sequence transactions
(e.g. An endorsement before the original transaction).
Tide keeps detailed records of each and every alteration made,
regardless of how it was changed (user override, automated
enrichment, reference lookup etc.). This audit trail is stored with the
cleansed data and is made available for reporting, including pre-
built data transformation and data quality reports.
The audit trail extends to data stored by Tide that was not received
on Bordereaux or in XML messages, such as Contract details. This
data is also controlled by Roles ensuring, in all events, that only
authorised individuals are allowed to make changes.
Tide puts you in control of the service you receive. No support calls to raise, no waiting. Make the changes you want, when you want.
The administration possibilities include:
Add and edit divisions
associated with
your company
Review, and modify,
reference data values
and mapping
Review and upload
rates of exchange
Add, edit or remove
users and assign
permissions
Review, add,
edit business and
validation rules
Tide is hosted on Microsoft Windows Azure infrastructure, located
in the UK South datacentre. The UK West datacentre is used
for disaster recovery. Azure hosts many of the world’s largest
online services with the data centres having been designed and
Segregated data
Your data is yours, and only you can see it. Fact.
Tide is a shared platform which does mean that it contains data
owned by more than one client, but controls and techniques have
been embedded in Tide from the start to ensure that each user
only sees what s/he should. The details entered against contracts
Any browser, any device
Tide can be accessed from the latest browsers, regardless of the
device you are using, with each receiving a tailored experience
based upon the screen size. Whether you want a widescreen
constructed with stringent levels of physical security and access
control, power redundancy and efficiency, environmental control,
and recovery capabilities. The physical facilities have achieved
ISO 27001 and SOC/SSAE 16/SAS 70 Type II certification.
in Tide specify who can access data (Coverholder, TPA, Broker
and Insurers), and the roles also define who can put the data there
(Contract Administrator, Bordereau Administrator and Data Owner).
Every client of Tide has its own dedicated data warehouse;
therefore all reports produced from it can only contain your data.
dashboard experience on your desktop, or a bordereau summary
on your phone, it’s your choice.
Secure from the ground up
The data centres have protection for items such as DDoS attacks. Using Azure tools, Tide is monitored for unexpected and
potentially invalid spikes in traffic and must react accordingly. Where a spike is genuine user activity, Tide scales-up to handle
the load. Where it represents some form of attack, this will be blockedusing Microsoft’s data centre protections.
Form-based user authentication denies access to attackers from the outset. All user passwords are stored in hashed format
and cannot be decrypted.
Microsoft Multi-Factor authentication requires users to validate who they are with 2 forms of authentication, with an
Authentication key provided by email or SMS. Whenever a user logs on from a PC, Mac or Tablet not seen before by Tide,
the user is re-sent an authentication key and requested to revalidate.
All communication to and from Tide is encrypted using HTTPS. In addition, User Session details are cryptographically secured.
Comprehensive input validation ensures robustness against all forms of malicious data input from the user, infrastructure,
external entities or database systems.
You’re in control It’s all taken care of for you