Syed Babar H Rizvi has over 12 years of experience as a software developer and team lead working on projects in healthcare, telecom and manufacturing. He currently works at Parametric Technology Corporation as a module lead. He has extensive experience with ETL tools like Informatica and databases like Oracle. He also has knowledge of big data technologies such as Hadoop, Spark and machine learning techniques.
Meetup: Big Data NLP with HPCC Systems® - A Development Ride from Spray to TH...HPCC Systems
HPCC (High Performance Computing Cluster) Systems from LexisNexis is an open source massive parallel-processing computing platform that solves Big Data problems. In this talk, attendees will be given an overview of HPCC Systems and see a demonstration of its use to parse data from free-form and semi-structured text. This represents a combined text extraction task with human intervention. The code elements and massively parallel processing principles involved in accomplishing these tasks will be thoroughly discussed.
• 11+ Years of IT Industry experience in Analysis, Design, Development, Maintenance and Support of various software applications mainly in Data Warehousing (Informatica Power Center, OWB, SSIS and Business Objects), Oracle (SQL, PL/SQL) and Teradata in industry verticals like Finance, Telecom, Retail and Healthcare.
• Work experience in client facing roles in UK and Ireland.
• Performed numerous roles in Business Intelligence projects as Data warehouse System Analyst, ETL Designer, Onshore coordinator, Technical Lead and Senior Data warehouse Developer roles with multinational IT result-driven organizations
• Extensive experience on Data integration projects accessing sources like Teradata, Oracle and SQL server.
• Created robust EDW Solution from various types of sources like Flat files, XML Files, EDCDIC Cobol copybook from Mainframe systems, DB2 Unload files.
• Extensive experience on Data discovery, cleansing using Informatica IDQ.
• Resolved Inconsistent and Duplicate Data issues during Data Analysis to Support Strategic EDW Goals.
• Extensive experience of Data Integration using Informatica Power center Tool stack.
• Strong knowledge on Data Warehousing concepts, ETL concepts, Data Modeling, Dimensional Modeling.
• Conducted training on Informatica and have achieved awards for proficient training capabilities.
• Excellent understanding of OLTP and OLAP concepts and expert in writing SQL, Stored procedure on Teradata, Oracle and SQL Server.
• Extensive experience in implementing Data Warehousing methodologies including STAR SCHEMA and SNOW-FLAKE SCHEMAS & 3NF for huge data warehouses.
• Extensive knowledge on Change Data Capture (CDC) and SCD Type 1, Type 2, Type 3 Implementations.
• Excellent understanding of Kimball and Inmon Methodologies.
• Provided leadership when addressing high level technical issues and questions with the functionality of the reporting and business intelligence applications.
• Managed the current and strategize to foresee and plan for the future engineering needs in Data Integration space.
• Performed roles as a interface and coordinator between Database Administration, ETL Development, Testing teams and reporting teams to eliminate the road blocks for smooth flow of information.
• Hands on experience in tuning ETL mappings, identifying and resolving performance bottlenecks in various levels like sources, targets, mappings, and sessions.
• Expert in designing and developing of complicate ETL mappings using Informatica PowerCenter.
• Proficient in optimizing performance issues using Informatica PowerCenter and Teradata.
• Having experience on using Teradata utilities (TPT, BTEQ, Fast Load, MultiLoad, FastExport, Tpump).
• Exposure in writing Shell Scripting as per given requirement.
• Work extensively in Teradata GCFR tool.
• Experience in SAP ECC integration with Informatica.
• Got training in Tableau, Qlik View& SAP BW 3.5and done POC for the same.
Meetup: Big Data NLP with HPCC Systems® - A Development Ride from Spray to TH...HPCC Systems
HPCC (High Performance Computing Cluster) Systems from LexisNexis is an open source massive parallel-processing computing platform that solves Big Data problems. In this talk, attendees will be given an overview of HPCC Systems and see a demonstration of its use to parse data from free-form and semi-structured text. This represents a combined text extraction task with human intervention. The code elements and massively parallel processing principles involved in accomplishing these tasks will be thoroughly discussed.
• 11+ Years of IT Industry experience in Analysis, Design, Development, Maintenance and Support of various software applications mainly in Data Warehousing (Informatica Power Center, OWB, SSIS and Business Objects), Oracle (SQL, PL/SQL) and Teradata in industry verticals like Finance, Telecom, Retail and Healthcare.
• Work experience in client facing roles in UK and Ireland.
• Performed numerous roles in Business Intelligence projects as Data warehouse System Analyst, ETL Designer, Onshore coordinator, Technical Lead and Senior Data warehouse Developer roles with multinational IT result-driven organizations
• Extensive experience on Data integration projects accessing sources like Teradata, Oracle and SQL server.
• Created robust EDW Solution from various types of sources like Flat files, XML Files, EDCDIC Cobol copybook from Mainframe systems, DB2 Unload files.
• Extensive experience on Data discovery, cleansing using Informatica IDQ.
• Resolved Inconsistent and Duplicate Data issues during Data Analysis to Support Strategic EDW Goals.
• Extensive experience of Data Integration using Informatica Power center Tool stack.
• Strong knowledge on Data Warehousing concepts, ETL concepts, Data Modeling, Dimensional Modeling.
• Conducted training on Informatica and have achieved awards for proficient training capabilities.
• Excellent understanding of OLTP and OLAP concepts and expert in writing SQL, Stored procedure on Teradata, Oracle and SQL Server.
• Extensive experience in implementing Data Warehousing methodologies including STAR SCHEMA and SNOW-FLAKE SCHEMAS & 3NF for huge data warehouses.
• Extensive knowledge on Change Data Capture (CDC) and SCD Type 1, Type 2, Type 3 Implementations.
• Excellent understanding of Kimball and Inmon Methodologies.
• Provided leadership when addressing high level technical issues and questions with the functionality of the reporting and business intelligence applications.
• Managed the current and strategize to foresee and plan for the future engineering needs in Data Integration space.
• Performed roles as a interface and coordinator between Database Administration, ETL Development, Testing teams and reporting teams to eliminate the road blocks for smooth flow of information.
• Hands on experience in tuning ETL mappings, identifying and resolving performance bottlenecks in various levels like sources, targets, mappings, and sessions.
• Expert in designing and developing of complicate ETL mappings using Informatica PowerCenter.
• Proficient in optimizing performance issues using Informatica PowerCenter and Teradata.
• Having experience on using Teradata utilities (TPT, BTEQ, Fast Load, MultiLoad, FastExport, Tpump).
• Exposure in writing Shell Scripting as per given requirement.
• Work extensively in Teradata GCFR tool.
• Experience in SAP ECC integration with Informatica.
• Got training in Tableau, Qlik View& SAP BW 3.5and done POC for the same.
A Software Engineer with experience in client application development in the IT industry, Involved in all phases of Project life Cycle (SDLC) -Analysis, Design, Development, Testing, Documentation and Maintenance of Web Applications with the extensive knowledge of Application development and related process.
Several years’ on teams practicing Agile or Scrum.
Extensive work with ASP.NET, ASP.NET MVC,ADO.NET, Entity Framework, LINQ, SQL Server, JQuery, AngularJS, Telerik Controls, Crystal Reports, User Interface Design, client side development, middle tier, and back end development.
• Associate Consultant pursuing Executive MBA with 3+ years of experience in Healthcare ,Banking domain & software development, implementation in the areas of Data warehousing using IBM web sphere Data stage 8.1 tool and IBM Info Sphere Data stage 8.7,ETL Architecture, enhancement, maintenance, Production support, Data Modeling, Data profiling, Reporting including Business requirement, system requirement gathering.
Explore our comprehensive data analysis project presentation on predicting product ad campaign performance. Learn how data-driven insights can optimize your marketing strategies and enhance campaign effectiveness. Perfect for professionals and students looking to understand the power of data analysis in advertising. for more details visit: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
StarCompliance is a leading firm specializing in the recovery of stolen cryptocurrency. Our comprehensive services are designed to assist individuals and organizations in navigating the complex process of fraud reporting, investigation, and fund recovery. We combine cutting-edge technology with expert legal support to provide a robust solution for victims of crypto theft.
Our Services Include:
Reporting to Tracking Authorities:
We immediately notify all relevant centralized exchanges (CEX), decentralized exchanges (DEX), and wallet providers about the stolen cryptocurrency. This ensures that the stolen assets are flagged as scam transactions, making it impossible for the thief to use them.
Assistance with Filing Police Reports:
We guide you through the process of filing a valid police report. Our support team provides detailed instructions on which police department to contact and helps you complete the necessary paperwork within the critical 72-hour window.
Launching the Refund Process:
Our team of experienced lawyers can initiate lawsuits on your behalf and represent you in various jurisdictions around the world. They work diligently to recover your stolen funds and ensure that justice is served.
At StarCompliance, we understand the urgency and stress involved in dealing with cryptocurrency theft. Our dedicated team works quickly and efficiently to provide you with the support and expertise needed to recover your assets. Trust us to be your partner in navigating the complexities of the crypto world and safeguarding your investments.
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
1. SYED BABAR H RIZVI
Page 1 of 4
Name : Syed Babar Hussain Rizvi
EmailID : babar.rizvi.pu@gmail.com
ContactNo: 919739283080
BigData & Machine Learning Practitioner
Executive Summary
A professional with 12+ years of experience in Software Development / Onsite Co-ordination / Team Leading
with deep insight in to Healthcare / Telecom/Manufacturing industry.
Working with Parametric Technology Corporation as Module lead.
An expert in leading teams to successful project implem entation with leadership skills, quality assurance and
schedule adherence.
Worked on all phases ofdata warehouse developmentlifecycle,from gathering requirements,ETL design and
implementation,and supportor new and existing applications.
Excellent technical and analytical skills with clear understanding ofETL design and projectarchitecture based on
reporting requirements.
Demonstrated expertise-utilizing ETL tools,including prior version of Informatica 7.X, 8.X to currentversion
(9.5.1) and RDBM systems like Oracle.
Worked in Agile and SDLC models.Worked on Jira board to track and reportprogress to management/Clients.
Estimated user storypoints,helped on Sprintplanning and handled go/no-go calls.
Knowledge ofBig data technologies like Hadoop,Spark,Hive and Python.
Working Knowledge of various tools like Control-M, Bitbucket, SVN, Splunk etc.
Areas of Exposure:
ETL: Informatica 7.x, 8.x, 9.x
SQL: SQL Developer, Toad
RDBMS: Oracle 10g, Oracle 9i
Programming Languages: SQL, PL-SQL, Python
Version Control: VSS, VCTL, Github
Professional Qualification:
Post-Graduate Certificate Program on BigData Analytics & Optimization (360 contact hours +
project hackathon & Viva) from INSOFE, accredited by Language Technologies Institute of
Carnegie Mellon University, USA.
BE (IT):- Aggregate 62% Year-2006
College:- Government Engineering College, Jagdalpur
University: Pt Ravishankar Shukla University
Data Science Projects
Image Classification: The objective of this projectwas to correctly identify any image thatis provided as a test.
Approach: A train data setwas provided which consisted of10 folders having differentimages.The total numbers of
images in all these folders were around 3000
• Data was pre-processed and broughtin the correct format.
• Data Augmentation technique was used to further enhances the image classification accuracy.
• To further see the improvement,VGG16 architecture was used and it improved the accuracy since we had used
the transfer learning concept.
2. SYED BABAR H RIZVI
Page 2 of 4
Text Classification: The objective of this projectwas to correctly identify the urgency (1, 2, 3 or 4) of the support
tickets based on the text in the body of the email.
Approach: A train data setwas provided which consisted of47000 records with 9 columns
• Data was pre-processed and was cleaned ofall the NA values to bring it in correct format.
• Tokenizer API was used to convert all the words into numeric format
• Glove embedding weights for our text data so that they can be further used for the embedding layer
• 4 layer CNN model along with embedding layer was builtto train and testthe data
Fraud Detection: Predicting the suspicious transactions using retail salesmen reports analysis & Salesmen
Segmentation
Approach: A train Data setwas provided which contained approx.48000 sale transaction records.Records were
classified as Fraud-Yes,
• Fraud-No or Fraud-Indeterminate.
• Data was pre-processed and Ration Analysis was done to create new features along with Outlier detection
methods.
• Benford's Law was used to analyse the prospective fraudulenttransactions.
• SMOTE was done for the class imbalance problem.
• Various models were builtto predictthe correct classification ofthe records.
• Clustering was also used to classifythe Salesperson as High Risk,Medium Risk and Low Risk based on his Sales
record data.
Tool Used:
Deep Learning (MLP, CNN, RNN and LSTM),
Machine- Learning (Regression, PCA, SVM, Clustering,Naive Bayes, KNN, Ensemble and DT),Linear
Programming,NLP,Image and Video processing, Monte Carlo Simulation,Genetic algorithm, Using Hadoop and
Spark Ecosystem,Statistics & Probability, Spark and Python.
PROFESSIONAL EXPERIENCE
Organization: - Parametric Technology Corporation, Pune
Duration: - (from Sept 13-Current)
Recent Project Name : Amtrak
Sector: USA Govt.
Designation: ETL developer and Information Analyst
Client: Vodafone Italy
Tools/Technology: Informatica PowerCenter 9.5, Oracle11g, Amazon Redshift, S3, EC2.
Accountabilities:
Preparing the POC and design path for EDW migration..
Prepared E2E for Oracle to Redshiftmovementofthe warehouse.
Worked on high level ERDs.
Database migration,Production Support,RCA.
Functional requirementanalysis, QualityImprovement,Regression testing.
3. SYED BABAR H RIZVI
Page 3 of 4
Organization: - Vodafone India Services pvt Ltd, Pune
Duration: - (from Mar 11- Aug 2013)
Project Name : Vodafone Online
Sector Telecom
Duration: Mar’11- Till Date
Designation: Onsite Co-ordinator and Information Analyst
Client: Vodafone Italy
Tools/Technology: Informatica 7.1 and 8.5.1 as ETL Tool, Oracle9i and Oracle10g as Database and PL/SQL
and Shell scripts.
Details: At its core is the Vodafone-developed interchange system. The n-tier design and service
oriented architecture delivers enhanced flexibility, scalability, and reliability. To meet the
reporting and analysis needs of Post/Pre paid, the data warehouse, Management
Reporting System and Surveillance and Utilization Review System components are
integrated into a cohesive Billing solution built to use high-performance data marts. The
individual business process RADs discusses the detailed use of the interChange
application to supportthe specific DSS business processes. The complete requirements
analysis for the various DSS business processes is available within the individual
business processes. The Business Objects and Informatica Project within the Vodafone
Italy Online Practice of the business is the first step in VF-Italy goal to migrate from their
current model to a new model where Vodafone provides all their Business Intelligence
developmentand supportresources and services in an onshore /offshore model (20-80%
mix).
Accountabilities:
Responsible for Preparing Technical Specs,analyzing Functional Specs,Developmentand
maintenance ofcode.
Responsible for the complexity determination and estimation.
Developing complexETL mapping and its corresponding sessions & worklets,workflows.
End-to-end testing of Data warehouse/Data Mart load.
Distributing the work between offshore team member and tracking the developmentprogress.
Discussing the technical way of developmentwith onsite and offshore team lead.
Doing the review of code developed from offshore.
Organization: - Tela Sourcing Pvt ltd, Pune
Duration: - (from May 09- Dec 2011)
2) ProjectName : HealthAspex
Skill/Tools : Informatica 8.5.1, UNIX, Oracle 10g, SQLServer 2005
Role : ETL Developer
HealthAspex system is business integration solution provider focused exclusively on healthcare payers.
Systems help in processing ofthe claims filed by the differentproviders for remunerations from the client
(payers) with both online and offline functionality ofprocessing claims.Our solution helps TPAs and PPOs
save time and money by converting paper claims into easy-to-manage electronic data.
4. SYED BABAR H RIZVI
Page 4 of 4
ResponsibilitiesasaTeam Member:
Extraction, Transformation and loading ofdata from flat file, Oracle, SQL server 2005 sources to
oracle database.
Created Informatica mappings and mapplets using differenttransformations.
Used differenttransformations like source qualifier,filter,aggregator,expression,connected and
unconnected lookup,sequence generator,router and update strategy.
Analysis of certain existing mappings,which were producing errors and modifying them to produce
correct results.
Used workflow manager to create session task and other tasks.
Involved in unittesting.
Documentations ofmappings as per standards.
Involved in the code review and preparing the code review document.
Organization: - Mphasis An EDS Company, Pune
Duration: - (from Jan 07- Feb 09)
1) ProjectName : USGS (iCE 4.1)
Skill/Tools : Informatica 8.5.1, Unix, Oracle 9i
Role : ETL Developer
This application is developed for Medicaid system (US government’s health care project). It’s an N-tier based application
that manages automation for different systems of Medicaid such as; Claims, Financial, Provider, Recipient, Drug Rebate and
Third Party Liabilities, etc. The architecture of this portal consists of 2 major parts; the core application and the state-wise
application. The core application is the underlying platform that has all the standard features for Medicaid automation and state-
wise application is the extended and customized version of the core to suit the requirements for specific state.
The Social Security Act is a Federal/State entitlement program that pays for medical assistance for certain individuals and
families with low incomes and resources. This project aimed at development, testing and deployment of Healthcare Insurance
infrastructure for all the States in USA. MMIS is an automated claims processing and information retrieval system that helps
State government administer the Medicaid program. This contains different subsystems like claims, managed care, recipient,
provider, third party liability, MAR, portal etc.
Worked on and mastered Claims and Financial, two of the most complex subsystems in USGS and helped team members on
the technical and functional minutes of Claims and their resolutions.
Responsibilitiesas aTeam Member:
Extraction, Transformation and Loading of the data Using Informatica
Used relational sources and flat files to populate the data mart
Involved in the development of informatica mappings
Created informatica mappings to load the data mart and monitored them.
Used workflow manager to create session task and other tasks
Involved in unit testing