Working as an Associate Engineer with hands-on experience in Python, Machine Learning, Database Technologies, Flask and Big data technologies like Apache Spark, Airflow
- The document contains the resume of Deepankar Sehdev which details his 3.5 years of experience in developing big data and data warehouse applications using technologies like Hadoop, Hive, Pig, Sqoop, AWS Redshift. It lists his roles and responsibilities in 4 past projects involving migration of data from mainframe and databases to Hadoop clusters and data warehouses.
Pardeep Kumar is a Software Development Engineer at Amazon with over 4 years of experience. He previously worked at Fair Isaac India Software Pvt. Ltd. (FICO) as a Tool Development Engineer. Some of his responsibilities included participating in the full development life cycle, active bug fixing, and developing features for projects like the GMRA portal and Blaze Advisor. He holds a Bachelor's degree in Computer Science and Engineering from the National Institute of Technology, Warangal. His areas of technical experience include projects involving distributed systems, Hadoop/Spark, and web development using technologies like Java, Python, C++ and SQL.
Arindam Sengupta has over 17 years of experience in architecting, developing, implementing, and customizing client-server and web-based applications. He has extensive experience with technologies like Hadoop, HDFS, MapReduce, Pig, Hive, HBase, Spark, Java, Oracle, SQL Server, and IBM DB2. Some of his recent projects involve designing Hadoop-based solutions for data ingestion, analytics, and visualization using technologies like Flume, Sqoop, HBase, MapReduce, Spark, and REST services.
Nagapandu Potti seeks a software engineering role that utilizes his technical skills. He has strong skills in Java, C, C++, Ruby, Scala, C#, databases like MySQL and MongoDB, web development technologies like JavaScript, AngularJS, and Ruby on Rails. He has work experience developing applications using these skills at Citrix and Cerner. Potti has a Master's degree in Computer Science from the University of Florida and a Bachelor's degree in Computer Science from Manipal University.
This document is a resume for Haridarshan H.S. that provides information about his education, technical skills, work experience, projects, and certifications. It summarizes that Haridarshan is pursuing a Master's in Computer Science from UT Dallas and has a Bachelor's in Electronics and Communication from RNSIT, Bangalore. He has skills in languages like Java, C++, Python and frameworks like AngularJS, Node.js. He has worked as a Systems Engineer at Infosys and led projects involving web development, data warehousing, and providing support. His academic and freelance projects involved areas like networking, web development, machine learning, Android apps, and Internet of Things.
This document contains the resume summary of Nageswara Rao Dasari. It outlines his 3.1 years of experience as a Software Engineer working on BIG DATA Technologies like Hadoop, HDFS, MapReduce, Hive and Pig. It also mentions his 1.4 years of experience in core Java and lists his technical skills like Java, SQL, JavaScript, CSS, Oracle, MySQL. It summarizes his most recent roles on projects for Barclays Bank and Target, where he performed tasks like data loading, writing MapReduce programs and Hive queries, and resolving JIRA tickets.
Scott Hovey has 17 years of experience developing automated tests, lab environments, and performance monitoring tools using languages like Python, Java, Perl, C++, and tools like Hyperic, JConsole, Linux/Windows native tools, and databases like Oracle, Postgres, MySQL. He has worked as a consultant for HeartMath Institute developing Cesium and Tableau maps, and as a Senior MTS Engineer and Tools and Performance Engineer at VMware maintaining virtual ESX code for testing and improving Tableau, Site Recovery Manager, and HP Quality Center performance and monitoring.
Mrigendra Kumar Bharti is a software developer with 2 years of experience developing web applications using Java. He has expertise in client/server development, data structures, algorithms, and Hadoop technologies like Spark, Kafka, Hive and HDFS. He has worked on projects involving network asset management, data ingestion using Apache NiFi, and integration with databases like Oracle, PostgreSQL and MongoDB. He has a Bachelor's degree in Information Technology and has received awards for his work. He maintains technical skills in Java, databases, frameworks and has experience with source control and build tools.
- The document contains the resume of Deepankar Sehdev which details his 3.5 years of experience in developing big data and data warehouse applications using technologies like Hadoop, Hive, Pig, Sqoop, AWS Redshift. It lists his roles and responsibilities in 4 past projects involving migration of data from mainframe and databases to Hadoop clusters and data warehouses.
Pardeep Kumar is a Software Development Engineer at Amazon with over 4 years of experience. He previously worked at Fair Isaac India Software Pvt. Ltd. (FICO) as a Tool Development Engineer. Some of his responsibilities included participating in the full development life cycle, active bug fixing, and developing features for projects like the GMRA portal and Blaze Advisor. He holds a Bachelor's degree in Computer Science and Engineering from the National Institute of Technology, Warangal. His areas of technical experience include projects involving distributed systems, Hadoop/Spark, and web development using technologies like Java, Python, C++ and SQL.
Arindam Sengupta has over 17 years of experience in architecting, developing, implementing, and customizing client-server and web-based applications. He has extensive experience with technologies like Hadoop, HDFS, MapReduce, Pig, Hive, HBase, Spark, Java, Oracle, SQL Server, and IBM DB2. Some of his recent projects involve designing Hadoop-based solutions for data ingestion, analytics, and visualization using technologies like Flume, Sqoop, HBase, MapReduce, Spark, and REST services.
Nagapandu Potti seeks a software engineering role that utilizes his technical skills. He has strong skills in Java, C, C++, Ruby, Scala, C#, databases like MySQL and MongoDB, web development technologies like JavaScript, AngularJS, and Ruby on Rails. He has work experience developing applications using these skills at Citrix and Cerner. Potti has a Master's degree in Computer Science from the University of Florida and a Bachelor's degree in Computer Science from Manipal University.
This document is a resume for Haridarshan H.S. that provides information about his education, technical skills, work experience, projects, and certifications. It summarizes that Haridarshan is pursuing a Master's in Computer Science from UT Dallas and has a Bachelor's in Electronics and Communication from RNSIT, Bangalore. He has skills in languages like Java, C++, Python and frameworks like AngularJS, Node.js. He has worked as a Systems Engineer at Infosys and led projects involving web development, data warehousing, and providing support. His academic and freelance projects involved areas like networking, web development, machine learning, Android apps, and Internet of Things.
This document contains the resume summary of Nageswara Rao Dasari. It outlines his 3.1 years of experience as a Software Engineer working on BIG DATA Technologies like Hadoop, HDFS, MapReduce, Hive and Pig. It also mentions his 1.4 years of experience in core Java and lists his technical skills like Java, SQL, JavaScript, CSS, Oracle, MySQL. It summarizes his most recent roles on projects for Barclays Bank and Target, where he performed tasks like data loading, writing MapReduce programs and Hive queries, and resolving JIRA tickets.
Scott Hovey has 17 years of experience developing automated tests, lab environments, and performance monitoring tools using languages like Python, Java, Perl, C++, and tools like Hyperic, JConsole, Linux/Windows native tools, and databases like Oracle, Postgres, MySQL. He has worked as a consultant for HeartMath Institute developing Cesium and Tableau maps, and as a Senior MTS Engineer and Tools and Performance Engineer at VMware maintaining virtual ESX code for testing and improving Tableau, Site Recovery Manager, and HP Quality Center performance and monitoring.
Mrigendra Kumar Bharti is a software developer with 2 years of experience developing web applications using Java. He has expertise in client/server development, data structures, algorithms, and Hadoop technologies like Spark, Kafka, Hive and HDFS. He has worked on projects involving network asset management, data ingestion using Apache NiFi, and integration with databases like Oracle, PostgreSQL and MongoDB. He has a Bachelor's degree in Information Technology and has received awards for his work. He maintains technical skills in Java, databases, frameworks and has experience with source control and build tools.
Machine Learning on Google Cloud with H2OSri Ambati
This document provides an overview of H2O.ai, a leading AI platform company. It discusses that H2O.ai was founded in 2012, is funded with $75 million, and has products including its open source H2O machine learning platform and its Driverless AI automated machine learning product. It also describes H2O.ai's leadership in the machine learning platform market according to Gartner, its team of 90 AI experts, and its global presence across several offices. Finally, it outlines H2O.ai's machine learning capabilities and how customers can use its platform and products.
Qamar Ali is a software developer currently working at Livastar.com in Hyderabad, India. He has over 3 years of experience in data science and software development. Some of his responsibilities have included automating report generation using Python scripts and Twitter data collection and analysis using Tweepy. He received his B.Tech in Computer Science from IIIT-Hyderabad in 2014 with a CGPA of 7.23. His skills include Python, Java, C, MySQL, Django and other web technologies.
This document contains Anil Kumar's resume. It summarizes his contact information, professional experience working with Hadoop and related technologies like MapReduce, Pig, and Hive. It also lists his technical skills and qualifications, including being a MapR certified Hadoop Professional. His work experience includes developing MapReduce algorithms, installing and configuring MapR Hadoop clusters, and working on projects for clients like Pfizer and American Express involving data analytics using Hadoop, Spark, and Hive.
Vishnu has over 5 years of experience in application development using Java and big data technologies like Hadoop. He has worked on projects involving web application development, data analytics using Hadoop components like HDFS, MapReduce, Pig and Hive. His skills include Java, J2EE, databases, version control and he has experience developing applications for both web and mobile. He is currently working as a Hadoop developer at Capgemini.
The document contains details about Nageswara Rao Dasari including his contact information, career objective, professional summary, technical summary, educational summary, and assignments. It outlines his 4+ years of experience as a Software Engineer working with technologies like Hadoop, Java, SQL, and tools like Eclipse. It provides details on 3 projects he worked on involving building platforms for banking customer data, retail customer data processing, and a web application.
Geetha Reddy has over 2 years of experience as a software developer specializing in web development using technologies like Python, Java, PHP, JavaScript, and databases like MySQL and PostgreSQL. She has worked on several projects including websites for SO2, Tahera Ecommerce, Xpertdata Works, and Shriya Engineers. Her responsibilities included designing responsive websites, integrating payment gateways, handling form data with PHP, and providing production support and maintenance. She has a Bachelor's Degree in Computer Science from GITAM University and is proficient in technologies like HTML, CSS, AngularJS, and frameworks like MVC.
Sanath Pabba has over 5 years of experience working with big data technologies like Hadoop, Spark, Hive, Pig, Kafka and NoSQL databases. He has expertise in data extraction, transformation and loading processes. Some of his responsibilities include writing Sqoop and Spark jobs to load and prepare data, developing automation scripts to monitor cluster utilization, and implementing validation rules for data quality. He has worked on various projects involving data warehousing, reporting, stream processing and analytics using technologies like SQL Server, Hive and Spark.
Introduction to Data Science with H2O- Mountain ViewSri Ambati
This document provides an overview of H2O.ai, an open source in-memory machine learning platform. It discusses that H2O.ai was founded in 2011 and is venture-backed, with a team of 37 people working on distributed systems for machine learning. It also summarizes that H2O provides easy to use APIs for Java, R, Python and other languages, and allows for scalable machine learning on large datasets using distributed algorithms to make full use of data without downsampling. Finally, it highlights how H2O works with other technologies like Spark, Hadoop, and HDFS to enable reading of large datasets for machine learning.
Introduction to data science with H2O-ChicagoSri Ambati
This document provides an overview of H2O.ai, an open source in-memory machine learning platform. It describes H2O.ai's product as an in-memory prediction engine, its team of 37 distributed systems engineers doing machine learning, and its headquarters in Mountain View, CA. It also provides details on how to use H2O with R and Python for scalable machine learning on large datasets across distributed systems.
Michal Malohlava talks about the PySparkling Water package for Spark and Python users.
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
This document discusses moving machine learning models from prototype to production. It outlines some common problems with the current workflow where moving to production often requires redevelopment from scratch. Some proposed solutions include using notebooks as APIs and developing analytics that are accessed via an API. It also discusses different data science platforms and architectures for building end-to-end machine learning systems, focusing on flexibility, security, testing and scalability for production environments. The document recommends a custom backend integrated with Spark via APIs as the best approach for the current project.
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-mlShubham Mallick
Shubham has over 7 years of experience in data analytics and engineering. He has extensive experience with technologies like MapR-Hadoop, Spark, Python, Hive, Kafka and machine learning algorithms. He is currently a senior data analyst where he builds data pipelines and analytics solutions. Previously he has led teams and taken on roles with responsibilities like requirement gathering, data modeling, ETL development, database administration and cloud migrations. He is pursuing an M.Tech in data science and has received several awards and certifications for his work.
Shabarish Kesa has experience as a software developer and systems engineer. They have a Master's in Information Technology from the University of Cincinnati with a 4.0 GPA. As a developer, they led a team migrating an application from AngularJs to React Js, decreasing onboarding time by 75% through automated scripts. As an engineer, they developed an Angular web app to decrease issue reporting time by 50% and quickly fixed a production server bug. Their projects include a resume application using React, Node and PostgreSQL and a Burger Builder app using React and Firebase.
Big Data with hadoop, Spark and BigQuery (Google cloud next Extended 2017 Kar...Imam Raza
Google Next Extended (https://cloudnext.withgoogle.com/) is an annual Google event focusing on Google cloud technologies. This presentation is from tech talk held in Google Next Extended 2017 Karachi event
This document is a resume for Shabarish Kesa summarizing their technical skills and work experience. They have experience developing full stack applications using technologies like React, Node, Java, Python, and databases like MongoDB and PostgreSQL. Their most recent role involved developing a data visualization application for asset monitoring using React, Node, and Docker. They have also worked on migrating applications from Angular to React and writing tests achieving high code coverage. They have a Master's degree in Information Technology and have contributed to open source projects on GitHub.
Tejas Bichave is a software professional with over 3 years of experience in Python, Java, and testing tools like Postman. He has worked on projects involving resource adapters, advertisement portals, auto provisioning servers, and cryptographic algorithm development. He holds an M-Tech in computer science and has published papers on caching techniques. He is seeking a new role where he can apply and grow his technical skills.
Sri Harsha Vajjhala is a software engineer with 2 years of experience in application development using technologies like Java, Spring Framework, Struts, Hibernate, and databases like Oracle and SQL Server. He has a B.Tech in computer science and is looking to join a high-tech company. His work experience includes projects for Hewlett Packard in areas like product lifecycle management and supply chain systems. He is proficient in languages like Java, C, and technologies like web services and has expertise in software development methodologies like Agile.
🔥 Top 5 Skills For Data Engineer In 2023 | Data Engineer Skills Required For ...Simplilearn
This video is based on Top 5 Skills For Data Engineer In 2023. In this video, we delve into the role of Data Engineers and the future salary trends. Learn about key skills like Big Data technologies, Data Modeling, and proficiency in programming languages that are crucial for excelling in the field. Stay ahead by mastering the expertise needed to thrive as a Data Engineer in the dynamic landscape of data-driven decision-making.
Minh N. Pham is currently pursuing a BS in Computer Science and Mathematical Sciences at Worcester Polytechnic Institute. He has experience in data science, machine learning, and full-stack web development. Some of his projects include building predictive models for income and stock price movement using Python and developing web applications using technologies like Ruby on Rails, Java, and Vue.js. He has worked on data analysis and machine learning problems in internships and research assistant roles.
Machine Learning on Google Cloud with H2OSri Ambati
This document provides an overview of H2O.ai, a leading AI platform company. It discusses that H2O.ai was founded in 2012, is funded with $75 million, and has products including its open source H2O machine learning platform and its Driverless AI automated machine learning product. It also describes H2O.ai's leadership in the machine learning platform market according to Gartner, its team of 90 AI experts, and its global presence across several offices. Finally, it outlines H2O.ai's machine learning capabilities and how customers can use its platform and products.
Qamar Ali is a software developer currently working at Livastar.com in Hyderabad, India. He has over 3 years of experience in data science and software development. Some of his responsibilities have included automating report generation using Python scripts and Twitter data collection and analysis using Tweepy. He received his B.Tech in Computer Science from IIIT-Hyderabad in 2014 with a CGPA of 7.23. His skills include Python, Java, C, MySQL, Django and other web technologies.
This document contains Anil Kumar's resume. It summarizes his contact information, professional experience working with Hadoop and related technologies like MapReduce, Pig, and Hive. It also lists his technical skills and qualifications, including being a MapR certified Hadoop Professional. His work experience includes developing MapReduce algorithms, installing and configuring MapR Hadoop clusters, and working on projects for clients like Pfizer and American Express involving data analytics using Hadoop, Spark, and Hive.
Vishnu has over 5 years of experience in application development using Java and big data technologies like Hadoop. He has worked on projects involving web application development, data analytics using Hadoop components like HDFS, MapReduce, Pig and Hive. His skills include Java, J2EE, databases, version control and he has experience developing applications for both web and mobile. He is currently working as a Hadoop developer at Capgemini.
The document contains details about Nageswara Rao Dasari including his contact information, career objective, professional summary, technical summary, educational summary, and assignments. It outlines his 4+ years of experience as a Software Engineer working with technologies like Hadoop, Java, SQL, and tools like Eclipse. It provides details on 3 projects he worked on involving building platforms for banking customer data, retail customer data processing, and a web application.
Geetha Reddy has over 2 years of experience as a software developer specializing in web development using technologies like Python, Java, PHP, JavaScript, and databases like MySQL and PostgreSQL. She has worked on several projects including websites for SO2, Tahera Ecommerce, Xpertdata Works, and Shriya Engineers. Her responsibilities included designing responsive websites, integrating payment gateways, handling form data with PHP, and providing production support and maintenance. She has a Bachelor's Degree in Computer Science from GITAM University and is proficient in technologies like HTML, CSS, AngularJS, and frameworks like MVC.
Sanath Pabba has over 5 years of experience working with big data technologies like Hadoop, Spark, Hive, Pig, Kafka and NoSQL databases. He has expertise in data extraction, transformation and loading processes. Some of his responsibilities include writing Sqoop and Spark jobs to load and prepare data, developing automation scripts to monitor cluster utilization, and implementing validation rules for data quality. He has worked on various projects involving data warehousing, reporting, stream processing and analytics using technologies like SQL Server, Hive and Spark.
Introduction to Data Science with H2O- Mountain ViewSri Ambati
This document provides an overview of H2O.ai, an open source in-memory machine learning platform. It discusses that H2O.ai was founded in 2011 and is venture-backed, with a team of 37 people working on distributed systems for machine learning. It also summarizes that H2O provides easy to use APIs for Java, R, Python and other languages, and allows for scalable machine learning on large datasets using distributed algorithms to make full use of data without downsampling. Finally, it highlights how H2O works with other technologies like Spark, Hadoop, and HDFS to enable reading of large datasets for machine learning.
Introduction to data science with H2O-ChicagoSri Ambati
This document provides an overview of H2O.ai, an open source in-memory machine learning platform. It describes H2O.ai's product as an in-memory prediction engine, its team of 37 distributed systems engineers doing machine learning, and its headquarters in Mountain View, CA. It also provides details on how to use H2O with R and Python for scalable machine learning on large datasets across distributed systems.
Michal Malohlava talks about the PySparkling Water package for Spark and Python users.
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
This document discusses moving machine learning models from prototype to production. It outlines some common problems with the current workflow where moving to production often requires redevelopment from scratch. Some proposed solutions include using notebooks as APIs and developing analytics that are accessed via an API. It also discusses different data science platforms and architectures for building end-to-end machine learning systems, focusing on flexibility, security, testing and scalability for production environments. The document recommends a custom backend integrated with Spark via APIs as the best approach for the current project.
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-mlShubham Mallick
Shubham has over 7 years of experience in data analytics and engineering. He has extensive experience with technologies like MapR-Hadoop, Spark, Python, Hive, Kafka and machine learning algorithms. He is currently a senior data analyst where he builds data pipelines and analytics solutions. Previously he has led teams and taken on roles with responsibilities like requirement gathering, data modeling, ETL development, database administration and cloud migrations. He is pursuing an M.Tech in data science and has received several awards and certifications for his work.
Shabarish Kesa has experience as a software developer and systems engineer. They have a Master's in Information Technology from the University of Cincinnati with a 4.0 GPA. As a developer, they led a team migrating an application from AngularJs to React Js, decreasing onboarding time by 75% through automated scripts. As an engineer, they developed an Angular web app to decrease issue reporting time by 50% and quickly fixed a production server bug. Their projects include a resume application using React, Node and PostgreSQL and a Burger Builder app using React and Firebase.
Big Data with hadoop, Spark and BigQuery (Google cloud next Extended 2017 Kar...Imam Raza
Google Next Extended (https://cloudnext.withgoogle.com/) is an annual Google event focusing on Google cloud technologies. This presentation is from tech talk held in Google Next Extended 2017 Karachi event
This document is a resume for Shabarish Kesa summarizing their technical skills and work experience. They have experience developing full stack applications using technologies like React, Node, Java, Python, and databases like MongoDB and PostgreSQL. Their most recent role involved developing a data visualization application for asset monitoring using React, Node, and Docker. They have also worked on migrating applications from Angular to React and writing tests achieving high code coverage. They have a Master's degree in Information Technology and have contributed to open source projects on GitHub.
Tejas Bichave is a software professional with over 3 years of experience in Python, Java, and testing tools like Postman. He has worked on projects involving resource adapters, advertisement portals, auto provisioning servers, and cryptographic algorithm development. He holds an M-Tech in computer science and has published papers on caching techniques. He is seeking a new role where he can apply and grow his technical skills.
Sri Harsha Vajjhala is a software engineer with 2 years of experience in application development using technologies like Java, Spring Framework, Struts, Hibernate, and databases like Oracle and SQL Server. He has a B.Tech in computer science and is looking to join a high-tech company. His work experience includes projects for Hewlett Packard in areas like product lifecycle management and supply chain systems. He is proficient in languages like Java, C, and technologies like web services and has expertise in software development methodologies like Agile.
🔥 Top 5 Skills For Data Engineer In 2023 | Data Engineer Skills Required For ...Simplilearn
This video is based on Top 5 Skills For Data Engineer In 2023. In this video, we delve into the role of Data Engineers and the future salary trends. Learn about key skills like Big Data technologies, Data Modeling, and proficiency in programming languages that are crucial for excelling in the field. Stay ahead by mastering the expertise needed to thrive as a Data Engineer in the dynamic landscape of data-driven decision-making.
Minh N. Pham is currently pursuing a BS in Computer Science and Mathematical Sciences at Worcester Polytechnic Institute. He has experience in data science, machine learning, and full-stack web development. Some of his projects include building predictive models for income and stock price movement using Python and developing web applications using technologies like Ruby on Rails, Java, and Vue.js. He has worked on data analysis and machine learning problems in internships and research assistant roles.
PyData London 2024: Mistakes were made (Dr. Rebecca Bilbro)Rebecca Bilbro
To honor ten years of PyData London, join Dr. Rebecca Bilbro as she takes us back in time to reflect on a little over ten years working as a data scientist. One of the many renegade PhDs who joined the fledgling field of data science of the 2010's, Rebecca will share lessons learned the hard way, often from watching data science projects go sideways and learning to fix broken things. Through the lens of these canon events, she'll identify some of the anti-patterns and red flags she's learned to steer around.
Build applications with generative AI on Google CloudMárton Kodok
We will explore Vertex AI - Model Garden powered experiences, we are going to learn more about the integration of these generative AI APIs. We are going to see in action what the Gemini family of generative models are for developers to build and deploy AI-driven applications. Vertex AI includes a suite of foundation models, these are referred to as the PaLM and Gemini family of generative ai models, and they come in different versions. We are going to cover how to use via API to: - execute prompts in text and chat - cover multimodal use cases with image prompts. - finetune and distill to improve knowledge domains - run function calls with foundation models to optimize them for specific tasks. At the end of the session, developers will understand how to innovate with generative AI and develop apps using the generative ai industry trends.
Did you know that drowning is a leading cause of unintentional death among young children? According to recent data, children aged 1-4 years are at the highest risk. Let's raise awareness and take steps to prevent these tragic incidents. Supervision, barriers around pools, and learning CPR can make a difference. Stay safe this summer!
We are pleased to share with you the latest VCOSA statistical report on the cotton and yarn industry for the month of May 2024.
Starting from January 2024, the full weekly and monthly reports will only be available for free to VCOSA members. To access the complete weekly report with figures, charts, and detailed analysis of the cotton fiber market in the past week, interested parties are kindly requested to contact VCOSA to subscribe to the newsletter.
06-18-2024-Princeton Meetup-Introduction to MilvusTimothy Spann
06-18-2024-Princeton Meetup-Introduction to Milvus
tim.spann@zilliz.com
https://www.linkedin.com/in/timothyspann/
https://x.com/paasdev
https://github.com/tspannhw
https://github.com/milvus-io/milvus
Get Milvused!
https://milvus.io/
Read my Newsletter every week!
https://github.com/tspannhw/FLiPStackWeekly/blob/main/142-17June2024.md
For more cool Unstructured Data, AI and Vector Database videos check out the Milvus vector database videos here
https://www.youtube.com/@MilvusVectorDatabase/videos
Unstructured Data Meetups -
https://www.meetup.com/unstructured-data-meetup-new-york/
https://lu.ma/calendar/manage/cal-VNT79trvj0jS8S7
https://www.meetup.com/pro/unstructureddata/
https://zilliz.com/community/unstructured-data-meetup
https://zilliz.com/event
Twitter/X: https://x.com/milvusio https://x.com/paasdev
LinkedIn: https://www.linkedin.com/company/zilliz/ https://www.linkedin.com/in/timothyspann/
GitHub: https://github.com/milvus-io/milvus https://github.com/tspannhw
Invitation to join Discord: https://discord.com/invite/FjCMmaJng6
Blogs: https://milvusio.medium.com/ https://www.opensourcevectordb.cloud/ https://medium.com/@tspann
Expand LLMs' knowledge by incorporating external data sources into LLMs and your AI applications.
06-20-2024-AI Camp Meetup-Unstructured Data and Vector DatabasesTimothy Spann
Tech Talk: Unstructured Data and Vector Databases
Speaker: Tim Spann (Zilliz)
Abstract: In this session, I will discuss the unstructured data and the world of vector databases, we will see how they different from traditional databases. In which cases you need one and in which you probably don’t. I will also go over Similarity Search, where do you get vectors from and an example of a Vector Database Architecture. Wrapping up with an overview of Milvus.
Introduction
Unstructured data, vector databases, traditional databases, similarity search
Vectors
Where, What, How, Why Vectors? We’ll cover a Vector Database Architecture
Introducing Milvus
What drives Milvus' Emergence as the most widely adopted vector database
Hi Unstructured Data Friends!
I hope this video had all the unstructured data processing, AI and Vector Database demo you needed for now. If not, there’s a ton more linked below.
My source code is available here
https://github.com/tspannhw/
Let me know in the comments if you liked what you saw, how I can improve and what should I show next? Thanks, hope to see you soon at a Meetup in Princeton, Philadelphia, New York City or here in the Youtube Matrix.
Get Milvused!
https://milvus.io/
Read my Newsletter every week!
https://github.com/tspannhw/FLiPStackWeekly/blob/main/141-10June2024.md
For more cool Unstructured Data, AI and Vector Database videos check out the Milvus vector database videos here
https://www.youtube.com/@MilvusVectorDatabase/videos
Unstructured Data Meetups -
https://www.meetup.com/unstructured-data-meetup-new-york/
https://lu.ma/calendar/manage/cal-VNT79trvj0jS8S7
https://www.meetup.com/pro/unstructureddata/
https://zilliz.com/community/unstructured-data-meetup
https://zilliz.com/event
Twitter/X: https://x.com/milvusio https://x.com/paasdev
LinkedIn: https://www.linkedin.com/company/zilliz/ https://www.linkedin.com/in/timothyspann/
GitHub: https://github.com/milvus-io/milvus https://github.com/tspannhw
Invitation to join Discord: https://discord.com/invite/FjCMmaJng6
Blogs: https://milvusio.medium.com/ https://www.opensourcevectordb.cloud/ https://medium.com/@tspann
https://www.meetup.com/unstructured-data-meetup-new-york/events/301383476/?slug=unstructured-data-meetup-new-york&eventId=301383476
https://www.aicamp.ai/event/eventdetails/W2024062014
06-20-2024-AI Camp Meetup-Unstructured Data and Vector Databases
Kajal_Rathi_Resume.pdf
1. Work History
Education
Certifications
Associate Engineer
Kajal Rathi
Contact
Address
Hyderabad, India 500032
Phone
7483445831
E-mail
kajalrathi1998@gmail.com
Skills
Python
MySQL, NoSQL(MongoDB),
PostgreSQL
Machine Learning, Pandas,
Numpy, Matplotlib, Scikit-
Learn, Seaborn, SDV
(synthetic data vault)
Power BI
Flask
Tools : Swagger, Postman,
Docker, GitHub
Apache Spark, Hadoop, Hive,
Airflow
Azure fundamentals
Working as an Associate Engineer with hands-on experience in Python,
Machine Learning, Database Technologies, Flask and Big data
technologies like Apache Spark, Airflow
Associate Engineer
Innominds Software Pvt Ltd, Hyderabad
Project Name: CorGr-DigiWise Airflow Requirements
Task for this project is to substitute current Perl scripts
and Cron jobs with Python Scripts and Airflow.
Digiwise possesses approximately 70 project modules
that utilize Perl Scripts and are executed through shell
scripts to gather and distribute files on MapR cluster
node.
Project Name: AtlPar-Cash Recon Project
This project involved converting numerous bank PDFs
to Excel provided by client, which were from different
states in US.
Python scripts were run on Excel files to validate
accuracy of work.
Data was then verified and modified using human
intelligence to fill in any missing information.
Project Name: Synthetic Data Generator
This project focuses on harnessing power of Synthetic
Data Generator Library (SDV) to create realistic
synthetic data from simple dataset.
By employing advanced algorithms, SDV Library will
generate sample rows that closely resemble statistical
characteristics of original data.
Project Name: Coin-Annotation
Objective of this project is to offer data services to
machine learning team in order to create NLP models
that could help Chat bot to increase it's efficiency by
providing high-quality, human-verified annotations on
user queries and responses.
2021-11 -
Current
PG Diploma Course in Big Data Analytics
CDAC Sunbeam - Pune, India
2021-05 -
2021-10
BE
Basaveshwar Engineering College - Bagalkot, India
2016-08 -
2020-09
Data Science with Python by Simplilearn
2022-04
Python Libraries for Data Science by Simplilearn
2023-02
Azure Fundamentals by Simplilearn
2023-03