AAAI-19 paper: Neural Relation ExtractionWithin and Across Sentence Boundaries
Authors: Pankaj Gupta, Subburam Rajaram, Hinrich Schuetze and Thomas Runkler
TOP 5 Most View Article From Academia in 2019sipij
TOP 5 Most View Article From Academia in 2019
Signal & Image Processing : An International Journal (SIPIJ)
ISSN : 0976 - 710X (Online) ; 2229 - 3922 (print)
http://www.airccse.org/journal/sipij/index.html
Big Data Analytics for Obesity PredictionAhsan Bilal
FS is essential for the analysis of datasets with millions of features. In such a context, Big Data tools are paramount, but the use of standard ML models is limited for data with such low instances to features ratios. Since Apache Spark 2.0 is unable to cope with our dataset containing ≈ 0.74 million features, we propose here a pipeline to solve this problem using partitioning strategies, both vertical and horizontal.
Signal with amplitude outside the range accepted by the sensor
Enter damaged image. get restored image
Post-processing of damaged images at the moment of acquisition
sRGB Color Space
Restoration with aesthetic purposes
What we expect: Color correction,Texture Edges / Lines / Image gradient; Structures;
Modeling based on convolutional neural networks
TOP 5 Most View Article From Academia in 2019sipij
TOP 5 Most View Article From Academia in 2019
Signal & Image Processing : An International Journal (SIPIJ)
ISSN : 0976 - 710X (Online) ; 2229 - 3922 (print)
http://www.airccse.org/journal/sipij/index.html
Big Data Analytics for Obesity PredictionAhsan Bilal
FS is essential for the analysis of datasets with millions of features. In such a context, Big Data tools are paramount, but the use of standard ML models is limited for data with such low instances to features ratios. Since Apache Spark 2.0 is unable to cope with our dataset containing ≈ 0.74 million features, we propose here a pipeline to solve this problem using partitioning strategies, both vertical and horizontal.
Signal with amplitude outside the range accepted by the sensor
Enter damaged image. get restored image
Post-processing of damaged images at the moment of acquisition
sRGB Color Space
Restoration with aesthetic purposes
What we expect: Color correction,Texture Edges / Lines / Image gradient; Structures;
Modeling based on convolutional neural networks
textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...Pankaj Gupta, PhD
ICLR 2019 conference paper.
Improving Topic Modeling with language structures (e.g., word ordering, local context, syntactic and semantic information); Neural Composite Generative Model: A Neural Topic + A Neural Language Model; Expressing both short-and long-range dependencies
Invited Talk @ Google AI, New York City USA.
Talk includes: Neural Relation Extraction (AAAI-2019 paper) and Neural Topic Modeling (AAAI-2019 and ICLR-2019 papers).
Neural Relation ExtractionWithin and Across Sentence BoundariesPankaj Gupta, PhD
AAAI-19 paper: Neural Relation ExtractionWithin and Across Sentence Boundaries
Authors: Pankaj Gupta, Subburam Rajaram, Hinrich Schuetze and Thomas Runkler
textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE...Pankaj Gupta, PhD
Unified neural model of topic and language modeling to introduce language structure in topic models for contextualized topic vectors
Representation learning for short text and long text documents
Generate contextualized topic vectors of variable document sizes, even in limited context settings.
Neural topic models with word embeddings
PhD in Machine Learning / Deep Learning / Natural Language Processing
Profile: https://www.linkedin.com/in/pankaj-gupta-6b95bb17/
Research Contributions: https://scholar.google.com/citations?user=_YjIJF0AAAAJ&hl=en
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...Pankaj Gupta, PhD
Analyzing and Interpreting Neural network (RNNs) for natural language text, especially in relation extraction.
Poster presented in the workshop #BlackBoxNLP at EMNLP 2018, Brussels Belgium
Lecture 07: Representation and Distributional Learning by Pankaj GuptaPankaj Gupta, PhD
Lecture on "Representation and Distributional Learning" at University of Munich (LMU), as part of "Deep Learning & AI" lecture series.
Includes: Fundamentals of representation learning, probabilistic graphical models, generative modeling, unsupervised learning, RBMs, RSM, DocNADE, etc.
Lecture 05: Recurrent Neural Networks / Deep Learning by Pankaj GuptaPankaj Gupta, PhD
Lecture on Recurrent Neural Network at University of Munich (LMU), as part of Deep Learning & AI lecture series.
Includes: Fundamentals of RNNs, Need for LSTM and GRU.
Analysis insight about a Flyball dog competition team's performanceroli9797
Insight of my analysis about a Flyball dog competition team's last year performance. Find more: https://github.com/rolandnagy-ds/flyball_race_analysis/tree/main
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...Pankaj Gupta, PhD
ICLR 2019 conference paper.
Improving Topic Modeling with language structures (e.g., word ordering, local context, syntactic and semantic information); Neural Composite Generative Model: A Neural Topic + A Neural Language Model; Expressing both short-and long-range dependencies
Invited Talk @ Google AI, New York City USA.
Talk includes: Neural Relation Extraction (AAAI-2019 paper) and Neural Topic Modeling (AAAI-2019 and ICLR-2019 papers).
Neural Relation ExtractionWithin and Across Sentence BoundariesPankaj Gupta, PhD
AAAI-19 paper: Neural Relation ExtractionWithin and Across Sentence Boundaries
Authors: Pankaj Gupta, Subburam Rajaram, Hinrich Schuetze and Thomas Runkler
textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE...Pankaj Gupta, PhD
Unified neural model of topic and language modeling to introduce language structure in topic models for contextualized topic vectors
Representation learning for short text and long text documents
Generate contextualized topic vectors of variable document sizes, even in limited context settings.
Neural topic models with word embeddings
PhD in Machine Learning / Deep Learning / Natural Language Processing
Profile: https://www.linkedin.com/in/pankaj-gupta-6b95bb17/
Research Contributions: https://scholar.google.com/citations?user=_YjIJF0AAAAJ&hl=en
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...Pankaj Gupta, PhD
Analyzing and Interpreting Neural network (RNNs) for natural language text, especially in relation extraction.
Poster presented in the workshop #BlackBoxNLP at EMNLP 2018, Brussels Belgium
Lecture 07: Representation and Distributional Learning by Pankaj GuptaPankaj Gupta, PhD
Lecture on "Representation and Distributional Learning" at University of Munich (LMU), as part of "Deep Learning & AI" lecture series.
Includes: Fundamentals of representation learning, probabilistic graphical models, generative modeling, unsupervised learning, RBMs, RSM, DocNADE, etc.
Lecture 05: Recurrent Neural Networks / Deep Learning by Pankaj GuptaPankaj Gupta, PhD
Lecture on Recurrent Neural Network at University of Munich (LMU), as part of Deep Learning & AI lecture series.
Includes: Fundamentals of RNNs, Need for LSTM and GRU.
Analysis insight about a Flyball dog competition team's performanceroli9797
Insight of my analysis about a Flyball dog competition team's last year performance. Find more: https://github.com/rolandnagy-ds/flyball_race_analysis/tree/main
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdfEnterprise Wired
In this guide, we'll explore the key considerations and features to look for when choosing a Trusted analytics platform that meets your organization's needs and delivers actionable intelligence you can trust.
Adjusting OpenMP PageRank : SHORT REPORT / NOTESSubhajit Sahu
For massive graphs that fit in RAM, but not in GPU memory, it is possible to take
advantage of a shared memory system with multiple CPUs, each with multiple cores, to
accelerate pagerank computation. If the NUMA architecture of the system is properly taken
into account with good vertex partitioning, the speedup can be significant. To take steps in
this direction, experiments are conducted to implement pagerank in OpenMP using two
different approaches, uniform and hybrid. The uniform approach runs all primitives required
for pagerank in OpenMP mode (with multiple threads). On the other hand, the hybrid
approach runs certain primitives in sequential mode (i.e., sumAt, multiply).
Global Situational Awareness of A.I. and where its headedvikram sood
You can see the future first in San Francisco.
Over the past year, the talk of the town has shifted from $10 billion compute clusters to $100 billion clusters to trillion-dollar clusters. Every six months another zero is added to the boardroom plans. Behind the scenes, there’s a fierce scramble to secure every power contract still available for the rest of the decade, every voltage transformer that can possibly be procured. American big business is gearing up to pour trillions of dollars into a long-unseen mobilization of American industrial might. By the end of the decade, American electricity production will have grown tens of percent; from the shale fields of Pennsylvania to the solar farms of Nevada, hundreds of millions of GPUs will hum.
The AGI race has begun. We are building machines that can think and reason. By 2025/26, these machines will outpace college graduates. By the end of the decade, they will be smarter than you or I; we will have superintelligence, in the true sense of the word. Along the way, national security forces not seen in half a century will be un-leashed, and before long, The Project will be on. If we’re lucky, we’ll be in an all-out race with the CCP; if we’re unlucky, an all-out war.
Everyone is now talking about AI, but few have the faintest glimmer of what is about to hit them. Nvidia analysts still think 2024 might be close to the peak. Mainstream pundits are stuck on the wilful blindness of “it’s just predicting the next word”. They see only hype and business-as-usual; at most they entertain another internet-scale technological change.
Before long, the world will wake up. But right now, there are perhaps a few hundred people, most of them in San Francisco and the AI labs, that have situational awareness. Through whatever peculiar forces of fate, I have found myself amongst them. A few years ago, these people were derided as crazy—but they trusted the trendlines, which allowed them to correctly predict the AI advances of the past few years. Whether these people are also right about the next few years remains to be seen. But these are very smart people—the smartest people I have ever met—and they are the ones building this technology. Perhaps they will be an odd footnote in history, or perhaps they will go down in history like Szilard and Oppenheimer and Teller. If they are seeing the future even close to correctly, we are in for a wild ride.
Let me tell you what we see.
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdfGetInData
Recently we have observed the rise of open-source Large Language Models (LLMs) that are community-driven or developed by the AI market leaders, such as Meta (Llama3), Databricks (DBRX) and Snowflake (Arctic). On the other hand, there is a growth in interest in specialized, carefully fine-tuned yet relatively small models that can efficiently assist programmers in day-to-day tasks. Finally, Retrieval-Augmented Generation (RAG) architectures have gained a lot of traction as the preferred approach for LLMs context and prompt augmentation for building conversational SQL data copilots, code copilots and chatbots.
In this presentation, we will show how we built upon these three concepts a robust Data Copilot that can help to democratize access to company data assets and boost performance of everyone working with data platforms.
Why do we need yet another (open-source ) Copilot?
How can we build one?
Architecture and evaluation