Elaborer la stratégie de communication - tout ce qu'il faut savoir.Amar LAKEL, PhD
Depuis plus de 15 ans, je conseille et forme les responsables et futurs responsables de communication à la stratégie de communication. J'ai voulu dans cette présentation, longue et lourde mais néanmoins synthétique, résumer 15 ans d'expériences en stratégie de communication, campagne de communication, stratégie de marque et mediaplanning
Méthodologie - Comment construire une stratégie de communication touristique ...Cap'Com
François Genin - directeur général de Leon Travel & Tourism
Comment construire une stratégie de communication touristique ? Sur quels outils s'appuyer, notamment vis-à-vis du marketing territorial ? Quelles sont les étapes et les acteurs à ne pas oublier ?
Enjeux de recherches pour les Humanités DigitalesAmar LAKEL, PhD
Les humanités digitales sont un énorme défi aux SHS : Introduire les TICs dans l’extraction, l’archivage, l’analyse automatique, la qualification des corpus, la visualisation des données… Nous voulons un projet pour fédérer les énergies ouvertes et innovantes afin de résoudre les défis de compréhension d'une communication complexe au cœur des sociétés.
Elaborer la stratégie de communication - tout ce qu'il faut savoir.Amar LAKEL, PhD
Depuis plus de 15 ans, je conseille et forme les responsables et futurs responsables de communication à la stratégie de communication. J'ai voulu dans cette présentation, longue et lourde mais néanmoins synthétique, résumer 15 ans d'expériences en stratégie de communication, campagne de communication, stratégie de marque et mediaplanning
Méthodologie - Comment construire une stratégie de communication touristique ...Cap'Com
François Genin - directeur général de Leon Travel & Tourism
Comment construire une stratégie de communication touristique ? Sur quels outils s'appuyer, notamment vis-à-vis du marketing territorial ? Quelles sont les étapes et les acteurs à ne pas oublier ?
Enjeux de recherches pour les Humanités DigitalesAmar LAKEL, PhD
Les humanités digitales sont un énorme défi aux SHS : Introduire les TICs dans l’extraction, l’archivage, l’analyse automatique, la qualification des corpus, la visualisation des données… Nous voulons un projet pour fédérer les énergies ouvertes et innovantes afin de résoudre les défis de compréhension d'une communication complexe au cœur des sociétés.
Le web administratif en France - Rapport Forum CPP 2009Amar LAKEL, PhD
Combien sont-ils en France, cette année-là, à comprendre l’ampleur
de l’événement, quand le 11 janvier 1994, Al Gore, vice-président des États-
Unis, défend, au Super Highway Summit de Los Angeles, la National
Information Infrastructure (NII), lancée en septembre 1993, devant une
assemblée d’entrepreneurs des nouvelles technologies. Al Gore souhaite non
seulement y expliquer le programme, mais y lancer un appel à l’initiative
privée sous couvert de l’Etat. Entre interventionnisme et pur libéralisme, il
souhaite redéfinir un nouveau modèle de coopération Etat/Privé : celui d’une
puissance économique mondiale s’appuyant sur le libre marché intérieur et
conduit par le gouvernement fédéral, chargé de pourfendre les barrières
commerciales qui tenteraient d’endiguer le « free flow of information »
(MATTELART, 2003). Ce discours fait suite à celui de Washington, devant le
National Press Club, où il avait insisté sur les principes qui guidaient le
programme gouvernemental. Ces deux discours vont former la base de la
célèbre intervention de Gore devant l’International Communication Union, le
21 mars 1994. Discours plus politique et plus consensuel, il tente d’engager la
communauté internationale dans la mise en place d’une Global Information
Infrastructure (GIS), et ce sur la même base que la politique américaine. Ce
discours connaîtra une postface, le 17 octobre 1994, au CenterCom,
réunissant les plus grands intérêts privés de la planète en matière de nouvelles
technologies de l’information et de la communication...
Data Marketing, l’ère de l’intelligence numérique ?Amar LAKEL, PhD
Après l’effet de mode du Big Data, quel bilan des solution professionnelle pour la P&ME dans l'usage des datas au service du Digital Marketing.
CCI, 28 mai 2014. Suivez notre actualité : https://www.facebook.com/calltoactdatamarketing
Mobilité, rupture ou continuité dans la communication électronique publique ?Amar LAKEL, PhD
Si l'on parle encore de « nouvelles » technologies de l’information de la communication (NTIC), c'est que l'innovation en matière de dispositifs de communication électronique a connu trois phénomènes majeurs qui se sont conjugués.
En premier lieu, les terminaux informatiques se sont miniaturisés tout en augmentant sans cesse leur puissance de calcul.
Ensuite, les canaux de télécommunication continuent de connaître une convergence de tous les formats de données pour aboutir aux solutions triple play (audio-vidéo, téléphonie, données textuelles…).
Enfin la diminution des coûts de production des produits de télécommunication a permis la massification des usages, en faisant exploser la consommation des services de communication.
Aujourd'hui, ce processus continu d'innovation semble atteindre un seuil critique que l'on résume généralement sous le terme de mobilité, étape la plus récente de la société de l’information.
Piloter sa campagne digital marketing de A à ZAmar LAKEL, PhD
Tout mais vraiment tout pour piloter une campagne marketing digitale. Ce cours est pensé pour couvrir l'ensemble des techniques de marketing digital. Emailing, Social Media, SEO, Ads, etc. Tout y passe comme une to do list au service de votre plan de com digital
Open Entrepreneurship_Teigland, Di Gangi, YetisRobin Teigland
Our presentation at the Innovation and Market Creation in and around Virtual Worlds in May 2012 at Copenhagen Business School. More information here: http://nordicworlds.net/2012/04/13/innovation-and-market-creation-in-and-around-virtual-worlds-2/.
See the WEBCAST as well!! mms://wmedia.it.su.se/SUB/NordLib/3.wmv
Presentation at Nordlib 2.0 in Stockholm, November 21th 2008
http://www.nordlib20.org/programme/
The modern library web environment consists of multiple content sources and applications that perform essential functions that often overlap and could potentially create a fractured user experience. For example, content in a library’s Drupal website may be replicated in LibGuides or WordPress blogs. Search functionality in a discovery platform may be replicated in a federated search tool or the ILS OPAC. This presentation provides tips, tackles technical and political challenges to building a single web experience for users, discusses solutions and use of APIs (application programming interfaces), provides concrete examples, and more.
LiMoSINe Press kit introduces this project that integrates the studies of leading researchers over diverse topics with a view to enable new kinds of language-based technology search. Now we are developing 5 demonstrators: ORMA, ThemeStreams, FlickrDemo, DEESSE and Streamwatchr. http://limosine-project.eu/
Le web administratif en France - Rapport Forum CPP 2009Amar LAKEL, PhD
Combien sont-ils en France, cette année-là, à comprendre l’ampleur
de l’événement, quand le 11 janvier 1994, Al Gore, vice-président des États-
Unis, défend, au Super Highway Summit de Los Angeles, la National
Information Infrastructure (NII), lancée en septembre 1993, devant une
assemblée d’entrepreneurs des nouvelles technologies. Al Gore souhaite non
seulement y expliquer le programme, mais y lancer un appel à l’initiative
privée sous couvert de l’Etat. Entre interventionnisme et pur libéralisme, il
souhaite redéfinir un nouveau modèle de coopération Etat/Privé : celui d’une
puissance économique mondiale s’appuyant sur le libre marché intérieur et
conduit par le gouvernement fédéral, chargé de pourfendre les barrières
commerciales qui tenteraient d’endiguer le « free flow of information »
(MATTELART, 2003). Ce discours fait suite à celui de Washington, devant le
National Press Club, où il avait insisté sur les principes qui guidaient le
programme gouvernemental. Ces deux discours vont former la base de la
célèbre intervention de Gore devant l’International Communication Union, le
21 mars 1994. Discours plus politique et plus consensuel, il tente d’engager la
communauté internationale dans la mise en place d’une Global Information
Infrastructure (GIS), et ce sur la même base que la politique américaine. Ce
discours connaîtra une postface, le 17 octobre 1994, au CenterCom,
réunissant les plus grands intérêts privés de la planète en matière de nouvelles
technologies de l’information et de la communication...
Data Marketing, l’ère de l’intelligence numérique ?Amar LAKEL, PhD
Après l’effet de mode du Big Data, quel bilan des solution professionnelle pour la P&ME dans l'usage des datas au service du Digital Marketing.
CCI, 28 mai 2014. Suivez notre actualité : https://www.facebook.com/calltoactdatamarketing
Mobilité, rupture ou continuité dans la communication électronique publique ?Amar LAKEL, PhD
Si l'on parle encore de « nouvelles » technologies de l’information de la communication (NTIC), c'est que l'innovation en matière de dispositifs de communication électronique a connu trois phénomènes majeurs qui se sont conjugués.
En premier lieu, les terminaux informatiques se sont miniaturisés tout en augmentant sans cesse leur puissance de calcul.
Ensuite, les canaux de télécommunication continuent de connaître une convergence de tous les formats de données pour aboutir aux solutions triple play (audio-vidéo, téléphonie, données textuelles…).
Enfin la diminution des coûts de production des produits de télécommunication a permis la massification des usages, en faisant exploser la consommation des services de communication.
Aujourd'hui, ce processus continu d'innovation semble atteindre un seuil critique que l'on résume généralement sous le terme de mobilité, étape la plus récente de la société de l’information.
Piloter sa campagne digital marketing de A à ZAmar LAKEL, PhD
Tout mais vraiment tout pour piloter une campagne marketing digitale. Ce cours est pensé pour couvrir l'ensemble des techniques de marketing digital. Emailing, Social Media, SEO, Ads, etc. Tout y passe comme une to do list au service de votre plan de com digital
Open Entrepreneurship_Teigland, Di Gangi, YetisRobin Teigland
Our presentation at the Innovation and Market Creation in and around Virtual Worlds in May 2012 at Copenhagen Business School. More information here: http://nordicworlds.net/2012/04/13/innovation-and-market-creation-in-and-around-virtual-worlds-2/.
See the WEBCAST as well!! mms://wmedia.it.su.se/SUB/NordLib/3.wmv
Presentation at Nordlib 2.0 in Stockholm, November 21th 2008
http://www.nordlib20.org/programme/
The modern library web environment consists of multiple content sources and applications that perform essential functions that often overlap and could potentially create a fractured user experience. For example, content in a library’s Drupal website may be replicated in LibGuides or WordPress blogs. Search functionality in a discovery platform may be replicated in a federated search tool or the ILS OPAC. This presentation provides tips, tackles technical and political challenges to building a single web experience for users, discusses solutions and use of APIs (application programming interfaces), provides concrete examples, and more.
LiMoSINe Press kit introduces this project that integrates the studies of leading researchers over diverse topics with a view to enable new kinds of language-based technology search. Now we are developing 5 demonstrators: ORMA, ThemeStreams, FlickrDemo, DEESSE and Streamwatchr. http://limosine-project.eu/
Come to the Fiesta! Join the OLE ProjectDoreen Herold
Led by Duke University, the OLE Project intends to build a design document for an open source library management system which will be based on the software design philosophy of service oriented architecture (SOA). SOA is becoming a dominant trend in technology as early adopters have shown that it provides the benefit of an agile system, one that is flexible in response to information demands. Lehigh’s Doreen Herold and Tim McGeary will present the status of the OLE Project, its process, its goals, and how other PALINET members can participate.
My keynote at the Ontologies Come of Age workshop at the International Semantic Web Conference in Bonn Germany. This workshop was named after a paper I wrote about a decade ago.
Keynote presentation for the International Semantic Web Conference in Athens Greece, on November 9, 2023. The talk addresses the generative AI explosion and its potential impacts on the Semantic Web and Knowledge Graph communities and, in fact, may spark a research Renaissance.
Abstract:
We are living in an age of rapidly advancing technology. History may view this period as one in which generative artificial intelligence is seen as reshaping the landscape and narrative of many technology-based fields of research and application. Times of disruptions often present both opportunities and challenges. We will discuss some areas that may be ripe for consideration in the field of Semantic Web research and semantically-enabled applications. Semantic Web research has historically focused on representation and reasoning and enabling interoperability of data and vocabularies. At the core are ontologies along with ontology-enabled (or ontology-compatible) knowledge stores such as knowledge graphs. Ontologies are often manually constructed using a process that (1) identifies existing best practice ontologies (and vocabularies) and (2) generates a plan for how to leverage these ontologies by aligning and augmenting them as needed to address requirements. While semi-automated techniques may help, there is typically a significant portion of the work that is often best done by humans with domain and ontology expertise. This is an opportune time to rethink how the field generates, evolves, maintains, and evaluates ontologies. We consider how hybrid approaches, i.e., those that leverage generative AI components along with more traditional knowledge representation and reasoning approaches to create improved processes. The effort to build a robust ontology that meets a use case can be large. Ontologies are not static however and they need to evolve along with knowledge evolution and expanded usage. There is potential for hybrid approaches to help identify gaps in ontologies and/or refine content. Further, ontologies need to be documented with term definitions and their provenance. Opportunities exist to consider semi-automated techniques for some types of documentation, provenance, and decision rationale capture for annotating ontologies. The area of human-AI collaboration for population and verification presents a wide range of areas of research collaboration and impact. Ontologies need to be populated with class and relationship content. Knowledge graphs and other knowledge stores need to be populated with instance data in order to be used for question answering and reasoning. Population of large knowledge graphs can be time consuming. Generative AI holds the promise to create candidate knowledge graphs that are compatible with the ontology schema. The knowledge graph should contain provenance information identifying how the content was populated and its source and correctness and currency should be checked. A human-AI assistant approach is presented.
NYAI #27: Cognitive Architecture & Natural Language Processing w/ Dr. Catheri...Maryam Farooq
For more AI talks, visit: nyai.co
These slides are from NYAI #27: Cognitive Architecture & Natural Language Processing w/ Dr. Catherine Havasi, which took place Tues, 12/18/19 at Kirkland & Ellis NYC.
[Speaker Bio] Dr. Catherine Havasi is a technology strategist, artificial intelligence researcher, and entrepreneur. In the late 90s, she co-founded the Common Sense Computing Initiative, or ConceptNet, the first crowd-sourced project for artificial intelligence and the largest open knowledge graph for language understanding. ConceptNet has played a role in thousands of AI projects and will be turning 20 next year. She has started several companies commercializing AI research, including Luminoso where she acts as Chief Strategy Officer. She is currently a visiting scientist at the MIT Media Lab where she works on computational creativity and previously directed the Digital Intuition group.
[Abstract] People who build everything from entertainment experiences to financial management face a dilemma: how can you scale what you’re building for broader consumption, yet maintain the personalization that makes it special? A fundamental tension exists between building something individualized, and scaling it to consumers such as visitors at a theme park, or gamers exploring the latest Zelda adventure. True disruption happens when we overcome the idea that one must sacrifice personalization to achieve mass production — like it has in advertising, recommendations, and web search.
Artificial Intelligence practitioners, especially in natural language understanding, dialogue, and cognitive modeling, face the same issue: how can we personalize our models for all audiences without relying on unscalable efforts such as writing specific rules, building dialogue trees, or designing knowledge graphs? Catherine Havasi believes we can remove this dichotomy and achieve “mass personalization.” In this session we’ll discuss how to understand domain text and build believable digital characters. We’ll talk about how adding a little common sense, cognitive architectures, and planning is making this all possible.
nyai.co
Similar to Digital Humanities research issues (20)
Aux origines de la sociologie : du XIXème au début XX ème siècleAmar LAKEL, PhD
Ce cours d'introduction à la sociologie est conçu pour les étudiants de première année en licence de sociologie. Il vise à fournir une compréhension fondamentale des concepts clés, des théories et des méthodes de la sociologie, en mettant l'accent sur les contributions des fondateurs de la discipline à la fin du 19e et au début du 20e siècle. À travers des lectures, des discussions en classe et des analyses de cas, les étudiants exploreront comment la sociologie permet de comprendre les structures sociales, les interactions humaines et les changements sociaux.
Les théories de la propagande étudient et perfectionnent l'utilisation des mass medias depuis plus d'un siècle. Influence, leader d'opinion, opinion publique, élite... comment l'usage des mass medias permet aux élites d'une société d'agir sur les représentations du monde de la masse pour ordonner la société en amont.
[Update 2022] Elaborer la stratégie de communication - tout ce qu'il faut savoirAmar LAKEL, PhD
Depuis plus de 15 ans, je conseille et forme les responsables et futurs responsables de communication à la stratégie de communication. J'ai voulu dans cette présentation, longue et lourde mais néanmoins synthétique, résumer 15 ans d'expériences en stratégie de communication, campagne de communication, stratégie de marque et mediaplanning
Enjeux de la recherche en SHS à l'ère des digital studiesAmar LAKEL, PhD
C'est une bien étrange conférence que j'ai donnée là, durant la Master Class DNHD 2020 de Bordeaux. En février 2020, alors même que nous étions à la porte d'une crise pandémique mondiale qui allait faire entrer notre monde au XXIème, c'est en toute innocence et avec un sentiment étrange que j'improvisais ce qui aura pour but, encore inconscient, d'exprimer ma position pour une révolution de la recherche en SHS en France.
Dans l'ombre d'une énième réforme universitaire qui achèvera l'oeuvre de destruction de la recherche en sciences sociales et humaines (mais aussi à l'ombre d'une révolution interne dans ma pratique de chercheur depuis 4 ans), j'ai voulu accoucher de mon manifeste pour une recherche ouverte et libre que j'appellerai plus tard : les Z.A.R. - Les Zones Autonomes de Recherche. Entre Bureaucratie centralisatrice au service de la radicalisation idéologique du néolibéralisme et retrait dépressif de chercheurs broyéd par la machinerie technocratique, j'ai fait un rêve. L'émergence d'une N.A.R., c'est-à-dire une Non Agence de la Recherche. Rassemblement rhizotopique de chercheurs libres voulant travailler ensemble, conduit par une seule pulsion : le désir de savoir.
Nouveaux régimes de vérité à l'heure des réseaux sociauxAmar LAKEL, PhD
L'Ecole Supérieure d'Art et de Design du TALM m'a invité à faire retour sur une décennie de recherche sur la construction de l'espace public à l'heure d'Internet. La conférence a eu lieu le mercredi 6 janvier 2021 à 18 h 00, en visioconférence via Microsoft Teams.
Une très grande majorité de la classe intellectuelle (journalistes, universitaires, artistes, etc.) semble convenir d’un postulat consensuel : l’arrivée d’Internet et des réseaux sociaux ont bouleversé la construction des espaces de socialisation. Dominé par ce que Louis Althusser appelait (dans la suite de Gramsci) des Appareils Idéologiques (d’État ou d’Entreprise), l’espace public à l’ère du numérique a vu sans aucun doute un nombre incalculable de nouveaux entrants sur la scène médiatique. Mais peut-on parler d’une révolution des conditions de possibilité de la lutte idéologique à l’ère de l’internet ? Quels sont ces fameux nouveaux acteurs ? Quels rôles jouent-ils dans la construction de nos horizons communs ? Nous reviendrons dans une première partie sur la théorie de la propagande, pour mieux contextualiser une seconde partie critique consacrée aux concepts de gouvernance, de pouvoir et d’influence en communication. Une dernière partie sera consacrée aux résultats récents de ses recherches empiriques sur le sujet.
Amar LAKEL est Maître de Conférence en Sciences de l’Information et de la Communication à l’Université Bordeaux Montaigne. Depuis plus de 15 ans, il conduit ses recherches sur la construction de l’espace public à l’heure d’internet. Après avoir étudié les modes de gouvernance du web par les institutions publiques et privées, il dirige un programme de recherche, My Web intelligence, qui tente de développer des outils et une méthodologie solide pour explorer la fabrique de l’opinion sur Internet.
Ce cours de L1 est une introduction à la transformation numérique des sociétés.
Comprendre en quoi la révolution informatique est une
rupture civilisationnelle qui nous a fait basculer dans un
autre monde.
Comprendre les structures technologiques du numérique qui
guident l’émergence des innovations (avec un focus sur
l’IA).
Comprendre comment on transforme les organisations par
une stratégie de transition numérique progressive mais
certaine.
Essayer de faire une prospective des 50 prochaines années
(vos années!) sur la base de cette révolution.
Ce TD est celui du cours "L1 COURS Numérique et Société".
Il vise, par une introduction à la manipulation des données, la compréhension de l'open data dans la perspectives d'une data analyse. On y apprend à utiliser Un Tableau et un outil de visualisation des données.
Approches méthodologique d’analyse du web : en finir avec la division quali ...Amar LAKEL, PhD
13 novembre 2019 de 17 heures à 19 heures – Séance 1 (2 heures): Approches méthodologique d’analyse du web : en finir avec la division quali et quanti.
IUT Bordeaux Montaigne, place Renaudel. Amphi 2
Introduction au concept d’opérationnalisation et de reproductibilité d’une étude. Nous introduisons l’environnement de travail numérique essentiel pour assurer une recherche de qualité.
Pour en savoir plus sur le cycle de séminaires : http://mywebintelligence.net/introduction-aux-digital-methods/
Juan Branco, une stratégie marketing d'un nouvel intellectuel numériqueAmar LAKEL, PhD
Comment un auteur de 28 ans, quasi inconnu, publiant dans une maison d’édition confidentielle, sans aucun soutien de la presse qui « fait » le livre en France, devient en moins d’un mois Auteur du pamphlet politique contre le Président de la République le plus vendu en France.
Pour une documentologie digitale : La vie numérique de Robert Escarpit de 200...Amar LAKEL, PhD
Conférence donnée en 2017 sur la question épistémologique de la nature du document numérique. En quoi ce document numérique nous révèle sur les nouvelles performences communicationnelles et plus loin sur la fabrique des espaces publics numérics.
Amar Lakel - Patrimonialisation numérique et avenir digital du document.Amar LAKEL, PhD
Les humanités numériques sont la rencontre entre la numérisation des traces de l’activité humaine et la démarche scientifique des SHS. Si le digital turn est une révolution sociétale, la pratique du chercheur ne peut échapper aux enjeux majeurs d’une innovation radicale de ses pratiques. Dans le cadre de cette master class, Amar LAKEL s’est interrogé sur la capacité d’innovation du chercheur en SHS qu’il distingue de ses capacités de professionnel de la recherche. Il enjoint alors la recherche à réussir une série d’épreuves pour réussir son entrer dans le XXIe siècle.
Vous trouverez la présentation video ici : https://youtu.be/YHj953Yb95w
Et l’article sur lequel se fonde cette présentation dans la revue Etudes digitales N°6 : https://www.researchgate.net/publication/332671854_Recherches_digitales_et_production_des_donnees_bouleversement_des_agencements_pour_le_chercheur_en_SIC
Du néolithique à la smart city : généalogie de l'espace public au coeur des p...Amar LAKEL, PhD
Comprendre la ville comme un dispositif de socialisation et de subjectivation de l'homme contemporain au moment même où nous entrant dans une ère de virtualisation du social.
La méthode Foucault : De l’énoncé à l’archive, enjeux du pouvoir dans la comm...Amar LAKEL, PhD
Cet intervention dans le cadre du séminaire de mon équipe de recherche E3D vise à réfléchir sur l'héritage épistémologique et méthodologique de l'oeuvre de Michel Foucault et de son utilisation dans les problèmatiques des Sciences de l'Information et de la Communiaction (SIC)
Transition digitale et Espaces publics numéricsAmar LAKEL, PhD
Comprendre comment la transition digitale t l'innovation numérique change les conditions de vie dans les espaces publics de demain : de la smart city à la "start-up nation", quels enjeux
Global Situational Awareness of A.I. and where its headedvikram sood
You can see the future first in San Francisco.
Over the past year, the talk of the town has shifted from $10 billion compute clusters to $100 billion clusters to trillion-dollar clusters. Every six months another zero is added to the boardroom plans. Behind the scenes, there’s a fierce scramble to secure every power contract still available for the rest of the decade, every voltage transformer that can possibly be procured. American big business is gearing up to pour trillions of dollars into a long-unseen mobilization of American industrial might. By the end of the decade, American electricity production will have grown tens of percent; from the shale fields of Pennsylvania to the solar farms of Nevada, hundreds of millions of GPUs will hum.
The AGI race has begun. We are building machines that can think and reason. By 2025/26, these machines will outpace college graduates. By the end of the decade, they will be smarter than you or I; we will have superintelligence, in the true sense of the word. Along the way, national security forces not seen in half a century will be un-leashed, and before long, The Project will be on. If we’re lucky, we’ll be in an all-out race with the CCP; if we’re unlucky, an all-out war.
Everyone is now talking about AI, but few have the faintest glimmer of what is about to hit them. Nvidia analysts still think 2024 might be close to the peak. Mainstream pundits are stuck on the wilful blindness of “it’s just predicting the next word”. They see only hype and business-as-usual; at most they entertain another internet-scale technological change.
Before long, the world will wake up. But right now, there are perhaps a few hundred people, most of them in San Francisco and the AI labs, that have situational awareness. Through whatever peculiar forces of fate, I have found myself amongst them. A few years ago, these people were derided as crazy—but they trusted the trendlines, which allowed them to correctly predict the AI advances of the past few years. Whether these people are also right about the next few years remains to be seen. But these are very smart people—the smartest people I have ever met—and they are the ones building this technology. Perhaps they will be an odd footnote in history, or perhaps they will go down in history like Szilard and Oppenheimer and Teller. If they are seeing the future even close to correctly, we are in for a wild ride.
Let me tell you what we see.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data LakeWalaa Eldin Moustafa
Dynamic policy enforcement is becoming an increasingly important topic in today’s world where data privacy and compliance is a top priority for companies, individuals, and regulators alike. In these slides, we discuss how LinkedIn implements a powerful dynamic policy enforcement engine, called ViewShift, and integrates it within its data lake. We show the query engine architecture and how catalog implementations can automatically route table resolutions to compliance-enforcing SQL views. Such views have a set of very interesting properties: (1) They are auto-generated from declarative data annotations. (2) They respect user-level consent and preferences (3) They are context-aware, encoding a different set of transformations for different use cases (4) They are portable; while the SQL logic is only implemented in one SQL dialect, it is accessible in all engines.
#SQL #Views #Privacy #Compliance #DataLake
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
Adjusting OpenMP PageRank : SHORT REPORT / NOTESSubhajit Sahu
For massive graphs that fit in RAM, but not in GPU memory, it is possible to take
advantage of a shared memory system with multiple CPUs, each with multiple cores, to
accelerate pagerank computation. If the NUMA architecture of the system is properly taken
into account with good vertex partitioning, the speedup can be significant. To take steps in
this direction, experiments are conducted to implement pagerank in OpenMP using two
different approaches, uniform and hybrid. The uniform approach runs all primitives required
for pagerank in OpenMP mode (with multiple threads). On the other hand, the hybrid
approach runs certain primitives in sequential mode (i.e., sumAt, multiply).
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdfGetInData
Recently we have observed the rise of open-source Large Language Models (LLMs) that are community-driven or developed by the AI market leaders, such as Meta (Llama3), Databricks (DBRX) and Snowflake (Arctic). On the other hand, there is a growth in interest in specialized, carefully fine-tuned yet relatively small models that can efficiently assist programmers in day-to-day tasks. Finally, Retrieval-Augmented Generation (RAG) architectures have gained a lot of traction as the preferred approach for LLMs context and prompt augmentation for building conversational SQL data copilots, code copilots and chatbots.
In this presentation, we will show how we built upon these three concepts a robust Data Copilot that can help to democratize access to company data assets and boost performance of everyone working with data platforms.
Why do we need yet another (open-source ) Copilot?
How can we build one?
Architecture and evaluation
2. AN OPEN SOURCE PLATFORM AS
FOUNDATION FOR DH TOOLS
An open source
platform to bind them
all
Digital Humanities are a major challenge to
social sciences: adding ITs in extracting,
archiving, automated analysis, corpus
qualification, data visualization…
3. AN OPEN SOURCE PLATFORM AS
FOUNDATION FOR DH TOOLS
An open source
platform to bind them
all
Digital Humanities are a major challenge to
social sciences: adding ITs in extracting,
archiving, automated analysis, corpus
qualification, data visualization…
CREATE A UNIFICATION DYNAMICS
Too many one-shot projects, high-value innovations
without consolidating experience. One platform to
bind them all ?
4. AN OPEN SOURCE PLATFORM AS
FOUNDATION FOR DH TOOLS
An open source
platform to bind them
all
Digital Humanities are a major challenge to
social sciences: adding ITs in extracting,
archiving, automated analysis, corpus
qualification, data visualization…
CREATE A UNIFICATION DYNAMICS
Too many one-shot projects, high-value innovations
without consolidating experience. One platform to
bind them all ?
OPEN GOVERNANCE FROM DAY ONE
My Web Intelligence is built around collaborative
tools (Github, Trello, etc.). These have been public
from day one, thus publicizing all research
progresses.
5. AN OPEN SOURCE PLATFORM AS
FOUNDATION FOR DH TOOLS
An open source
platform to bind them
all
Digital Humanities are a major challenge to
social sciences: adding ITs in extracting,
archiving, automated analysis, corpus
qualification, data visualization…
CREATE A UNIFICATION DYNAMICS
Too many one-shot projects, high-value innovations
without consolidating experience. One platform to
bind them all ?
OPEN GOVERNANCE FROM DAY ONE
My Web Intelligence is built around collaborative
tools (Github, Trello, etc.). These have been public
from day one, thus publicizing all research
progresses.
FOR THE COMMON GOOD
My Web Intelligence is meant to be as common as
possible so that intelligence tools benefit to all
(easy to install, well documented, etc.)
6. AN OPEN SOURCE PLATFORM AS
FOUNDATION FOR DH TOOLS
An open source
platform to bind them
all
Digital Humanities are a major challenge to
social sciences: adding ITs in extracting,
archiving, automated analysis, corpus
qualification, data visualization…
CREATE A UNIFICATION DYNAMICS
Too many one-shot projects, high-value innovations
without consolidating experience. One platform to
bind them all ?
OPEN GOVERNANCE FROM DAY ONE
My Web Intelligence is built around collaborative
tools (Github, Trello, etc.). These have been public
from day one, thus publicizing all research
progresses.
FOR THE COMMON GOOD
My Web Intelligence is meant to be as common as
possible so that intelligence tools benefit to all (easy
to install, well documented, etc.)
COLLABORATIVE FIRST
My Web Intelligence chooses openness and
collaboration to answer the challenges posed by
new technologies and media.
7. The content manager :
the heterogeneous
archive management
challenge
Allow SHS study the digital humanities is
offer primarily a platform able to extract and
retain huge amounts of expressions from
heterogeneous sources.
MASTERING THE EXTRACTION AND
ARCHIVING AGENTS (CRAWLERS) AMID
BIG DATA.
8. The content manager :
the heterogeneous
archive management
challenge
Allow SHS study the digital humanities is
offer primarily a platform able to extract and
retain huge amounts of expressions from
heterogeneous sources.
AUTOMATICALLY EXTRACTING CORPUS ON
THE NEEDS
Give a crawler accessing heterogeneous sources
with enough modularity to meet all users' projects
MASTERING THE EXTRACTION AND
ARCHIVING AGENTS (CRAWLERS) AMID
BIG DATA.
9. The content manager :
the heterogeneous
archive management
challenge
Allow SHS study the digital humanities is
offer primarily a platform able to extract and
retain huge amounts of expressions from
heterogeneous sources.
AUTOMATICALLY EXTRACTING CORPUS ON
THE NEEDS
Give a crawler accessing heterogeneous sources with
enough modularity to meet all users' projects
GIVE A USER INTERFACE TO MANAGE
CORPUS.
Clean, delete, sort, rearrange, according to its own
heuristics, is a must for any DH project.
MASTERING THE EXTRACTION AND
ARCHIVING AGENTS (CRAWLERS) AMID
BIG DATA.
10. The content manager :
the heterogeneous
archive management
challenge
Allow SHS study the digital humanities is
offer primarily a platform able to extract and
retain huge amounts of expressions from
heterogeneous sources.
AUTOMATICALLY EXTRACTING CORPUS ON
THE NEEDS
Give a crawler accessing heterogeneous sources with
enough modularity to meet all users' projects
GIVE A USER INTERFACE TO MANAGE
CORPUS.
Clean, delete, sort, rearrange, according to its own
heuristics, is a must for any DH project.
A COLLABORATIVE MANAGEMENT TOOLS
FOR DATA STUDIES.
We did not win the HD challenge alone. A platform
of this ambition will integrate a team management
module to the data processing service .
MASTERING THE EXTRACTION AND
ARCHIVING AGENTS (CRAWLERS) AMID
BIG DATA.
11. The content manager :
the heterogeneous
archive management
challenge
Allow SHS study the digital humanities is
offer primarily a platform able to extract and
retain huge amounts of expressions from
heterogeneous sources.
AUTOMATICALLY EXTRACTING CORPUS ON
THE NEEDS
Give a crawler accessing heterogeneous sources with
enough modularity to meet all users' projects
GIVE A USER INTERFACE TO MANAGE
CORPUS.
Clean, delete, sort, rearrange, according to its own
heuristics, is a must for any DH project.
A COLLABORATIVE MANAGEMENT TOOLS FOR
DATA STUDIES.
We did not win the HD challenge alone. A platform of
this ambition will integrate a team management
module to the data processing service .
RECRUITING INTELLIGENT AGENTS
The democratization of machine learning and
artificial intelligence now allows hiring processing
algorithms to assist you in the mass management of
your data.
MASTERING THE EXTRACTION AND
ARCHIVING AGENTS (CRAWLERS) AMID
BIG DATA.
12. Analysis of content :
The challenge of
automating the
qualification.
The language processing has made
enormous progress. However some open
solution provides opportunities to qualify
the body of masses. Our project aims to
bring together the foundations of research
in this area.
QUALIFY AUTOMATICALLY DATA ABOUT
COMMUNICATION SITUATIONS .
13. Analysis of content :
The challenge of
automating the
qualification.
The language processing has made
enormous progress. However some open
solution provides opportunities to qualify
the body of masses. Our project aims to
bring together the foundations of research
in this area.
QUALIFY THE COMMUNICATION SITUATION .
Each expression have to be contextualized in a
mediated communication situation and need to be
qualified automatically.
QUALIFY AUTOMATICALLY DATA ABOUT
COMMUNICATION SITUATIONS .
14. Analysis of content :
The challenge of
automating the
qualification.
The language processing has made
enormous progress. However some open
solution provides opportunities to qualify
the body of masses. Our project aims to
bring together the foundations of research
in this area.
QUALIFY THE COMMUNICATION SITUATION .
Each expression have to be contextualized in a
mediated communication situation and need to be
qualified automatically.
ANALYZE THE IMPACT OF ACTS DISCURSIVE .
Save impact indicators of all expressions to be able
to not only measure their influence but their
resonance with the representations of the receivers
of the message.
QUALIFY AUTOMATICALLY DATA ABOUT
COMMUNICATION SITUATIONS .
15. Analysis of content :
The challenge of
automating the
qualification.
The language processing has made
enormous progress. However some open
solution provides opportunities to qualify
the body of masses. Our project aims to
bring together the foundations of research
in this area.
QUALIFY THE COMMUNICATION SITUATION .
Each expression have to be contextualized in a
mediated communication situation and need to be
qualified automatically.
ANALYZE THE IMPACT OF ACTS DISCURSIVE .
Save impact indicators of all expressions to be able to
not only measure their influence but their resonance
with the representations of the receivers of the
message.
ANALYZE THE CONTENT AUTOMATICALLY.
Lemmatization of texts, the main objects of
expressions, arguments trees ... The content analysis
allows automatic classification of the corpus serving
the detection of collective representations.
QUALIFY AUTOMATICALLY DATA ABOUT
COMMUNICATION SITUATIONS .
16. Analysis of content :
The challenge of
automating the
qualification.
The language processing has made
enormous progress. However some open
solution provides opportunities to qualify
the body of masses. Our project aims to
bring together the foundations of research
in this area.
QUALIFY THE COMMUNICATION SITUATION .
Each expression have to be contextualized in a
mediated communication situation and need to be
qualified automatically.
ANALYZE STYLISTIC FORMS TO IDENTIFY
PATTERNS SPEAKER .
The style , feeling, language level , type of
vocabulary ... the detection of styles enriches
patterns of speakers to better identify their
intention of communication.
ANALYZE THE IMPACT OF ACTS DISCURSIVE .
Save impact indicators of all expressions to be able to
not only measure their influence but their resonance
with the representations of the receivers of the
message.
ANALYZE THE CONTENT AUTOMATICALLY.
Lemmatization of texts, the main objects of
expressions, arguments trees ... The content analysis
allows automatic classification of the corpus serving
the detection of collective representations.
QUALIFY AUTOMATICALLY DATA ABOUT
COMMUNICATION SITUATIONS .
17. The algorithms of
speech: At the source
of the positions.
The generation of discourse responds to
more or less stereotyped behaviors. The
algorithms that detect patterns used to
measure but also to predict them...
DETECTION AND MEASURING PATTERNS
FROM SOURCE OF SPEECH TO
UNDERSTAND THE ECONOMY
GENERATIVE.
18. The algorithms of
speech: At the source
of the positions.
The generation of discourse responds to
more or less stereotyped behaviors. The
algorithms that detect patterns used to
measure but also to predict them...
ANALYZE THE POSITIONS OF SPEAKERS .
By the qualification of expressions depending on
the discursive act model, it is possible to quantify
the production of discourse through multi varied
statistical processing (type AFC , ACP, trees ... )
DETECTION AND MEASURING PATTERNS
FROM SOURCE OF SPEECH TO
UNDERSTAND THE ECONOMY
GENERATIVE.
19. The algorithms of
speech: At the source
of the positions.
The generation of discourse responds to
more or less stereotyped behaviors. The
algorithms that detect patterns used to
measure but also to predict them...
ANALYZE THE POSITIONS OF SPEAKERS .
By the qualification of expressions depending on the
discursive act model, it is possible to quantify the
production of discourse through multi varied
statistical processing (type AFC , ACP, trees ... )
PREDICTING PRODUCTION OF SPEECH
Predictive algorithms enable not only qualify
incomplete data but also to generate hypotheses
about future positions taken by developing future
scenarios
DETECTION AND MEASURING PATTERNS
FROM SOURCE OF SPEECH TO
UNDERSTAND THE ECONOMY
GENERATIVE.
20. The algorithms of
speech: At the source
of the positions.
The generation of discourse responds to
more or less stereotyped behaviors. The
algorithms that detect patterns used to
measure but also to predict them...
ANALYZE THE POSITIONS OF SPEAKERS .
By the qualification of expressions depending on the
discursive act model, it is possible to quantify the
production of discourse through multi varied
statistical processing (type AFC , ACP, trees ... )
PREDICTING PRODUCTION OF SPEECH
Predictive algorithms enable not only qualify
incomplete data but also to generate hypotheses
about future positions taken by developing future
scenarios
THE SOCIAL NETWORK ANALYSIS AS SOCIAL
CONTEXT OF SPEECH
The structural analysis of networks applied to the
analysis of discourse in their co-citation retrieves
the frame that binds and socializes enunciators.
DETECTION AND MEASURING PATTERNS
FROM SOURCE OF SPEECH TO
UNDERSTAND THE ECONOMY
GENERATIVE.
21. The algorithms of
speech: At the source
of the positions.
The generation of discourse responds to
more or less stereotyped behaviors. The
algorithms that detect patterns used to
measure but also to predict them...
ANALYZE THE POSITIONS OF SPEAKERS .
By the qualification of expressions depending on the
discursive act model, it is possible to quantify the
production of discourse through multi varied
statistical processing (type AFC , ACP, trees ... )
SNA AS THE ANALYSIS OF COGNITIVE
STRUCTURES OF SPEECH.
The SNA provides a new perspective in the analysis
of the argumentative co-presence in the large
corpus by introducing its own notions (centrality,
betwenness, etc.).
PREDICTING PRODUCTION OF SPEECH
Predictive algorithms enable not only qualify
incomplete data but also to generate hypotheses
about future positions taken by developing future
scenarios
THE SOCIAL NETWORK ANALYSIS AS SOCIAL
CONTEXT OF SPEECH
The structural analysis of networks applied to the
analysis of discourse in their co-citation retrieves the
frame that binds and socializes enunciators.
DETECTION AND MEASURING PATTERNS
FROM SOURCE OF SPEECH TO
UNDERSTAND THE ECONOMY
GENERATIVE.
22. Data visualization: The
look as a source of
intelligence ?
The data visualization challenge is to give
interpretive schemes for large masses of
data in a specific study context. My Web
Intelligence explore the relationship
between visualization and digital expression.
VIEW AND INTERPRETING DIGITAL
EXPRESSIONS WEB.
23. Data visualization: The
look as a source of
intelligence ?
The data visualization challenge is to give
interpretive schemes for large masses of
data in a specific study context. My Web
Intelligence explore the relationship
between visualization and digital expression.
NAVIGATING THE CORPUS OF EXPRESSION.
View and Navigate the expressions through
dashboards of act of enunciation (type, media,
speakers, hearing, etc.).
VIEW AND INTERPRETING DIGITAL
EXPRESSIONS WEB.
24. Data visualization: The
look as a source of
intelligence ?
The data visualization challenge is to give
interpretive schemes for large masses of
data in a specific study context. My Web
Intelligence explore the relationship
between visualization and digital expression.
NAVIGATING THE CORPUS OF EXPRESSION.
View and Navigate the expressions through
dashboards of act of enunciation (type, media,
speakers, hearing, etc.).
SORT AND INDEXING CONTENT.
Explorer viewing by keyword clouds, dynamic
indexes and other representations of the text to
facilitate the conceptual analysis.
VIEW AND INTERPRETING DIGITAL
EXPRESSIONS WEB.
25. Data visualization: The
look as a source of
intelligence ?
The data visualization challenge is to give
interpretive schemes for large masses of
data in a specific study context. My Web
Intelligence explore the relationship
between visualization and digital expression.
NAVIGATING THE CORPUS OF EXPRESSION.
View and Navigate the expressions through
dashboards of act of enunciation (type, media,
speakers, hearing, etc.).
SORT AND INDEXING CONTENT.
Explorer viewing by keyword clouds, dynamic indexes
and other representations of the text to facilitate the
conceptual analysis.
MAPPING THE SOURCES OF INFORMATION.
The mapping of the speakers enables contextual
navigation supports media by analyzing their
relevant relationships as social context of utterance.
VIEW AND INTERPRETING DIGITAL
EXPRESSIONS WEB.
26. Data visualization: The
look as a source of
intelligence ?
The data visualization challenge is to give
interpretive schemes for large masses of
data in a specific study context. My Web
Intelligence explore the relationship
between visualization and digital expression.
NAVIGATING THE CORPUS OF EXPRESSION.
View and Navigate the expressions through
dashboards of act of enunciation (type, media,
speakers, hearing, etc.).
MAPPING COLLECTIVE REPRESENTATIONS.
The use of SNA in concept mapping offers the
prospect of a new visualization of collective
representations and therefore the context of
knowledge and episteme studied sayings.
SORT AND INDEXING CONTENT.
Explorer viewing by keyword clouds, dynamic indexes
and other representations of the text to facilitate the
conceptual analysis.
MAPPING THE SOURCES OF INFORMATION.
The mapping of the speakers enables contextual
navigation supports media by analyzing their relevant
relationships as social context of utterance.
VIEW AND WEB. INTERPRETING DIGITAL
EXPRESSIONS
27. MY WEB INTELLIGENCE
Architecture Patterns Issues
PROJECT MANAGER
(territories and requests)
ORACLES
first list of approved
expression for starting the
graph`
READER
Donwload and index the
document like an
expression
CRAWLER
Deep crawling web
SCRAPPER
Read heteregenous files
for build an expression
APPROBATION
Algorithm for approving
linked expression
QUALIFICATION
Enrich expression and
domain with data
RANKING
Build Kpi to rank
expression and domain
APIs
Making a bridge with 3rd
Soft
EXPORT FILES
CSV, GEXPH and all kind
of models
VIZUALISATION
Use Viz librairies to
navigate in this data
(graph, tree, etc.)
Input : absorb the corpus Annotate : qualify your datas Output : show the patterns
The My Web Intelligence challenge is to
absorb heterogeneous corpus and archive
and index them in a model of Author -
Media- expression data.
This is both to respect the specificity of
media but at the same time to use the
communication meta analysis models to
analyze both, meaning of speech, but also
the pragmatic analyzes of communicative
acts
It's not about only what is said in a kind of
naive sociology but also to understand the
social dynamics and strategies at work in
the production of meaning .
The second part of the development of My
Web Intelligence is to develop intelligent
agents able the playing the role of
librarians in their functions :
- Analysis of the relevance of the
document in respect of the project. One of
the major issues in the management of
datas is the cleaning of noise and digital
debris. But beyond that, the relevance to
the research issues are the key usability
HD platforms.
- Data Enrichment (or annotation). The
second role of "librarian function" of each
research project is to work , annotate, as
described by algorithms that human
agents , both by external and internal
sources.
My Web Intelligence thinks like a Core
Framework in the ecosystem of data
analysis projects. A major issue is the
interconnection of the process to third-
party applications in both upstream
inputs/outputs .
Designing an API outputs/inputs to make
compatible third management solutions
and data processing.
Facilitate the production and export of BD
-compatible files with the processing in
third-party software (ex . Gephi , SPSS , R,
etc. )
Use data visualization solutions to navigate
, process and analyze large corpus and
identify meaningful patterns.