DiscoverKnowledge Graph Insights
Knowledge Graph Insights
Claim Ownership

Knowledge Graph Insights

Author: Larry Swanson

Subscribed: 13Played: 176
Share

Description

Interviews with experts on semantic technology, ontology design and engineering, linked data, and the semantic web.
37 Episodes
Reverse
Casey Hart Ontology engineering has its roots in the idea of ontology as defined by classical philosophers. Casey Hart sees many other connections between professional ontology practice and the academic discipline of philosophy and shows how concepts like epistemology, metaphysics, and rhetoric are relevant to both knowledge graphs and AI technology in general. We talked about: his work as a lead ontologist at Ford and as an ontology consultant his academic background in philosophy the variety of pathways into ontology practice the philosophical principles like metaphysics, epistemology, and logic that inform the practice of ontology his history with the the Cyc project and employment at Cycorp how he re-uses classes like "category" and similar concepts from upper ontologies like gist his definition of "AI" - including his assertion that we should use term to talk about a practice, not a particular technology his reminder that ontologies are models and like all models can oversimplify reality Casey's bio Casey Hart is the lead ontologist for Ford, runs an ontology consultancy, and pilots a growing YouTube channel. He is enthusiastic about philosophy and ontology evangelism. After earning his PhD in philosophy from the University of Wisconsin-Madison (specializing in epistemology and the philosophy of science), he found himself in the private sector at Cycorp. Along his professional career, he has worked in several domains: healthcare, oil & gas, automotive, climate science, agriculture, and retail, among others. Casey believes strongly that ontology should be fun, accessible, resemble what is being modelled, and just as complex as it needs to be. He lives in the Pacific Northwest with his wife and three daughters and a few farm animals. Connect with Casey online LinkedIn ontologyexplained at gmail dot com Ontology Explained YouTube channel Video Here’s the video version of our conversation: https://youtu.be/siqwNncPPBw Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 38. When the subject of philosophy comes up in relation to ontology practice, it's typically cited as the origin of the term, and then the subject is dropped. Casey Hart sees many other connections between ontology practice and it its philosophical roots. In addition to logic as the foundation of OWL, he shows how philosophy concepts like epistemology, metaphysics, and rhetoric are relevant to both knowledge graphs and AI technology in general. Interview transcript Larry: Hi, everyone. Welcome to episode number 38 of the Knowledge Graph Insights podcast. I am really delighted today to welcome to the show Casey Hart. Casey has a really cool YouTube channel on the philosophy behind ontology engineering and ontology practice. Casey is currently an ontologist at Ford, the motor car company. So welcome Casey, tell the folks a little bit more about what you're up to these days. Casey: Hi. Thanks, Larry. I'm super excited to be here. I've listened to the podcast, and man, your intro sounds so smooth. I was like, "I wonder how many edits that takes." No, you just fire them off, that's beautiful. Casey: Yeah, so like you said, these days I'm the ontologist at Ford, so building out data models for sensor data and vehicle information, all those sorts of fun things. I am also working as a consultant. I've got a couple of different startup healthcare companies and some cybersecurity stuff, little things around the edge. I love evangelizing ontology, talking about it and thinking about it. And as you mentioned for the YouTube channel, that's been my creative outlet. My background is in philosophy and I was interested in, I got my PhD in philosophy, I was going to teach it. You write lots of papers, those sorts of things, and I miss that to some extent getting out into industry, and that's been my way back in to, all right, come up with an idea,
Chris Mungall Capturing knowledge in the life sciences is a huge undertaking. The scope of the field extends from the atomic level up to planetary-scale ecosystems, and a wide variety of disciplines collaborate on the research. Chris Mungall and his colleagues at the Berkeley Lab tackle this knowledge-management challenge with well-honed collaborative methods and AI-augmented computational tooling that streamlines the organization of these precious scientific discoveries. We talked about: his biosciences and genetics work at the Berkeley Lab how the complexity and the volume of biological data he works with led to his use of knowledge graphs his early background in AI his contributions to the gene ontology the unique role of bio-curators, non-semantic-tech biologists, in the biological ontology community the diverse range of collaborators involved in building knowledge graphs in the life sciences the variety of collaborative working styles that groups of bio-creators and ontologists have created some key lessons learned in his long history of working on large-scale, collaborative ontologies, key among them, meeting people where they are some of the facilitation methods used in his work, tools like GitHub, for example his group's decision early on to commit to version tracking, making change-tracking an entity in their technical infrastructure how he surfaces and manages the tacit assumptions that diverse collaborators bring to ontology projects how he's using AI and agentic technology in his ontology practice how their decision to adopt versioning early on has enabled them to more easily develop benchmarks and evaluations some of the successes he's had using AI in his knowledge graph work, for example, code refactoring, provenance tracking, and repairing broken links Chris's bio Chris Mungall is Department Head of Biosystems Data Science at Lawrence Berkeley National Laboratory. His research interests center around the capture, computational integration, and dissemination of biological research data, and the development of methods for using this data to elucidate biological mechanisms underpinning the health of humans and of the planet. He is particularly interested in developing and applying knowledge-based AI methods, particularly Knowledge Graphs (KGs) as an approach for integrating and reasoning over multiple types of data. Dr. Mungall and his team have led the creation of key biological ontologies for the integration of resources covering gene function, anatomy, phenotypes and the environment. He is a principal investigator on major projects such as the Gene Ontology (GO) Consortium, the Monarch Initiative, the NCATS Biomedical Data Translator, and the National Microbiome Data Collaborative project. Connect with Chris online LinkedIn Berkeley Lab Video Here’s the video version of our conversation: https://youtu.be/HMXKFQgjo5E Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 37. The span of the life sciences extends from the atomic level up to planetary ecosystems. Combine this scale and complexity with the variety of collaborators who manage information about the field, and you end up with a huge knowledge-management challenge. Chris Mungall and his colleagues have developed collaborative methods and computational tooling that enable the construction of ontologies and knowledge graphs that capture this crucial scientific knowledge. Interview transcript Larry: Hi everyone. Welcome to episode number 37 of the Knowledge Graph Insights podcast. I am really delighted today to welcome to the show Chris Mungall. Chris is a computational scientist working in the biosciences at the Lawrence Berkeley National Laboratory. Many people just call it the Berkeley Lab. He's the principal investigator in a group there, has his own lab working on a bunch of interesting stuff, which we're going to talk about today.
Emeka Okoye Semantic technologies permit powerful connections across a variety of linked data resources across the web. Until recently, developers had to learn the RDF language to discover and use these resources. Leveraging the new Model Context Protocol (MCP) and LLM-powered natural-language interfaces, Emeka Okoye has created the RDF Explorer, an MCP service that lets any developer surf the semantic web without having to learn its specialized language. We talked about: his long history in knowledge engineering and AI agents his deep involvement in the business and technology communities in Nigeria, including founding the country's first internet startup how he was building knowledge graphs before Google coined the term an overview of MCP, the Model Context Protocol, and its benefits the RDF Explorer MCP server he has developed how the MCP protocol and helps ease some of the challenges that semantic web developers have traditionally faced the capabilities of his RDF Explorer: facilitating communication between AI applications, language models, and RDF data enabling graph exploration and graph data analysis via SPARQL queries browsing, accessing, and evaluating linked-open-data RDF resources the origins of RDF Explorer in his attempt to improve ontology engineering tooling his objections to "vibe ontology" creation the ability of RDF Explorer to let non-RDF developers users access knowledge graph data how accessing knowledge graph data addresses the problem of the static nature of the data in language models the natural connections he sees between neural network AI and symbolic AI like knowledge graphs, and the tech tribalism he sees in the broader AI world that prevents others from seeing them how the ability of LLMs to predict likely language isn't true intelligence or actual knowledge some of the lessons he learned by building the RDF Explorer, e.g., how the MCP protocol removes a lot of the complexity in building hybrid AI solutions how MCP helps him validate the ontologies he creates Emeka's bio Emeka is a Knowledge Engineer, Semantic Architect, and Generative AI Engineer who leverages his over two decades of expertise in ontology and knowledge engineering and software development to architect, develop, and deploy innovative, data-centric AI products and intelligent cognitive systems to enable organizations in their Digital Transformation journey to enhance their data infrastructure, harness their data assets for high-level cognitive tasks and decision-making processes, and drive innovation and efficiency enroute to achieving their organizational goals. Emeka’s experience has embraced a breadth of technologies his primary focus being solution design, engineering and product development while working with a cross section of professionals across various cultures in Africa and Europe in solving problems at a complex level. Emeka can understand and explain technologies from deep diving under the hood to the value proposition level. Connect with Emeka online LinkedIn Making Knowledge Graphs Accessible: My Journey with MCP and RDF Explorer RDF Explorer (GitHub) Video Here’s the video version of our conversation: https://youtu.be/GK4cqtgYRfA Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 36. The widespread adoption of semantic technologies has created a variety of linked data resources on the web. Until recently, you had to learn semantic tools to access that data. The arrival of LLMs, with their conversational interfaces and ability to translate natural language into knowledge graph queries, combined with the new Model Context Protocol, has empowered semantic web experts like Emeka Okoye to build tools that let any developer surf the semantic web. Interview transcript Larry: Hi, everyone. Welcome to episode number 36 of the Knowledge Graph Insights podcast.
Tom Plasterer Shortly after the semantic web was introduced, the demand for discoverable and shareable data arose in both research and industry. Tom Plasterer was instrumental in the early conception and creation of the FAIR data principle, the idea that data should be findable, accessible, interoperable, and reusable. From its origins in the semantic web community, scientific research, and the pharmaceutical industry, the FAIR data idea has spread across academia, research, industry, and enterprises of all kinds. We talked about: his recent move from a big pharma company to Exponential Data where he leads the knowledge graph and FAIR data practices the direct line from the original semantic web concept to FAIR data principles the scope of the FAIR acronym, not just four concepts, but actually 15 how the accessibility requirement in FAIR distinguishes the standard from the open data the role of knowledge graphs in the implementation of a FAIR data program the intentional omission of prescribed implementations in the development of FAIR and the ensuing variety of implementation patterns how the desire for consensus in the biology community smoothed the development of the FAIR standard the role of knowledge graphs in providing a structure for sharing terminology and other information in a scientific community how his interest in omics led him to computer science and then to the people skills crucial to knowledge graph work the origins of the impetus for FAIR in European scientific research and the pharmaceutical industry the growing adoption of FAIR as enterprises mature their web thinking and vendors offer products to help with implementations the roles of both open science and the accessibility needs in industry contributed to the development of FAIR the interesting new space at the intersection of generative AI and FAIR and knowledge graph the crucial foundational role of FAIR in AI systems Tom's bio Dr. Tom Plasterer is a leading expert in data strategy and bioinformatics, specializing in the application of knowledge graphs and FAIR data principles within life sciences and healthcare. With over two decades of experience in both industry and academia, he has significantly contributed to bioinformatics, systems biology, biomarker discovery, and data stewardship. His entrepreneurial ventures include co-founding PanGenX, a Personalized Medicine/Pharmacogenetics Knowledge Base start-up, and directing Project Planning and Data Interpretation at BG Medicine. During his extensive tenure at AstraZeneca, he was instrumental in championing Data Centricity, FAIR Data, and Knowledge Graph initiatives across various IT and scientific business units. Currently, Dr. Plasterer serves as the Managing Director of Knowledge Graph and FAIR Data Capability at XponentL Data, where he defines strategy and implements advanced applications of FAIR data, knowledge graphs, and generative AI for the life science and healthcare industries. He is also a prominent figure in the community, having co-founded the Pistoia Alliance FAIR Data Implementation group and serving on its FAIR data advisory board. Additionally, he co-organizes the Health Care and Life Sciences symposium at the Knowledge Graph Conference and is a member of Elsevier’s Corporate Advisory Board. Connect with Tom online LinkedIn Video Here’s the video version of our conversation: https://youtu.be/Lt9Dc0Jvr4c Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 35. With the introduction of semantic web technologies in the early 2000s, the World Wide Web began to look something like a giant database. And with great data, comes great responsibility. In response to the needs of data stewards and consumers across science, industry, and technology, the FAIR data principle - F A I R - was introduced. Tom Plasterer was instrumental in the early efforts to make web data findable,
Mara Inglezakis Owens Mara Inglezakis Owens brings a human-centered focus to her work as an enterprise architect at a major US airline. Drawing on her background in the humanities and her pragmatic approach to business, she has developed a practice that embodies both "digital anthropology" and product thinking. The result is a knowledge architecture that works for its users and consistently demonstrates its value to key stakeholders. We talked about: her role as an enterprise architect at a major US airline how her background as a humanities scholar, and especially as a rhetoric teacher, prepared her for her current work as a trusted business advisor some important mentoring she received early in her career how "digital anthropology" and product thinking fit into her enterprise architecture practice how she demonstrates the financial value of her work to executives and other stakeholders her thoughtful approach to the digitalization process and systems design the importance of documentation in knowledge engineering work how to sort out and document stakeholders' self-reports versus their actual behavior the scope of her knowledge modeling work, not just physical objects in the world, but also processes and procedures two important lessons she's learned over her career: don't be afraid to justify financial investment in your work, and "don't be so attached to an ideal outcome that you miss the best possible" Mara's bio Mara Inglezakis Owens is an enterprise architect who specializes in digitalization and knowledge management. She has deep experience in end-to-end supply chain as well as in planning, product, and program management. Mara’s background is in epistemology (history and philosophy of science, information science, and literature), which gives a unique, humanistic flavor to her practice. When she is not working, Mara enjoys aviation, creative writing, gardening, and raising her children. She lives in Minneapolis. Connect with Mara online LinkedIn email: mara dot inglezakis dot owens at gmail dot com Video Here’s the video version of our conversation: https://youtu.be/d8JUkq8bMIc Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 34. When think about architecting knowledge systems for a giant business like a global airline, you might picture huge databases and complex spaghetti diagrams of enterprise architectures. These do in fact exist, but the thing that actually makes these systems work is an understanding of the needs of the people who use, manage, and finance them. That's the important, human-focused work that Mara Inglezakis Owens does as an enterprise architect at a major US airline. Interview transcript Larry: Hi, everyone. Welcome to episode 34 of the Knowledge Graph Insights Podcast. I am really delighted today to welcome to the show, Mara, I'm going to get this right, Inglezakis Owens. She's an enterprise architect at a major US airline. So, welcome, Mara. Tell the folks a little bit more about what you're up to these days. Mara: Hi, everybody. My name's Mara. And these days I am achieving my childhood dream of working in aviation, not as a pilot, but that'll happen, but as an enterprise architect. I've been doing EA, also data and information architecture, across the whole scope of supply chain for about 10 years, everything from commodity sourcing to SaaS, software as a service, to now logistics. And a lot of my days, I spend interviewing subject matter experts, convincing business leaders they should do stuff, and on my best days, I get to crawl around on my hands and knees in an airplane hangar. Larry: Oh, fun. That is ... Yeah. I didn't know ... I knew that there's that great picture of you sitting in the jet engine, but I didn't realize this was the fulfillment of a childhood dream. That's awesome. But everything you've just said ties in so well to the tagline on your LinkedIn pro...
Frank van Harmelen Much of the conversation around AI architectures lately is about neuro-symbolic systems that combine neural-network learning tech like LLMs and symbolic AI like knowledge graphs. Frank van Harmelen's research has followed this path, but he puts all of his AI research in the larger context of how these technical systems can best support people. While some in the AI world seek to replace humans with machines, Frank focuses on AI systems that collaborate effectively with people. We talked about: his role as a professor of AI at the Vrije Universiteit in Amsterdam how rapid change in the AI world has affected the 10-year, €20-million Hybrid Intelligence Centre research he oversees the focus of his research on the hybrid combination of human and machine intelligence how the introduction of conversational interfaces has advance AI-human collaboration a few of the benefits of hybrid human-AI collaboration the importance of a shared worldview in any collaborative effort the role of the psychological concept of "theory of mind" in hybrid human-AI systems the emergence of neuro-symbolic solutions how he helps his students see the differences between systems 1 and 2 thinking and its relevance in AI systems his role in establishing the foundations of the semantic web the challenges of running a program that spans seven universities and employs dozens of faculty and PhD students some examples of use cases for hybrid AI-human systems his take on agentic AI, and the importance of humans in agent systems some classic research on multi-agent computer systems the four research challenges - collaboration, adaptation, responsibility, and explainability - they are tackling in their hybrid intelligence research his take on the different approaches to AI in Europe, the US, and China the matrix structure he uses to allocate people and resources to three key research areas: problems, solutions, and evaluation his belief that "AI is there to collaborate with people and not to replace us" Frank's bio Since 2000 Frank van Harmelen has played a leading role in the development of the Semantic Web. He is a co-designer of the Web Ontology Language OWL, which has become a worldwide standard. He co-authored the first academic textbook of the field, and was one of the architects of Sesame, an RDF storage and retrieval engine, which is in wide academic and industrial use. This work received the 10-year impact award at the International Semantic Web Conference. Linked Open Data and Knowledge Graphs are important spin-offs from this work. Since 2020, Frank is is scientific director of the Hybrid Intelligence Centre, where 50 PhD students and as many faculty members from 7 Dutch universities investigate AI systems that collaborate with people instead of replacing them. The large scale of modern knowledge graphs that contain hundreds of millions of entities and relationships (made possible partly by the work of Van Harmelen and his team) opened the door to combine these symbolic knowledge representations with machine learning. Since 2018, Frank has pivoted his research group from purely symbolic Knowledge Representation to Neuro-Symbolic forms of AI. Connect with Frank online Hybrid Intelligence Centre Video Here’s the video version of our conversation: https://youtu.be/ox20_l67R7I Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 33. As the AI landscape has evolved over the past few years, hybrid architectures that combine LLMs, knowledge graphs, and other AI technology have become the norm. Frank van Harmelen argues that the ultimate hybrid system must also include humans. He's running a 10-year, €20 million research program in the Netherlands to explore exactly this. His Hybrid Intelligence Centre investigates AI systems that collaborate with people instead of replacing them. Interview transcript Larry: Hi,
Denny Vrandečić As the founder of Wikidata, Denny Vrandečić has thought a lot about how to better connect the world's knowledge. His current project is Abstract Wikipedia, an initiative that aims to let anyone anywhere on the planet contribute to, and benefit from, the world's collective knowledge, in their native language. It's an ambitious goal, but - inspired by the success of other contributor-driven Wikimedia Foundation projects - Denny is confident that community can make it happen We talked about: his work as Head of Special Projects at the Wikimedia Foundation and his current projects: Wikifunctions and Abstract Wikipedia the origin story of his first project at Wikimedia - Wikidata a precursor project that informed Wikidata - Semantic MediaWiki the resounding success of the Wikidata project, the most edited wiki in the world, with half a million contributors how the need for more expressivity than Wikidata offers led to the idea for Abstract Wikipedia an overview of the Abstract Wikipedia project the abstract language-independent notation that underlies Abstract Wikipedia how Abstract Wikipedia will permit almost instant updating of Wikipedia pages with the facts it provides the capability of Abstract Wikipedia to permit both editing and use of knowledge in an author's native language their exploration of using LLMs to use natural language to create structured representations of knowledge how the design of Abstract Wikipedia encourages and facilitates contributions to the project the Wikifunctions project, a necessary precondition to Abstract Wikipedia the role of Wikidata as the Rosetta Stone of the web some background on the Wikifunctions project the community outreach work that Wikimedia Foundation does and the role of the community in the development of Abstract Wikipedia and Wikifunctions the technical foundations for his how to contribute to Wikimedia Foundation projects his goal to remove language barriers to allow all people to work together in a shared knowledge space a reminder that Tim Berners-Lee's original web browser included an editing function Denny's bio Denny Vrandečić is Head of Special Projects at the Wikimedia Foundation, leading the development of Wikifunctions and Abstract Wikipedia. He is the founder of Wikidata, co-creator of Semantic MediaWiki, and former elected member of the Wikimedia Foundation Board of Trustees. He worked for Google on the Google Knowledge Graph. He has a PhD in Semantic Web and Knowledge Representation from the Karlsruhe Institute of Technology. Connect with Denny online user Denny at Wikimedia Wikidata profile Mastodon LinkedIn email: denny at wikimedia dot org Resources mentioned in this interview Wikimedia Foundation Wikidata Semantic MediaWiki Wikidata: The Making Of Wikifunctions Abstract Wikipedia Meta-Wiki Video Here’s the video version of our conversation: https://youtu.be/iB6luu0w_Jk Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 32. The original plan for the World Wide Web was that it would be a two-way street, with opportunities to both discover and share knowledge. That promise was lost early on - and then restored a few years later when Wikipedia added an "edit" button to the internet. Denny Vrandečić is working to make that edit function even more powerful with Abstract Wikipedia, an innovative platform that lets web citizens both create and consume the world's knowledge, in their own language. Interview transcript Larry: Hi, everyone. Welcome to episode number 32 of the Knowledge Graph Insights podcast. I am really delighted today to welcome to the show Denny Vrandecic. Denny is best known as the founder of Wikidata, which we'll talk about more in just a minute. He's currently the Head of Special Projects at the Wikimedia Foundation. He's also a visiting professor at King's College Lo...
Charles Ivie Since the semantic web was introduced almost 25 years ago, many have dismissed it as a failure. Charles Ivie shows that the RDF standard and the knowledge-representation technology built on it have actually been quite successful. More than half of the world's web pages now share semantic annotations and the widespread adoption of knowledge graphs in enterprises and media companies is only growing as enterprise AI architectures mature. We talked about: his long work history in the knowledge graph world his observation that the semantic web is "the most catastrophically successful thing which people have called a failure" some of the measures of the success of the semantic web: ubiquitous RDF annotations in web pages, numerous knowledge graph deployments in big enterprises and media companies, etc. the long history of knowledge representation the role of RDF as a Rosetta Stone between human knowledge and computing capabilities how the abstraction that RDF permits helps connect different views of knowledge within a domain the need to scope any ontology in a specific domain the role of upper ontologies his transition from computer science and software engineering to semantic web technologies the fundamental role of knowledge representation tech - to help humans communicate information, to innovate, and to solve problems how semantic modeling's focus on humans working things out leads to better solutions than tech-driven approaches his desire to start a conversation around the fundamental upper principles of ontology design and semantic modeling, and his hypothesis that it might look something like a network of taxonomies Charles' bio Charles Ivie is a Senior Graph Architect with the Amazon Neptune team at Amazon Web Services (AWS). With over 15 years of experience in the knowledge graph community, he has been instrumental in designing, leading, and implementing graph solutions across various industries. Connect with Charles online LinkedIn Video Here’s the video version of our conversation: https://youtu.be/1ANaFs-4hE4 Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 31. Since the concept of the semantic web was introduced almost 25 years ago, many have dismissed it as a failure. Charles Ivie points out that it's actually been a rousing success. From the ubiquitous presence of RDF annotations in web pages to the mass adoption of knowledge graphs in enterprises and media companies, the semantic web has been here all along and only continues to grow as more companies discover the benefits of knowledge-representation technology. Interview transcript Larry: Hi everyone. Welcome to episode number 31 of the Knowledge Graph Insights Podcast. I am really happy today to welcome to the show Charles Ivie. Charles is currently a senior graph architect at Amazon's Neptune department. He's been in the graph community for years, worked at the BBC, ran his own consultancies, worked at places like The Telegraph and The Financial Times and places you've heard of. So welcome Charles. Tell the folks a little bit more about what you're up to these days. Charles: Sure. Thanks. Thanks, Larry. Very grateful to be invited on, so thank you for that. And what have I been up to? Yeah, I've been about in the graph industry for about 14 years or something like that now. And these days I am working with the Amazon Neptune team doing everything I can to help people become more successful with their graph implementations and with their projects. And I like to talk at conferences and join things like this and write as much as I can. And occasionally they let me loose on some code too. So that's kind of what I'm up to these days. Larry: Nice. Because you have a background as a software engineer and we will talk more about that later because I think that's really relevant to a lot of what we'll talk about.
Andrea Gioia In recent years, data products have emerged as a solution to the enterprise problem of siloed data and knowledge. Andrea Gioia helps his clients build composable, reusable data products so they can capitalize on the value in their data assets. Built around collaboratively developed ontologies, these data products evolve into something that might also be called a knowledge product. We talked about: his work as CTO at Quantyca, a data and metadata management consultancy his description of data products and their lifecycle how the lack of reusability in most data products inspired his current approach to modular, composable data products - and brought him into the world of ontology how focusing on specific data assets facilitates the creation of reusable data products his take on the role of data as a valuable enterprise asset how he accounts for technical metadata and conceptual metadata in his modeling work his preference for a federated model in the development of enterprise ontologies the evolution of his data architecture thinking from a central-governance model to a federated model the importance of including the right variety business stakeholders in the design of the ontology for a knowledge product his observation that semantic model is mostly about people, and working with them to come to agreements about how they each see their domain Andrea's bio Andrea Gioia is a Partner and CTO at Quantyca, a consulting company specializing in data management. He is also a co-founder of blindata.io, a SaaS platform focused on data governance and compliance. With over two decades of experience in the field, Andrea has led cross-functional teams in the successful execution of complex data projects across diverse market sectors, ranging from banking and utilities to retail and industry. In his current role as CTO at Quantyca, Andrea primarily focuses on advisory, helping clients define and execute their data strategy with a strong emphasis on organizational and change management issues. Actively involved in the data community, Andrea is a regular speaker, writer, and author of 'Managing Data as a Product'. Currently, he is the main organizer of the Data Engineering Italian Meetup and leads the Open Data Mesh Initiative. Within this initiative, Andrea has published the data product descriptor open specification and is guiding the development of the open-source ODM Platform to support the automation of the data product lifecycle. Andrea is an active member of DAMA and, since 2023, has been part of the scientific committee of the DAMA Italian Chapter. Connect with Andrea online LinkedIn (#TheDataJoy) Github Video Here’s the video version of our conversation: https://www.youtube.com/watch?v=g34K_kJGZMc Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 30. In the world of enterprise architectures, data products are emerging as a solution to the problem of siloed data and knowledge. As a data and metadata management consultant, Andrea Gioia helps his clients realize the value in their data assets by assembling them into composable, reusable data products. Built around collaboratively developed ontologies, these data products evolve into something that might also be called a knowledge product. Interview transcript Larry: Hi, everyone. Welcome to episode number 30 of the Knowledge Graph Insights podcast. I'm really happy today to welcome to the show Andrea Gioia. Andrea's, he does a lot of stuff. He's a busy guy. He's a partner and the chief technical officer at Quantyca, a consulting firm that works on data and metadata management. He's the founder of Blindata, a SaaS product that goes with his consultancy. I let him talk a little bit more about that. He's the author of the book Managing Data as a Product, and he's also, he comes out of the data heritage but he's now one of these knowledge people like us.
Dave McComb During the course of his 25-year consulting career, Dave McComb has discovered both a foundational problem in enterprise architectures and the solution to it. The problem lies in application-focused software engineering that results in an inefficient explosion of redundant solutions that draw on overlapping data sources. The solution that Dave has introduced is a data-centric architecture approach that treats data like the precious business asset that it is. We talked about: his work as the CEO of Semantic Arts, a prominent semantic technology and knowledge graph consultancy based in the US the application-centric quagmire that most modern enterprises find themselves trapped in data centricity, the antidote to application centricity his early work in semantic modeling how the discovery of the "core model" in an enterprise facilitates modeling and building data-centric enterprise systems the importance of "baby step" approaches and working with actual customer data in enterprise data projects how building to "enduring business themes" rather than to the needs of individual applications creates a more solid foundation for enterprise architectures his current interest in developing a semantic model for the accounting field, drawing on his history in the field and on Semantic Arts' gist upper ontology the importance of the concept of a "commitment" in an accounting model how his approach to financial modeling permits near-real-time reporting his Data-Centric Architecture Forum, a practitioner-focused event held each June in Ft. Collins, Colorado Dave's bio Dave McComb is the CEO of Semantic Arts. In 2000 he co-founded Semantic Arts with the aim of bringing semantic technology to Enterprises. From 2000- 2010 Semantic Arts focused on ways to improve enterprise architecture through ontology modeling and design. Around 2010 Semantic Arts began helping clients more directly with implementation, which led to the use of Knowledge Graphs in Enterprises. Semantic Arts has conducted over 100 successful projects with a number of well know firms including Morgan Stanley, Electronic Arts, Amgen, Standard & Poors, Schneider-Electric, MD Anderson, the International Monetary Fund, Procter & Gamble, Goldman Sachs as well as a number of government agencies. Dave is the author of Semantics in Business Systems (2003), which made the case for using Semantics to improve the design of information systems, Software Wasteland (2018) which points out how application-centric thinking has led to the deplorable state of enterprise systems and The Data-Centric Revolution (2019) which outlines a alternative to the application-centric quagmire. Prior to founding Semantic Arts he was VP of Engineering for Velocity Healthcare, a dot com startup that pioneered the model driven approach to software development. He was granted three patents on the architecture developed at Velocity. Prior to that he was with a small consulting firm: First Principles Consulting. Prior to that he was part of the problem. Connect with Dave online LinkedIn email: mccomb at semanticarts dot com Semantic Arts Resources mentioned in this interview Dave's books: The Data-Centric Revolution: Restoring Sanity to Enterprise Information Systems Software Wasteland: How the Application-Centric Quagmire is Hobbling Our Enterprises Semantics in Business Systems: The Savvy Manager's Guide gist ontology Data-Centric Architecture Forum Video Here’s the video version of our conversation: https://youtu.be/X_hZG7cFOCE Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 29. Every modern enterprise wrestles with its data, trying to get the most out of it. The smartest businesses have figured out that it isn't just "the new oil" - data is the very bedrock of their enterprise architecture. For the past 25 years, Dave McComb has helped companies understand the...
Ole Olesen-Bagneux In every enterprise, says Ole Olesen-Bagneux, the information you need to understand your organization's metadata is already there. It just needs to be discovered and documented. Ole's Meta Grid can be as simple as a shared, curated collection of documents, diagrams, and data but might also be expressed as a knowledge graph. Ole appreciates "North Star" architectures like microservices and the Data Mesh but presents the Meta Grid as a simpler way to manage enterprise metadata. We talked about: his work as Chief Evangelist at Actian his forthcoming book, "Fundamentals of Metadata Management" how he defines his Meta Grid: an integration architecture that connects metadata across metadata repositories his definition of metadata and its key characteristic, that it's always in two places at once how the Meta Grid compares with microservices architectures and organizing concepts like Data Mesh the nature of the Meta Grid as a small, simple, and slow architecture which is not technically difficult to achieve his assertion that you can't build a Meta Grid because it already exists in every organization the elements of the Meta Grid: documents, diagrams or pictures, and examples of data how knowledge graphs fit into the Meta Grid his appreciation for "North Star" architectures like Data Mesh but also how he sees the Meta Grid as a more pragmatic approach to enterprise metadata management the evolution of his new book from a knowledge graph book to his elaboration on the "slow" nature of the Meta Grid, in particular how its metadata focus contrasts with faster real-time systems like ERPs the shape of the team topology that makes Meta Grid work Ole's bio Ole Olesen-Bagneux is a globally recognized thought leader in metadata management and enterprise data architecture. As VP, Chief Evangelist at Actian, he drives industry awareness and adoption of modern approaches to data intelligence, drawing on his extensive expertise in data management, metadata, data catalogs, and decentralized architectures. An accomplished author, Ole has written The Enterprise Data Catalog (O’Reilly, 2023). He is currently working on Fundamentals of Metadata Management (O’Reilly, 2025), introducing a novel metadata architecture known as the Meta Grid. With a PhD in Library and Information Science from the University of Copenhagen, his unique perspective bridges traditional information science with modern data management. Before joining Actian, Ole served as Chief Evangelist at Zeenea, where he played a key role in shaping and communicating the company’s technology vision. His industry experience includes leadership roles in enterprise architecture and data strategy at major pharmaceutical companies like Novo Nordisk.Ole is passionate about scalable metadata architectures, knowledge graphs, and enabling organizations to make data truly discoverable and usable. Connect with Ole online LinkedIn Substack Medium Resources mentioned in this interview Fundamentals of Metadata Management, Ole's forthcoming book Data Management at Scale by Piethein Strengholt Fundamentals of Data Engineering by Joe Reis and Matt Housley Meta Grid as a Team Topology, Substack article Stewart Brand's Pace Layers Video Here’s the video version of our conversation: https://youtu.be/t01IZoegKRI Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 28. Every modern enterprise wrestles with the scale, the complexity, and the urgency of understanding their data and metadata. So, by necessity, comprehensive architectural approaches like microservices and the data mesh are complex, big, and fast. Ole Olesen-Bagneux proposes a simple, small, and slow way for enterprises to cultivate a shared understanding of their enterprise knowledge, a decentralized approach to metadata strategy that he calls the Meta Grid. Interview transcript Larry: Hi,
Andrea Volpini Your organization's brand is what people say about you after you've left the room. It's the memories you create that determine how people think about you later. Andrea Volpini says that the same dynamic applies in marketing to AI systems. Modern brand managers, he argues, need to understand how both human and machine memory work and then use that knowledge to create digital memories that align with how AI systems understand the world. We talked about: his work as CEO at WordLift, a company that builds knowledge graphs to help companies automate SEO and other marketing activities a recent experiment he did during a talk at an AI conference that illustrates the ability of applications like Grok and ChatGPT to build and share information in real time the role of memory in marketing to current AI architectures his discovery of how the agentic approach he was taking to automating marketing tasks was actually creating valuable context for AI systems the mechanisms of memory in AI systems and an analogy to human short- and long-term memory the similarities he sees in how the human neocortex forms memories and how the knowledge about memory is represented in AI systems his practice of representing entities as both triples and vectors in his knowledge graph how he leverages his understanding of the differences in AI models in his work the different types of memory frameworks to account for in both the consumption and creation of AI systems: semantic, episodic, and procedural his new way of thinking about marketing: as a memory-creation process the shift in focus that he thinks marketers need to make, "creating good memories for AI in order to protect their brand values" Andrea's bio Andrea Volpini is the CEO of WordLift and co-founder of Insideout10. With 25 years of experience in semantic web technologies, SEO, and artificial intelligence, he specializes in marketing strategies. He is a regular speaker at international conferences, including SXSW, TNW Conference, BrightonSEO, The Knowledge Graph Conference, G50, Connected Data and AI Festival. Andrea has contributed to industry publications, including the Web Almanac by HTTP Archive. In 2013, he co-founded RedLink GmbH, a commercial spin-off focused on semantic content enrichment, natural language processing, and information extraction. Connect with Andrea online LinkedIn X Bluesky WordLift Video Here’s the video version of our conversation: https://youtu.be/do-Y7w47CZc Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 27. Some experts describe the marketing concept of branding as, What people say about you after you’ve left the room. It's the memories they form of your company that define your brand. Andrea Volpini sees this same dynamic unfolding as companies turn their attention to AI. To build a memorable brand online, modern marketers need to understand how both human and machine memory work and then focus on creating memories that align with how AI systems understand the world. Interview transcript Larry: Hi, everyone. Welcome to episode number 27 of the Knowledge Graph Insights podcast. I am really delighted today to welcome to the show Andrea Volpini. Andrea is the CEO and the founder at WordLift, a company based in Rome. Tell the folks a little bit more about WordLift and what you're up to these days, Andrea. Andrea: Yep. So we build knowledge graphs and to help brands automate their SEO and marketing efforts using large language model and AI in general. Larry: Nice. Yeah, and you're pretty good at this. You've been doing this a while and you had a recent success story, I think that shows, that really highlights some of your current interests in your current work. Tell me about your talk in Milan and the little demonstration you did with that. Andrea: Yeah, yeah, so it was last week at AI Festival,
Jacobus Geluk The arrival of AI agents creates urgency around the need to guide and govern them. Drawing on his 15-year history in building reliable AI solutions for banks and other enterprises, Jacobus Geluk sees a standards-based data-product marketplace as the key to creating the thriving data economy that will enable AI agents to succeed at scale. Jacobus launched the effort to create the DPROD data-product description specification, creating the supply side of the data market. He's now forming a working group to document the demand side, a "use-case tree" specification to articulate the business needs that data products address. We talked about: his work as CEO at Agnos.ai, an enterprise knowledge graph and AI consultancy the working group he founded in 2023 which resulted in the DPROD specification to describe data products an overview of the data-product marketplace and the data economy the need to account for the demand side of the data marketplace the intent of his current work on to address the disconnect between tech activities and business use cases how the capabilities of LLMs and knowledge graphs complement each other the origins of his "use-case tree" model in a huge banking enterprise knowledge graph he built ten years ago how use case trees improve LLM-driven multi-agent architectures some examples of the persona-driven, tech-agnostic solutions in agent architectures that use-case trees support the importance of constraining LLM action with a control layer that governs agent activities, accounting for security, data sourcing, and issues like data lineage and provenance the new Use Case Tree Work Group he is forming the paradox in the semantic technology industry now of a lack of standards in a field with its roots in W3C standards Jacobus' bio Jacobus Geluk is a Dutch Semantic Technology Architect and CEO of agnos.ai, a UK-based consulting firm with a global team of experts specializing in GraphAI — the combination of Enterprise Knowledge Graphs (EKG) with Generative AI (GenAI). Jacobus has over 20 years of experience in data management and semantic technologies, previously serving as a Senior Data Architect at Bloomberg and Fellow Architect at BNY Mellon, where he led the first large-scale production EKG in the financial industry. As a founding member and current co-chair of the Enterprise Knowledge Graph Forum (EKGF), Jacobus initiated the Data Product Workgroup, which developed the Data Product Ontology (DPROD) — a proposed OMG standard for consistent data product management across platforms. Jacobus can claim to have coined the term "Enterprise Knowledge Graph (EKG)" more than 10 years ago, and his work has been instrumental in advancing semantic technologies in financial services and other information-intensive industries. Connect with Jacobus online LinkedIn Agnos.ai Resources mentioned in this podcast DPROD specification Enterprise Knowledge Graph Forum Object Management Group Use Case Tree Method for Business Capabilities DCAT Data Catalog Vocabulary Video Here’s the video version of our conversation: https://youtu.be/J0JXkvizxGo Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 26. In an AI landscape that will soon include huge groups of independent software agents acting on behalf of humans, we'll need solid mechanisms to guide the actions of those agents. Jacobus Geluk looks at this situation from the perspective of the data economy, specifically the data-products marketplace. He helped develop the DPROD specification that describes data products and is now focused on developing use-case trees that describe the business needs that they address. Interview transcript Larry: Okay. Hi everyone. Welcome to episode number 26 of the Knowledge Graph Insights podcast. I am really happy today to welcome to the show, Jacobus Geluk. Sorry, I try to speak Dutch, do my best.
Rebecca Schneider Skills that Rebecca Schneider learned in library science school - taxonomy, ontology, and semantic modeling - have only become more valuable with the arrival of AI technologies like LLMs and the growing interest in knowledge graphs. Two things have stayed constant across her library and enterprise content strategy work: organizational rigor and the need to always focus on people and their needs. We talked about: her work as Co-Founder and Executive Director at AvenueCX, an enterprise content strategy consultancy her background as a "recovering librarian" and her focus on taxonomies, metadata, and structured content the importance of structured content in LLMs and other AI applications how she balances the capabilities of AI architectures and the needs of the humans that contribute to them the need to disambiguate the terms that describe the span of the semantic spectrum the crucial role of organization in her work and how you don't to have formally studied library science to do it the role of a service mentality in knowledge graph work how she measures the efficiency and other benefits of well-organized information how domain modeling and content modeling work together in her work her tech-agnostic approach to consulting the role of metadata strategy into her work how new AI tools permit easier content tagging and better governance the importance of "knowing your collection," not becoming a true subject matter expert but at least getting familiar with the content you are working with the need to clean up your content and data to build successful AI applications Rebecca's bio Rebecca is co-founder of AvenueCX, an enterprise content strategy consultancy. Her areas of expertise include content strategy, taxonomy development, and structured content. She has guided content strategy in a variety of industries: automotive, semiconductors, telecommunications, retail, and financial services. Connect with Rebecca online LinkedIn email: rschneider at avenuecx dot com Video Here’s the video version of our conversation: https://youtu.be/ex8Z7aXmR0o Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 25. If you've ever visited the reference desk at your local library, you've seen the service mentality that librarians bring to their work. Rebecca Schneider brings that same sensibility to her content and knowledge graph consulting. Like all digital practitioners, her projects now include a lot more AI, but her work remains grounded in the fundamentals she learned studying library science: organizational rigor and a focus on people and their needs. Interview transcript Larry: Hi, everyone. Welcome to episode number 25 of the Knowledge Graph Insights podcast. I am really excited today to welcome to the show Rebecca Schneider. Rebecca is the co-founder and the executive director at AvenueCX, a consultancy in the Boston area. Welcome, Rebecca. Tell the folks a little bit more about what you're up to these days. Rebecca: Hi, Larry. Thanks for having me on your show. Hello, everyone. My name is Rebecca Schneider. I am a recovering librarian. I was a trained librarian, worked in a library with actual books, but for most of my career, I have been focusing on enterprise content strategy. Furthermore, I typically focus on taxonomies, metadata, structured content, and all of that wonderful world that we live in. Larry: Yeah, and we both come out of that content background and have sort of converged on the knowledge graph background together kind of over the same time period. And it's really interesting, like those skills that you mentioned, the library science skills of taxonomy, metadata, structured, and then the application of that in structured content in the content world, how, as you've got in more and more into knowledge graph stuff, how has that background, I guess...
Ashleigh Faith With her 15-year history in the knowledge graph industry and her popular YouTube channel, Ashleigh Faith has informed and inspired a generation of graph practitioners and enthusiasts. She's an expert on semantic modeling, knowledge graph construction, and AI architectures and talks about those concepts in ways that resonate both with her colleagues and with newcomers to the field. We talked about: her popular IsA DataThing YouTube channel the crucial role of accurately modeling actual facts in semantic practice and AI architectures her appreciation of the role of knowledge graphs in aligning people in large organizations around concepts and the various words that describe them the importance of staying focused on the business case for knowledge graph work, which has become both more important with the arrival of LLMs and generative AI the emergence of more intuitive "talk to your graph" interfaces some of her checklist items for onboarding aspiring knowledge graph engineers how to decide whether to use a property graph or a knowledge graph, or both her hope that more RDF graph vendors will offer a free tier so that people can more easily experiment with them approaches to AI architecture orchestration the enduring importance of understanding how information retrieval works Ashleigh's bio Ashleigh Faith has her PhD in Advanced Semantics and over 15 years of experience working on graph solutions across the STEM, government, and finance industries. Outside of her day-job, she is the Founder and host of the IsA DataThing YouTube channel and podcast where she tries to demystify the graph space. Connect with Ashleigh online LinkedIn IsA DataThing YouTube channel Video Here’s the video version of our conversation: https://youtu.be/eMqLydDu6oY Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 24. One way to understand the entity resolution capabilities of knowledge graphs is to picture on old-fashioned telephone operator moving plugs around a switchboard to make the right connections. Early in her career, that's one way that Ashleigh Faith saw the power of knowledge graphs. She has since developed sophisticated approaches to knowledge graph construction, semantic modeling, and AI architectures and shares her deeply informed insights on her popular YouTube channel. Interview transcript Larry: Hi, everyone. Welcome to episode number 24 of the Knowledge Graph Insights Podcast. I am super extra delighted today to welcome to the show Ashleigh Faith. Ashleigh is the host of the awesome YouTube channel IsA DataThing, which has thousands of subscribers, thousands of monthly views. I think it's many people's entry point into the knowledge graph world. Welcome, Ashleigh. Great to have you here. Tell the folks a little bit more about what you're up to these days. Ashleigh: Thanks, Larry. I've known you for quite some time. I'm really excited to be here today. What about me? I do a lot of semantic and AI stuff for my day job. But yeah, I think my main passion is also helping others get involved, understand some of the concepts a little bit better for the semantic space and now the neuro-symbolic AI. That's AI and knowledge graphs coming together. That is quite a hot topic right now, so lots and lots of untapped potential in what we can talk about. I do most of that on my channel. Larry: Yeah. I will refer people to your channel because we've got only a half-hour today. It's ridiculous. Ashleigh: Yeah. Larry: We just talked for an hour before we went on the air. It's ridiculous. What I'd really like to focus on today is the first stage in any of this, the first step in any of these knowledge graph implementations or any of this stuff is modeling. I think about it from a designerly perspective. I do a lot of mental model discernment, user research kind of stuff, and then conceptual modeling to agree on things.
Panos Alexopoulos Any knowledge graph or other semantic artifact must be modeled before it's built. Panos Alexopoulos has been building semantic models since 2006. In 2020, O'Reilly published his book on the subject, "Semantic Modeling for Data." The book covers the craft of semantic data modeling, the pitfalls practitioners are likely to encounter, and the dilemmas they'll need to overcome. We talked about: his work as Head of Ontology at Textkernel and his 18-year history working with symbolic AI and semantic modeling his definition and description of the practice of semantic modeling and its three main characteristics: accuracy, explicitness, and agreement the variety of artifacts that can result from semantic modeling: database schemas, taxonomies, hierarchies, glossaries, thesauri, ontologies, etc. the difference between identifying entities with human understandable descriptions in symbolic AI and numerical encodings in sub-symbolic AI the role of semantic modeling in RAG and other hybrid AI architectures a brief overview of data modeling as a practice how LLMs fit into semantic modeling: as sources of information to populate a knowledge graph, as coding assistants, and in entity and relation extraction other techniques besides NLP and LLMs that he uses in his modeling practice: syntactic patterns, heuristics, regular expressions, etc. the role of semantic modeling and symbolic AI in emerging hybrid AI architectures the importance of defining the notion of "autonomy" as AI agents emerge Panos' bio Panos Alexopoulos has been working since 2006 at the intersection of data, semantics and software, contributing in building intelligent systems that deliver value to business and society. Born and raised in Athens, Greece, Panos currently works as a principal educator at OWLTECH, developing and delivering training workshops that provide actionable knowledge and insights for data and AI practitioners. He also works as Head of Ontology at Textkernel BV, in Amsterdam, Netherlands, leading a team of data professionals in developing and delivering a large cross-lingual Knowledge Graph in the HR and Recruitment domain. Panos has published several papers at international conferences, journals and books, and he is a regular speaker in both academic and industry venues. He is also the author of the O’Reilly book “Semantic Modeling for Data – Avoiding Pitfalls and Dilemmas”, a practical and pragmatic field guide for data practitioners that want to learn how semantic data modeling is applied in the real world. Connect with Panos online LinkedIn Video Here’s the video version of our conversation: https://youtu.be/ENothdlfYGA Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 23. In order to build a knowledge graph or any other semantic artifact, you first need to model the concepts you're working with, and that model needs to be accurate, to explicitly represent all of the ideas you're working with, and to capture human agreements about them. Panos Alexopoulos literally wrote the book on semantic modeling for data, covering both the principles of modeling as well as the pragmatic concerns of real-world modelers. Interview transcript Larry: Hi everyone. Welcome to episode number 23 of the Knowledge Graph Insights podcast. I am really excited today to welcome to the show Panos Alexopoulos. Panos is the head of ontology at Textkernel, a company in Amsterdam that works on knowledge graphs for the HR and recruitment world. Welcome, Panos. Tell the folks a little bit more about what you're doing these days. Panos: Hi Larry. Thank you very much for inviting me to your podcast. I'm really happy to be here. Yeah, so as you said, I'm head of ontology at Textkernel. Actually, I've been working in the field of data semantics, knowledge graph ontologies for almost now 18 years, even before the era of machine learning,
Mike Pool Mike Pool sees irony in the fact that semantic-technology practitioners struggle to use the word "semantics" in ways that meaningfully advance conversations about their knowlege-representation work. In a recent LinkedIn post, Mike even proposed a moratorium on the use of the word. We talked about: his multi-decade career in knowledge representation and ontology practice his opinion that we might benefit from a moratorium on the term "semantics" the challenges in pinning down the exact scope of semantic technology how semantic tech permits reusability and enables scalability the balance in semantic practice between 1) ascribing meaning in tech architectures independent of its use in applications and 2) considering end-use cases the importance of staying domain-focused as you do semantic work how to stay pragmatic in your choice of semantic methods how reification of objects is not inherently semantic but does create a framework for discovering meaning how to understand and capture subtle differences in meaning of seemingly clear terms like "merger" or "customer" how LLMs can facilitate capturing meaning Mike's bio Michael Pool works in the Office of the CTO at Bloomberg, where he is working on a tool to create and deploy ontologies across the firm. Previously, he was a principal ontologist on the Amazon Product Knowledge team, and has also worked to deploy semantic technologies/approaches and enterprise knowledge graphs at a number of big banks in New York City. Michael also spent a couple of years on the famous Cyc project and has evaluated knowledge representation technologies for DARPA. He has also worked on tooling to integrate probabilistic and semantic models and oversaw development of an ontology to support a consumer-facing semantic search engine. He lives in New York City and loves to run around in circles in Central Park. Connect with Mike online LinkedIn Video Here’s the video version of our conversation: https://youtu.be/JlJjBWGwSDg Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 22. The word "semantics" is often used imprecisely by semantic-technology practitioners. It can describe a wide array of knowledge-representation practices, from simple glossaries and taxonomies to full-blown enterprise ontologies, any of which may be summarized in a conversation as "semantics." Mike Pool thinks that this dynamic - using a word that lacks precise meaning while assuming that it communicates a lot - may justify a moratorium on the use of the term. Interview transcript Larry: Hi everyone, welcome to episode number 22 of the Knowledge Graph Insights podcast. I'm really happy today to welcome to the show Mike Pool. Mike is a longtime ontologist, a couple of decades plus. He recently took a position at Bloomberg. But he made this really provocative post on LinkedIn lately that I want to flesh out today, and we'll talk more about that throughout the rest of the show. Welcome, Mike, tell the folks a little bit more about what you're up to these days. Mike: Hey, thank you, Larry. Yeah. As you noted, I've just taken a position with Bloomberg and for these many years that you alluded to, I've been very heavily focused on building, doing knowledge representation in general. In the last let's say decade or so I've been particularly focused on using ontologies and knowledge graphs in large banks, or large organizations at least, to help organize disparate data, to make it more accessible, breakdown data silos, et cetera. It's particularly relevant in the finance industry where things can be sliced and diced in so many different ways. I find there's a really important use case in the financial space but in large organizations in general, in my opinion, for using ontology. So that's a lot of what I've been thinking about, to make that more accessible to the organization and to help them build these ontologies and utilize th...
Margaret Warren As a 10-year-old photographer, Margaret Warren would jot down on the back of each printed photo metadata about who took the picture, who was in it, and where it was taken. Her interest in image metadata continued into her adult life, culminating the creation of ImageSnippets, a service that lets anyone add linked open data descriptions to their images. We talked about: her work to make images more discoverable with metadata connected via a knowledge graph how her early childhood history as a metadata strategist, her background in computing technology, and her personal interest in art and a photography shows up in her product, ImageSnippets her takes on the basics of metadata strategy and practice the many types of metadata: descriptive, administrative, technical, etc. the role of metadata in the new AI world some of the good and bad reasons that social media platforms might remove metadata from images privacy implications of metadata in social media the linked data principles that she applies in ImageSnippets and how they're managed in the product's workflow her wish that CMSs and social media platforms would not strip the metadata from images as they ingest them the lightweight image ontology that underlies her ImageSnippets product her prediction that the importance of metadata that supports provenance, demonstrates originality, and sets context will continue to grow in the future Margaret's bio Margaret Warren is a technologist, researcher and artist/content creator. She is the founder and CEO of Metadata Authoring Systems whose mission is to make the most obscure images on the web findable, and easily accessible by describing and preserving them in the most precise ways possible. To assist with this mission, she is the creator of a system called, ImageSnippets which can be used by anyone to build linked data descriptions of images into graphs. She is also a research associate with the Florida Institute of Human and Machine Cognition, one of the primary organizers of a group called The Dataworthy Collective and is a member of the IPTC (International Press and Telecommunications Council) photo-metadata working group and the Research Data Alliance charter on Collections as Data. As a researcher, Margaret's primary focus is at the intersection of semantics, metadata, knowledge representation and information science particularly around visual content, search and findability. She is deeply interested in how people describe what they experience visually and how to capture and formalize this knowledge into machine readable structures. She creates tools and processes for humans but augmented by machine intelligence. Many of these tools are useful for unifying the many types of metadata and descriptions of images - including the very important context element - into ontology infused knowledge graphs. Her tools can be used for tasks as advanced as complex domain modeling but can also facilitate image content to be shared and published while staying linked to it's metadata across workflows. Learn more and connect with Margaret online LinkedIn Patreon Bluesky Substack ImageSnippets Metadata Authoring Systems personal and art site IPTC links IPTC Photo Metadata Software that supports IPTC Photo Metadata Get IPTC Photo Metadata Browser extensions for IPTC Photo Metadata Resource not mentioned in podcast (but very useful for examining structured metadata in web pages) OpenLink Structured Data Sniffer (OSDS) Video Here’s the video version of our conversation: https://youtu.be/pjoAAq5zuRk Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 21. Nowadays, we are all immersed in a deluge of information and media, especially images. The real value of these images is captured in the metadata about them. Without information about the history of an image, its technical details,
Jans Aasman Hybrid AI architectures get more complex every day. For Jans Aasman, large language models and generative AI are just the newest additions to his toolkit. Jans has been building advanced hybrid AI systems for more than 15 years, using knowledge graphs, symbolic logic, and machine learning - and now LLMs and gen AI - to build advanced AI systems for Fortune 500 companies. We talked about: his knowledge graph and neuro-symbolic work as the CEO of Franz the crucial role of a visionary knowledge graph champion in KG adoption in enterprises the two types of KG champions he has encountered: the magic-seeking, forward-looking technologist and the more pragmatic IT leader trying to better organize their operation the AI architectural patterns and themes he has seen emerge over the past 25 years: logic, reasoning, event-based KGs, machine learning, and of course gen AI and LLMs how gen AI lets him do things he couldn't have imagined five years ago the enduring importance of enterprise taxonomies, especially in RAG architectures which business entities need to be understood to answer complex business questions his approach to neuro-symbolic AI, seeing it as a "fluid interplay between a knowledge graph, symbolic logic, machine learning, and generative AI" the power of "magic predicates" a common combination of AI technologies and human interactions that can improve medical diagnosis and care decisions his strong belief in keeping humans in the loop in AI systems his observation that technology and business leaders seeing the need for "a symbolic approach next to generative AI" his take on the development of reasoning capabilities of LLMs how the code-generation capabilities of LLMs are more beneficial to senior programmers and may even impede the work of less experiences coders Jans' bio Jans Aasman is a Ph.D. psychologist and expert in Cognitive Science - as well as CEO of Franz Inc., an early innovator in Artificial Intelligence and provider of Knowledge Graph Solutions based on AllegroGraph. As both a scientist and CEO, Dr. Aasman continues to break ground in the areas of Artificial Intelligence and Knowledge Graphs as he works hand-in-hand with numerous Fortune 500 organizations as well as government entities worldwide. Connect with Jans online LinkedIn email: ja at franz dot com Video Here’s the video version of our conversation: https://www.youtube.com/watch?v=SZBZxC8S1Uk Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 20. The mix of technologies in hybrid artificial intelligence systems just keeps getting more interesting. This might seem like a new phenomenon, but long before our LinkedIn feeds were clogged with posts about retrieval augmented generation and neuro-symbolic architectures, Jans Aasman was building AI systems that combined knowledge graphs, symbolic logic, and machine learning. Large language models and generative AI are just the newest technologies in his AI toolkit. Interview transcript Larry: Hi, everyone. Welcome to episode number 20 of the Knowledge Graph Insights podcast. I am really delighted today to welcome to the show Jans Aasmann. Jans is, he originally started out as a psychologist and he got into cognitive science. For the past 20 years, he's run a company called Franz, where he's the CEO doing neuro-symbolic AI, so welcome, Jans. Tell the folks a little bit more about what you're doing these days. Jans: We help companies build knowledge graphs, but with the special angle that we now offer neuro-symbolic AI so that we, in a very fluid way, mix traditional symbolic logic and the traditional machine learning with the new generative AI. We do this in every possible combination that you could think of. Larry: Who? Jans: These applications might be in healthcare or in call centers or in publishing. It's many, many, many different domains it supplies. Larry:
Juan Sequeda Knowledge graph technology has been around for decades. The benefits so far accruing to only a few big enterprises and tech companies. Juan Sequeda sees large language models as a critical enabler for the broader adoption of KGs. With their capacity to accelerate the acquisition and use of valuable business knowledge, LLMs offer a path to a better return on your enterprise's investment in semantics. We talked about: his work data.world as Principal scientist and the head of the AI lab at data.world the new discovery and knowledge-acquisition capabilities that LLMs give knowledge engineers a variety of business benefits that unfold from these new capabilities the payoff of investing in semantics and knowledge: "one plus one is greater than two" how semantic understanding and the move from a data-first world to a knowledge-first world helps businesses make better decisions and become more efficient the pendulum swings in the history of the development of AI and knowledge systems his research with Dean Allemang on how knowledge graphs can help LLMs improve the accuracy of answers of questions posed to enterprise relational databases the role of industry benchmarks in understanding the return on your invest in semantics the importance of treating semantics as a first-class citizen how business leaders can recognize and take advantage of the semantics and knowledge work that is already happening in their organizations Juan's bio Juan Sequeda is the Principal Scientist and Head of the AI Lab at data.world. He holds a PhD in Computer Science from The University of Texas at Austin. Juan’s research and industry work has been on the intersection of data and AI, with the goal to reliably create knowledge from inscrutable data, specifically designing and building Knowledge Graph for enterprise data and metadata management. Juan is the co-author of the book “Designing and Building Enterprise Knowledge Graph” and the co-host of Catalog and Cocktails, an honest, no-bs, non-salesy data podcast. Connect with Juan online LinkedIn Catalog & Cocktails podcast Video Here’s the video version of our conversation: https://youtu.be/xZq12K7GvB8 Podcast intro transcript This is the Knowledge Graph Insights podcast, episode number 19. The AI pendulum has been swinging back and forth for many decades. Juan Sequeda argues that we're now at a point in the advancement of AI technology where businesses can fully reap its long-promised benefits. The key is a semantic understanding of your business, captured in a knowledge graph. Juan sees large language models as a critical enabler of this capability, in particular the ability of LLMs to accelerate the acquisition and use of valuable business knowledge. Interview transcript Larry: Hi, everyone. Welcome to episode number 19 of the Knowledge Graph Insights podcast. I am really delighted today to welcome to the show Juan Sequeda. Juan is the principal scientist and the head of the AI lab at data.world. He's also the co-host of the really good popular podcast, Catalog & Cocktails. So welcome, Juan. Tell the folks a little bit more about what you're up to these days. Juan: Hey, very great. Thank you so much for having me. Great to chat with you. So what am I up to now these days? Obviously, knowledge graphs is something that is my entire life of what I've been doing. This was before it was called knowledge graphs. I would say that the last year, year-and-a-half, almost two years now, I would say, is been understanding the relationship between knowledge graphs and LLMs. If people have been following our work, what we've been doing a lot has been on understanding how to use knowledge graphs to increase the accuracy for your chat with your data system, so be able to do question answering over your structured SQL databases and how knowledge graphs increase the accuracy of that. So we can chat about that. Juan:
loading
Comments