DiscoverThe MapScaping Podcast - GIS, Geospatial, Remote Sensing, earth observation and digital geography
The MapScaping Podcast - GIS, Geospatial, Remote Sensing, earth observation and digital geography
Claim Ownership

The MapScaping Podcast - GIS, Geospatial, Remote Sensing, earth observation and digital geography

Author: MapScaping

Subscribed: 957Played: 35,127
Share

Description

A podcast for geospatial people. Weekly episodes that focus on the tech, trends, tools, and stories from the geospatial world. Interviews with the people that are shaping the future of GIS, geospatial as well as practitioners working in the geo industry.

This is a podcast for the GIS and geospatial community subscribe or visit https://mapscaping.com to learn more
249 Episodes
Reverse
Geospatial Product Swiss Army Knife 1. The "Build It and They Won't Come" Trap We have all seen it: a talented geospatial professional spends months—perhaps years—perfecting a technically sophisticated web map or a niche data service, only to release it to a deafening silence. In our industry, the "build it and they will come" philosophy is a fast track to zero traction. Precision is the enemy of progress when it is applied to the wrong problem. Daniel and Stella Blake Kelly explored a remedy for this pattern. Stella—a New Zealand-born, Sydney-based strategist and founder of the consultancy Cartisan—didn’t start with a master plan. She "fell into" the industry after being inspired by a lecturer with bright blue hair and a passion for GIS that rivaled a Lego builder’s creativity. Today, she helps organizations move from "making things" to "building products that matter" using a framework she calls the Product Swiss Army Knife. -------------------------------------------------------------------------------- 2. The 7-Step Framework: More Than Just a Map Many geospatial experts suffer from a technology-first bias, prioritizing data accuracy over strategic utility. To counter this, Stella advocates for a disciplined, seven-tool toolkit designed to bridge the gap between GIS and Product Design: Vision: Establish a clear statement of what you are building and why it needs to exist. User Needs: Move beyond assumptions to identify real users and their specific friction points. Market & Context: Analyze the existing ecosystem (competitors, data, and workflows) to find your gap. Features: Ruthlessly prioritize "must-haves" to define a lean Minimum Viable Product (MVP). Prototypes & User Flows: Map out the user’s journey through the service before writing a line of code. Proof of Concept: Create a tangible, working version to prove the technical and market logic. Launch & Learn: Release early to gather real-world data and iterate based on evidence. This structure forces builders to treat the "spatial" element as a solution rather than the entire product. To illustrate User Needs (Tool #2), Stella suggests using formal User Stories to step out of the technical mindset: "As a solar panel marketer, I want to find potential customers with enough roof surface area so that I can reach out to them and provide an accurate quote." By grounding the project in a specific human problem, the developer stops building for themselves and starts building for the market. As Stella notes: "The thing about the product Swiss Army knife... is that it can be applied to almost any situation where there is an end consumer, where somebody is going to use the thing, the service that you make." -------------------------------------------------------------------------------- 3. The "200 Tools" Strategy: Programmatic Market Validation Daniel shared an unconventional approach to product discovery that serves as a masterclass in Market Context (Tool #3). Leveraging AI, he has built nearly 200 simple geospatial tools—such as a "Roof Area Calculator"—not as final products, but as a "sandbox" for discovery. This is Programmatic Market Validation. Instead of starting with a complex SaaS model, Daniel uses these micro-tools to find "winners" via organic search traffic. By observing where the internet already has unsolved spatial queries, he lets the market dictate which products deserve a full-scale build. In this new landscape, the barrier to entry has shifted: the competitive advantage is no longer "coding ability"—it is strategic experimentation. -------------------------------------------------------------------------------- 4. Not All Traffic is Equal: The High-Value Keyword Insight One of the most surprising takeaways from this experimentation is the direct link between specific geospatial problems and commercial value. A general GIS data tool might get thousands of views, but a "Roof Area Calculator" generates significantly higher programmatic advertising revenue. The reason? Market Context. The keyword "roofing" implies high-value intent; a user measuring their roof is likely in the market for a new one, making them incredibly valuable to advertisers. Understanding the commercial landscape surrounding a user's problem is the difference between a struggling hobby project and a viable MicroSaaS. -------------------------------------------------------------------------------- 5. The Precision Paradox: Why GIS Experts Struggle with UX There is a fundamental tension between the geospatial technical mindset and the product design mindset. GIS professionals are trained to be exact, precise, and correct. Designers, however, are taught to be wrong, gather feedback, and iterate. Daniel illustrated this with a "Hot Jar" anecdote. He once built a site where users were failing to move through the revenue funnel. Heat maps revealed the issue wasn't the data—it was the layout. Users weren't scrolling down far enough to see the critical action button. The data was perfect, but the UX was broken. Stella emphasizes that building a product requires the humility to accept that "the best designers of products are the users themselves." Success often comes from moving a button or simplifying a flow, not from adding another decimal point of precision to the underlying geometry. -------------------------------------------------------------------------------- 6. Launching "Soft" to De-Risk the Rollout The "perfectionism trap" is the primary reason geospatial products fail to launch. Builders fear that "releasing slop" will damage their brand. However, Stella suggests the Soft Launch (Tool #7) as a vital de-risking mechanism. A soft launch allows you to: Prevent Stagnation: Avoid the "quiet abandonment" of projects that never see the light of day. Validate Demand: Ensure people actually want the tool before committing to months of development. Build Brand and Trust: In a world where anyone can spin up a tool with AI, trust is the ultimate differentiator. Launching early ensures continuous improvement and prevents the high-stakes pressure of a single "grand opening" that may miss the mark entirely. -------------------------------------------------------------------------------- 7. Conclusion: The Final Ponderance Building successful geospatial products is about empathy and process, not just pixels and polygons. Whether you are building a global API or an internal tool for a government agency, the principles of the Swiss Army Knife remain the same. At the recent Phosphag workshop in Oakland, the range of products—from print maps to digital twins—all shared a common hurdle: the energy to push through the "perfection barrier." As you look at your current projects, ask yourself: Am I building this because the data exists, or because a human has a problem I can solve? Success in the modern landscape requires a diversity of skills—brand, marketing, and distribution. If you aren't embarrassed by your first version, you’ve already lost the market. Stop building in the dark. Get out there and build the thing.
Why Machine-Writing Code is the Best (and Most Dangerous) Thing for Geospatial:   The current discourse surrounding AI coding is nothing if not polarized. On one side, the technofuturists urge us to throw away our keyboards; on the other, skeptics dismiss Large Language Models (LLMs) as little more than "fancy autocomplete" that will never replace a "real" engineer. Both sides miss the nuanced reality of the shift we are living through right now.   I recently sat down with Matt Hansen, Director of Geospatial Ecosystems at Element 84, to discuss this transition. With a 30-year career spanning the death of photographic film to the birth of Cloud-Native Geospatial, Hansen has a unique vantage point on how technology shifts redefine our roles. He isn’t predicting a distant future; he is describing a present where the barrier between an idea and a functioning tool has effectively collapsed.   The "D" Student Who Built the Future Hansen’s journey into the heart of open-source leadership began with what he initially thought was a terminal failure. As a freshman at the Rochester Institute of Technology, he found himself in a C programming class populated almost entirely by seasoned professionals from Kodak. Intimidated and overwhelmed by the "syntax wall," he withdrew from the class the first time and scraped by with a "D" on his second attempt. For years, he believed software simply wasn't his path. Today, however, he is a primary architect of the SpatioTemporal Asset Catalog (STAC) ecosystem and a major open-source contributor. This trajectory is the perfect case study for the democratizing power of AI: it allows the subject matter expert—the person who understands "photographic technology" or "imaging science"—to bypass the mechanical hurdles of brackets and semi-colons. "I took your class twice and thought I was never software... and now here I am like a regular contributor to open source software for geospatial." — Matt Hansen to his former professor.   The Rise of "Vibe Coding" and the Fragmentation Trap   We are entering the era of "vibe coding," where developers prompt AI based on a general description or "vibe" of what they need. While this is exhilarating for the individual, it creates a systemic risk of "bespoke implementations." When a user asks an AI for a solution without a deep architectural understanding, the machine often generates a narrow, unvetted fragment of code rather than utilizing a secure, scalable library. The danger here is a catastrophic loss of signal. If thousands of users release these AI-generated fragments onto platforms like GitHub, we risk drowning out the vetted, high-quality solutions that the community has spent decades building. We are creating a "sea of noise" that could make it harder for both humans and future AI models to identify the standard, proper way to solve a problem.   Why Geospatial is Still "Special" (The Anti-meridian Test)   For a long time, the industry mantra has been "geospatial isn’t special," pushing for spatial data to be treated as just another data type, like in GeoParquet. However, Hansen argues that AI actually proves that domain expertise is more critical than ever. Without specific guidance, AI often fails to account for the unique edge cases of a spherical world. Consider the "anti-meridian" problem: polygons crossing the 180th meridian. When asked to handle spatial data, an AI will often "brute force" a custom logic that works for a small, localized dataset but fails the moment it encounters the wrap-around logic of a global scale. A domain expert knows to direct the AI toward Pete Kadomsky’s "anti-meridian" library. AI is not a subject matter expert; it is a powerful engine that requires an expert navigator to avoid the "Valley of Despair."   Documentation is Now SEO for the Machines   We are seeing a counterintuitive shift in how we value documentation. Traditionally, README files and tutorials were written by humans, for humans. In the age of AI, documentation has become the primary way we "market" our code to the machines. If your open-source project lacks a clean README or a rigorous specification, it is effectively invisible to the AI-driven future of development. By investing in high-quality documentation, developers are engaging in a form of technical SEO. You are ensuring that when an AI looks for the "signal" in the noise, it chooses your vetted library because it is the most readable and reliable option available.   From Software Developers to Software Designers   The role of the geospatial professional is shifting from writing syntax to what Hansen calls the "Foundry" model. Using tools like GitHub Specit, the human acts as a designer, defining rigorous blueprints, constraints, and requirements in human language. The machine then executes the "how," while the human remains the sole arbiter of the "what" and "why." Hansen’s advice for the next generation—particularly those entering a job market currently hostile to junior engineers—is to abandon generalism. Don't just learn to code; become a specialist in a domain like geospatial. The ability to write Python is becoming a commodity, but the ability to design a system that accounts for the nuances of remote sensing is an increasingly rare and valuable asset.   History Repeats: The "Priesthood" of Assembly   This shift mirrors the 1950s, when the "priesthood" of assembly programmers looked at the first compilers with deep suspicion. Kathleen Booth, who wrote the first assembly language, lived in a world where manual coding was an arcane, elite skill. Those early programmers argued that compilers were untrustworthy and that a human could always write "better" code by hand. They were technically right about efficiency, but they were wrong about the future. Just as the compiler was "good enough" to allow us to move "up the stack" and take on more complex problems, AI is the next level of abstraction. We might use a "Ralph Wiggum script"—a loop that feeds AI output back into itself until the task is "done"—and while it may be a brute-force method, it is often more productive than the perfection of the past.   Conclusion: The Future is a Specialist's Game   We are moving away from being the writers of code and toward being the designers of systems. While the "syntax wall" has been demolished, the requirement for domain knowledge has only grown higher. The keyboard isn't dying; it is being repurposed for higher-level architectural thought.   As the industry experiences a "recursive improvement" of these tools, the question for every professional is no longer about whether the machine can do your job. It’s whether you have the specialized expertise to tell the machine what a "good enough" job actually looks like. Are you prepared to stop being a coder and start being a designer?
How can you accurately aggregate and compare point-based data from different parts of the world? When analyzing crime rates, population, or environmental factors, how do you divide the entire globe into equal, comparable units for analysis?   For data scientists and geospatial analysts, these are fundamental challenges. The solution lies in a powerful class of tools called Discrete Global Grid Systems (DGGS). These systems provide a consistent framework for partitioning the Earth's surface into a hierarchy of cells, each with a unique identifier. The most well-known systems, Google's S2 and Uber's H3, have become industry standards for everything from database optimization to logistics.   However, these systems come with inherent trade-offs. Now, a new DGGS called A5 has been developed to solve some of the critical limitations of its predecessors, particularly concerning area distortion and analytical accuracy.   Why Gridding the Globe is Harder Than It Looks The core mathematical challenge of any DGGS is simple to state but difficult to solve: it is impossible to perfectly flatten a sphere onto a 2D grid without introducing some form of distortion. Think of trying to apply a perfect chessboard or honeycomb pattern to the surface of a ball; the shapes will inevitably have to stretch or warp to fit together without gaps.   All DGGS work by starting with a simple 3D shape, a polyhedron, and projecting its flat faces onto the Earth's surface. The choice of this initial shape and the specific projection method used are what determine the system's final characteristics. As a simple analogy, consider which object you’d rather be hit on the head with: a smooth ball or a spiky cube? The ball is a better approximation of a sphere. When you "inflate" a spiky polyhedron to the size of the Earth, the regions nearest the sharp vertices get stretched out the most, creating the greatest distortion.   A Quick Look at the Incumbents: S2 and H3   To understand what makes A5 different, it's essential to have some context on the most popular existing systems.   Google's S2: The Cube-Based Grid The S2 system is based on projecting a cube onto the sphere. On each face of this conceptual cube, a grid like a chessboard is applied. This approach is relatively simple but introduces significant distortion at the cube’s vertices, or "spikes." As the grid is projected onto the sphere, the cells near these vertices become stretched into diamond shapes instead of remaining square. S2 is widely used under the hood for optimizing geospatial queries in database systems like Google BigQuery.   Uber's H3: The Hexagonal Standard Uber's H3 system starts with an icosahedron—a 20-sided shape made of triangles. Because an icosahedron is a less "spiky" shape than a cube, H3 suffers from far less angular distortion. Its hexagonal cells look more consistent across the globe, making it popular for visualization. H3's immense success is also due to its excellent and user-friendly ecosystem of tools and libraries, making it easy for developers to adopt. However, H3 has one critical limitation for data analysis: it is not an equal-area system. This was a deliberate trade-off, not a flaw; H3 was built by a ride-sharing company trying to match drivers to riders, a use case where exact equal area doesn't particularly matter. To wrap a sphere in hexagons, you must also include exactly 12 pentagons—just like on a soccer ball. If you look closely at a football, you'll see the pentagonal panels are slightly smaller than the hexagonal ones. This same principle causes H3 cells to vary in size. The largest and smallest hexagons at a given resolution can differ in area by a factor of two, meaning that comparing raw counts in different cells is like comparing distances in miles and kilometers without conversion. For example, cells near Buenos Aires are smaller because of their proximity to one of the system's core pentagons, creating a potential source of error if not properly normalized.   Introducing A5: A New System Built for Accuracy A5 is a new DGGS designed from the ground up to prioritize analytical accuracy. It is based on a dodecahedron, a 12-sided shape with pentagonal faces that is, in the words of its creator, "even less spiky" than H3's icosahedron.   The motivation for A5 came from a moment of discovery. Its creator, Felix Palmer, stumbled upon a unique 2D tiling pattern made of irregular pentagons. This led to a key question: could this pattern be extended to cover the entire globe? The answer was yes, and it felt like uncovering something "very, very fundamental." This sense of intellectual curiosity, rather than a narrow business need, is the foundation upon which A5 is built. A5's single most important feature is that it is a true equal-area system. Using a specific mathematical projection, A5 ensures that every single cell at a given resolution level has the exact same area. This guarantee even accounts for the Earth's true shape as a slightly flattened ellipsoid, not a perfect sphere.   This is a game-changer for analysis. By providing cells of identical size, A5 eliminates the need for analysts to perform complex area-based normalization. This prevents a common source of error and dramatically simplifies workflows when calculating metrics like population density, risk exposure, or any other value that depends on a consistent spatial unit.   A5 vs. H3 vs. S2: A Head-to-Head Comparison The choice of base polyhedron and projection method results in significant differences between the major DGGS. Here is a direct comparison of their key technical characteristics. Metric A5 H3 S2 Base Polyhedron Dodecahedron (12 pentagonal faces) Icosahedron (20 triangular faces) Cube (6 square faces) Equal-Area Cells Yes (Exact) No (Up to 2x area variation) No Max Resolution ~30 square millimeters ~1 square meter ~1 square centimeter Global Hierarchy Yes (Single top-level world cell) No (122 top-level cells) Yes (6 top-level cells)   The A5 Ecosystem and its "Polyglot Mirroring" Approach   The success of H3 proves that a powerful mathematical system is not enough; it needs a rich ecosystem of accessible tools to gain adoption. A5 is being built with this principle in mind, but with a novel development strategy.   This approach is called "polyglot mirroring." Instead of building a single core library in C and creating language bindings, A5 maintains separate, complete, and equivalent codebases in multiple languages, including TypeScript, Python, and Rust. To keep these distinct codebases synchronized, Large Language Models (LLMs) are used to port changes and new features from one language to another. This strategy makes the system more accessible and maintainable for developers within each language's native community.   The power of this approach was proven in a true "wow moment" during A5's development. The creator, having never written a single line of Rust, fed the existing TypeScript and Python versions and a comprehensive test suite to an LLM. After about a week of guided iteration, the model produced a complete, working, high-performance Rust library. This demonstrates how modern tools can enable a single developer to build and maintain a truly multi-lingual ecosystem, something that would have been impossible just a few years ago.   Conclusion: When Should You Choose A5?   A5 offers a powerful and precise alternative to existing global grid systems. Its primary advantages make it the ideal choice for specific, demanding use cases.   • Statistical Validity: Any analysis where equal-area cells are paramount for accuracy is a prime candidate for A5. This includes density mapping, demographic studies, environmental modeling, and financial risk assessment. • Extreme Resolution: For applications requiring precision beyond what H3 or S2 can offer, A5's ability to index down to cells of approximately 30 square millimeters provides unmatched granularity. • Efficient Global Hierarchy: Workflows that need to query data at a global scale benefit from A5's simple hierarchy, which starts from a single cell representing the entire world. In contrast, loading global data with H3's 122 top-level cells could require 122 separate requests, creating unnecessary complexity and inefficiency. To explore the A5 system, see detailed visualizations, and understand the technical comparisons in more depth, visit the official website at a5geo.org.
The Open-Source Conundrum   Many successful open-source projects begin with passion, but the path from a community-driven tool to a sustainable business is often a trap.   The most common route—relying on high-value consulting contracts—can paradoxically lead to operational chaos. Instead of a "feast or famine" cycle, many companies find themselves with more than enough work, but this success comes at a cost: a fragmented codebase, an exhausted team, and a growing disconnect from the core open-source community.   This episode deconstructs a proven playbook for escaping this trap: the strategic transition from a service-based consultancy to a product-led company.   Through the story of Jeroen Ticheler and his company, GeoCat, we will analyze how this pivot creates a more stable business, a healthier open-source community, and ultimately, a better product for everyone.
Open-source software is often described as "free," a cornerstone of the modern digital world available for anyone to download, use, and modify. But this perception of "free" masks a growing and invisible cost—not one paid in dollars, but in the finite attention, time, and mounting pressure placed on the volunteer and community maintainers.   This hidden tax is most acute when it comes to security.   Jody from Geocat, a long-time contributor to the popular GeoServer project, pulled back the curtain on the immense strain that security vulnerabilities place on the open-source ecosystem.   His experiences reveal critical lessons for anyone who builds, uses, or relies on open-source software.
What if communities could map their own worlds using low-cost drones and open AI models instead of waiting for expensive satellite imagery? In this episode with Leen from HOT (Humanitarian OpenStreetMap Team), we explore how they're putting open mapping tools directly into communities' hands—from $500 drones that fly in parallel to create high-resolution imagery across massive areas, to predictive models that speed up feature extraction without replacing human judgment. Key topics: Why local knowledge beats perfect accuracy The drone tasking system: how multiple pilots map 80+ square kilometers simultaneously AI-assisted mapping with humans in the loop at every step Localizing AI models so they actually understand what buildings in Chad or Papua New Guinea look like The platform approach: plugging in models for trees, roads, rooftop material, waste detection, whatever communities need The tension between speed and OpenStreetMap's principles Why mapping is ultimately a power game—and who decides what's on the map
This conversation with Jed Sundwall, Executive Director of Radiant Earth, starts with a simple but crucial distinction: the difference between data and data products. And that distinction matters more than you might think. We dig into why so many open data portals feel like someone just threw up a bunch of files and called it a day. Sure, the data's technically "open," but is it actually useful? Jed argues we need to be way more precise with our language and intentional about what we're building. A data product has documentation, clear licensing, consistent formatting, customer support, and most importantly - it'll actually be there tomorrow. From there, we explore Source Cooperative, which Jed describes as "object storage for people who should never log into a cloud console." It's designed to be invisible infrastructure - the kind you take for granted because it just works. We talk about cloud native concepts, why object storage matters, and what it really means to think like a product manager when publishing data. The conversation also touches on sustainability - both the financial kind (how do you keep data products alive for 50 years?) and the cultural kind (why do we need organizations designed for the 21st century, not the 20th?). Jed introduces this idea of "gazelles" - smaller, lighter-weight institutions that can move together and actually get things done. We wrap up talking about why shared understanding matters more than ever, and why making data easier to access and use might be one of the most important things we can do right now.
Reflections from the FOSS4G 2025 conference    Processing, Analysis, and Infrastructure (FOSS4G is Critical Infrastructure) The high volume of talks on extracting meaning from geospatial data—including Python workflows, data pipelines, and automation at scale—reinforced the idea that FOSS4G represents critical infrastructure. AI Dominance: AI took up a lot of space at the conference. I was particularly interested in practical, near-term impact talks like AI assisted coding and how AI large language models can enhance geospatial workflows in QGIS. Typically, AI discussions focus on big data and earth observation, but these topics touch a larger audience. I sometimes wonder if adding "AI" to a title is now like adding a health warning: "Caution, a machine did this". Python Still Rules (But Rust is Chatting): Python remains the pervasive, default geospatial language. However, there was chatter about Rust. One person suggested rewriting QGIS in Rust might make it easier to attract new developers. Data Infrastructure, Formats, and Visualization When geospatial people meet, data infrastructure—the "plumbing" of how data is stored, organized, and accessed—always dominates. Cloud Native Won: Cloud native architecture captured all the attention. When thinking about formats, we are moving away from files on disk toward objects in storage and streaming subsets of data. Key cloud-native formats covered included COGs (Cloud Optimized GeoTIFFs), Zarr, GeoParquet, and PMTiles. A key takeaway was the need to choose a format that best suits the use case, defined by who will read the file and what they will use the data for, rather than focusing solely on writing it. The Spatial Temporal Asset Catalog (STAC) "stole the show" as data infrastructure, and DuckDB was frequently mentioned. Visualization is moving beyond interactive maps and toward "interactive experiences". There were also several presentations on Discrete Global Grid Systems (DGGS). Standards and Community Action Standards Matter: Standards are often "really boring," but they are incredibly important for interoperability and reaping the benefits of network effects. The focus was largely on OGC APIs replacing legacy APIs like WMS and WFS (making it hard not to mention PyGeoAPI). Community Empowerment: Many stories focused on community-led projects solving real-world problems. This represents a shift away from expert-driven projects toward community action supported by experts. Many used OSM (OpenStreetMap) as critical data infrastructure, highlighting the need for locals to fill in large empty chunks of the map. High-Level Takeaways for the Future If I had to offer quick guidance based on the conference, it would be: Learn Python. AI coding is constantly improving and worth thinking about. Start thinking about maps as experiences. Embrace the Cloud and understand cloud-native formats. Standards matter. AI is production-ready and will be an increasingly useful interface to analysis. Reflections: What Was Missing? The conference was brilliant, but a few areas felt underrepresented: Sustainable Funding Models: I missed a focus on how organizations can rethink their business models to maintain FOSS4G as critical infrastructure without maintainers feeling their time is an arbitrage opportunity. Niche Products: I would have liked more stories about side hustles and niche SAS products people were building, although I was glad to see the "Build the Thing" product workshop on the schedule. Natural Language Interface: Given the impact natural language is having on how we interact with maps and geo-data, I was surprised there wasn't more dedicated discussion around it. I believe it will be a dominant way we interact with the digital world. Art and Creativity: Beyond cartography and design talks, I was surprised how few talks focused on creative passion projects built purely for the joy of creation, not necessarily tied to making a part of something bigger.
Karl returns to the Mapscaping podcast to discuss his latest venture, Tyche Insights - a platform aimed at building a global community of geospatial storytellers working with open data. In this conversation, we explore the evolution from his previous company, Building Footprint USA (acquired by Lightbox), to this new mission of democratizing public data storytelling. Karl walks us through the challenges and opportunities of open data, the importance of unbiased storytelling, and how geospatial professionals can apply their skills to analyze and share insights about their own communities. Karl shares his vision for creating something akin to Wikipedia, but for civic data stories - complete with style guides, editorial processes, and community collaboration.   Featured Links Tyche Insights: Main website: https://tycheinsights.com Wiki platform: https://wiki.tycheinsights.com Example project: https://albanydatastories.com Mentioned in Episode: USAFacts: https://usafacts.org QField Partner Program: https://qfield.org/partner Open Data Watch: (monitoring global open data policies)
AI Slop: An Experiment in Discovery Solo Episode Reflection: I'm back behind the mic after about a year-long break. Producing this podcast takes more time than you might imagine, and I was pretty burnt out. The last year brought some major life events, including moving my family back to New Zealand from Denmark, dealing with depression, burying my father, starting a new business with my wife, and having a teenage daughter in the house. These events took up a lot of space. The Catalyst for Return: Eventually, you figure out how to deal with grief, stop mourning the way things were, and focus on the way things could be. When this space opened up in my life, AI came into the picture. AI got me excited about ideas again because for the first time, I could just build things myself without needing to pitch ideas or spend limited financial resources. On "AI Slop": I understand why some content is called "slop," but for those of us who see AI as a tool, I don't think the term is helpful. We don't refer to our first clumsy experiments with other technologies—like our first map or first lines of code—as slop. I believe that if we want to encourage curiosity and experimentation, calling the results of people trying to discover what's possible "slop" isn't going to help.   My AI Experimentation Journey My goal in sharing these experiments is to encourage you to go out and try AI yourself. Phase 1: SEO and Content Generation My experimentation began with generating SEO-style articles as a marketing tool. As a dyslexic person, I previously paid freelancers thousands of dollars over the years to help create content for my website because it was too difficult or time-consuming for me to create myself. Early Challenges & Learning: My initial SEO content wasn't great, and Google recognized this, which is why those early experiments don't rank in organic search. However, this phase taught me about context windows, the importance of prompting (prompt engineering), and which models and tools to use for specific tasks. Automation and Agents: I played around with automation platforms like Zapier, make.com, and n8n. I built custom agents, starting with Claude projects and custom GPTs. I even experimented with voice agents using platforms like Vappy and 11 Labs. Unexpected GIS Capabilities: During this process, I realized you can ask platforms like ChatGPT to perform GIS-related data conversions (e.g., geojson to KML or shapefile using geopandas), repro data, create buffers around geometries, and even upload a screenshot of a table from a PDF and convert it to a CSV file. While I wouldn't blindly trust an LLM for critical work, it's been interesting to learn where they make mistakes and what I can trust them for. AI as a Sparring Partner: I now use AI regularly to create QGIS plugins and automations. Since I often work remotely as the only GIS person on certain projects, I use AI—specifically talking to ChatGPT via voice on my phone—as a sparring partner to bounce ideas off of and help me solve problems when I get stuck. Multimodal Capabilities: The multimodal nature of Gemini is particularly interesting; if you share your screen while working in QGIS, Gemini can talk you through solving a problem (though you should consider privacy concerns).   The Shift to Single-Serve Map Applications I noticed that the digital landscape was changing rapidly. LLMs were becoming "answer engines," replacing traditional search on Google, which introduced AI Overviews. Since these models no longer distribute traffic to websites like mine the way they used to, I needed a new strategy. The Problem with Informational Content: Informational content on the internet is going to be completely dominated by AI. The Opportunity: Real Data: AI is great at generating content, but if you need actual data—like contours for your specific plot of land in New Zealand—you need real data, not generated data. New Strategy: My new marketing strategy is to create targeted, single-serve map applications and embed them in my website. These applications do one thing and one thing only, using open and valuable data to solve very specific problems. This allows me to rank in organic search because these are problems that LLMs have not yet mastered. Coding with AI: I started by using ChatGPT to code small client-side map applications, then moved to Claude, which is significantly better than OpenAI's models and is still my coding model of choice. Currently, I use Cursor AI as a development environment, swapping between Claude code, OpenAI's Codex, and other models. A Caveat: Using AI for coding can be incredibly frustrating. The quality of the code drops dramatically once it reaches a certain scale. However, even with flaws, it’s a thousand times better and faster than what I could do myself, making my ideas possible. Crucially, I believe that for the vast majority of use cases, mediocre code is good enough.   Success Story: GeoHound After practicing and refining my methods, I decided to build a Chrome extension. Every GIS professional can relate to the pain point of sifting through HTTP calls in the developer tools networking tab to find the URL for a web service to use in QGIS or ArcGIS. The Impossible Idea Made Possible: I had pitched this idea to multiple developers in the past, who were either uninterested or quoted between $10,000 and $15,000 to build it. The AI Result: Using AI, I had a minimum viable Chrome extension—GeoHound—that filtered out common geo web services within 3 hours. It took a few days of intermittent work before it was published to the Chrome and Edge web stores. Current Use: GeoHound has thousands of users (my own statistics suggest closer to or over 3,000 users, compared to the 1,000 shown on the Chrome store). While not perfect, it is clearly good enough, and this was something that was impossible for me just six months ago.   My Point: Now is the Time to Experiment AI is here, and it will lead to profound change. Experimenting with it is vital because it will: Help you develop the skills and knowledge needed to meet the needs of the people you serve. Help you better understand what is hype and what is not, allowing you to decipher which voices to listen to. We are moving from a world where information is ubiquitous to a world where knowledge is ubiquitous. Now is the time to be making sloppy mistakes. Don't let perfection stop you from learning how to make stuff that is going to be good enough. If your work consists of repetitive tasks that follow step-by-step recipes, that's going to be a tough gig going forward. Long-term, there will be new opportunities, but you need to be experimenting now to be in a position to take advantage of them. Resources Mentioned You will find a list of the tools I've been experimenting with in the show notes. Automation: make.com, n8n, Zapier Voice/Agents: 11 Labs, Vappy, custom GPT (MCP servers) Coding Models: Claude (current choice), OpenAI's Codex, ChatGPT Development Environment: Cursor AI LLMs/Multimodal: Gemini (studio.google.com) Browser Extension: GeoHound (for Chrome and Edge) https://chromewebstore.google.com/detail/nooldeimgcodenhncjkjagbmppdinhfe?utm_source=item-share-cb If you build anything interesting with these tools, please let me know! I'd love to hear about your own experiments.
Jonathan Wagner, CEO of Scribble Maps, is back on the podcast, and this time we're talking about Scribble—an AI agent he's built into his platform. Not a chatbot, an agent. There's a difference, and we get into that. https://mapscaping.com/podcast/the-business-of-web-maps/  So far, Scribble has access to 140 tools. It can view your map, select tools, build plugins, fetch data, and handle onboarding and customer education.   But here's the thing—should you care?   I think you should, because we're going to see more and more of these things. And whether you like it or not, for a lot of people, this is going to be the way they interact with geospatial data. I don't think we can put the genie back in the bottle. I personally, I'm not entirely sure I would if I could.   Yeah, sure, there's a lot of uncertainty around what these things can do and how they're going to impact us. I get that. I feel it too. But we can't afford to stick our heads in the sand and pretend like it's not happening.   In this conversation, Jonathan walks through why he built Scribble (spoiler: his wife was expecting and he needed to solve an onboarding problem), the real risks of adding AI to your product, and the technical decisions behind using Gemini over OpenAI. We also talk about privacy concerns, the Model Context Protocol (MCP), and what this all means for the future of GIS.   We touch on the QGIS MCP server, the democratization of mapping tools, and when maps aren't actually the answer. It's an honest look at where we are with AI agents in geospatial, from someone who's actually building one.   https://en.wikipedia.org/wiki/Lojban    https://github.com/jjsantos01/qgis_mcp    How's that?
Mapillary

Mapillary

2025-10-2748:05

Exploring the Evolution and Impact of Mapillary with Ed from Meta.  Topics include Ed's journey with Mapillary, the process of uploading and utilizing street-level imagery, and the integration with OpenStreetMap. Ed talks about the challenges of mapping with various devices, the role of community contributions, and future potentials in mapping technology, such as using neural radiance fields (NeRFs) for creating immersive 3D scenes. The episode provides insights into how Mapillary is advancing geospatial data collection and usage. 00:00 Introduction to the Map Scaping Podcast 00:57 Meet Ed: Product Manager at Meta 02:09 Ed's Journey with Mapillary 03:59 What is Mapillary? 07:00 The Evolution of 360 Cameras 09:20 Uploading Imagery to Mapillary 14:10 Building a 3D Model of the World 19:10 Meta's Use of Map Data 21:24 The Importance of Community in Mapping 24:15 The Importance of Authoritative Data 24:49 Meta's Contributions to Open Source Geo World 25:27 Real-World Applications: Vietnam's B Group 28:16 Innovative Mapping in Detroit 31:38 Future of Mapping: Lidar and Beyond 32:20 Exploring Neural Radiance Fields (NeRFs) 35:40 Challenges and Innovations in Mapping Technology 45:25 Community Contributions and Future Directions 46:37 Closing Remarks and Contact Information   Previous episodes that you might find interesting https://mapscaping.com/podcast/scaling-map-data-generation-using-computer-vision/ https://mapscaping.com/podcast/the-rapid-editor/ https://mapscaping.com/podcast/overture-maps-and-the-daylight-distribution/      
Telematics Data is Reshaping Our Understanding of Road Networks In this episode MIT Professor Hari Balakrishnan explains how Cambridge Mobile Telematics (CMT) is transforming traditional road network analysis by layering dynamic behavioural data onto static map geometries.  Telematics data creates "living maps" that go beyond traditional road geometry and attributes. By collecting movement data from 45 million users through phones and IoT devices, CMT has developed sophisticated models that can: - Generate dynamic risk maps showing crash probability for every road segment globally - Detect infrastructure issues that aren't visible in traditional mapping (like poorly placed bus stops) - Validate and correct map attributes like speed limits and lane connectivity - Differentiate between overpasses and intersections using movement patterns - Create contextual understanding of road segments based on actual usage patterns Particularly interesting for GIS professionals is CMT's approach to data fusion, combining traditional map geometry with temporal movement data to create predictive models. This has practical applications from infrastructure planning to autonomous vehicle navigation, where understanding the cultural context of road usage proves as important as precise geometry. The episode challenges traditional static approaches to road network mapping, suggesting that the future lies in dynamic, behavior-informed spatial data models that can adapt to changing conditions and usage patterns. For anyone working with transportation networks or smart city initiatives, this episode provides valuable insights into how movement data is changing our understanding of road infrastructure and spatial behaviour.   Connect with Hari on LinkedIn! https://www.linkedin.com/in/hari-balakrishnan-0702263/ Cambridge Mobile Telematics https://www.cmtelematics.com/   BTW,  I keep busy creating free mapping tools and publishing them there https://mapscaping.com/map-tools/ swing by and take a look!  
Hivemapper

Hivemapper

2024-12-0551:291

In this week’s episode, I’m thrilled to welcome back Ariel Seidman, founder of HiveMapper. Ariel was my very first podcast guest back in 2019, and HiveMapper has come a long way since then! We explore how HiveMapper has evolved from a drone-based mapping system to a cutting-edge platform collecting street-level data at a global scale. Ariel shares the challenges of scaling large-scale mapping efforts, the pivot to building their own hardware, and the role of blockchain-based incentives in driving adoption. Here are just a few topics we cover: Why HiveMapper shifted focus from drones to street-level mapping. The power of combining hardware and software to solve mapping challenges. How HiveMapper has already mapped 28% of the global road network. The revolutionary edge computing and data filtering techniques driving efficiency. What it takes to compete with industry giants like Google Maps. Whether you're fascinated by the intersection of geospatial technology and innovation or looking for insights into scaling impactful startups, this episode is packed with value. Let me know your thoughts or hit reply if you’d like to discuss the episode!   https://beemaps.com/ Connect with Ariel here https://www.linkedin.com/in/aseidman/   PS I have just finished creating a web-based tool that lets you explore and download OpenStreetMap data, It is a bit different from other tools and I would appreciate some feedback.  https://mapscaping.com/openstreetmap-category-viewer/       
Tracking Elephants

Tracking Elephants

2024-11-0645:361

Tracking elephants in Southern Africa’s Kavango-Zambezi (KAZA) region, the largest transfrontier conservation area in the world. Lead scientist Robin Naidoo from the World Wildlife Fund-US explains the complex, cross-border collaboration required to understand elephant movements across vast landscapes and the role of GNSS.   Connected with Robin  https://www.worldwildlife.org/experts/robin-naidoo   Read more information about this study here https://besjournals.onlinelibrary.wiley.com/doi/10.1111/1365-2664.14746   https://news.mongabay.com/2024/09/jumbo-collaring-effort-reveals-key-elephant-movement-corridors/   Check out https://www.movebank.org/
Today's episode touches on some pretty big topics like Imposter Syndrome, Mentorship, Career Progression, Adaptability and Diversity Today you are going to hear two stories from two very different voices. Two brilliant people who happen to be women in geospatial.    Ta Taneka https://www.linkedin.com/in/ta-taneka/ Mary Murphy https://www.linkedin.com/in/mary-murphy-12319433/   You can check out the GIS Directions Podcast here: https://esriaustralia.com.au/gis-directions-podcast or search for GIS Directions where every you listen to podcasts   Recommended Podcast Episodes Getting where you want to go in your geospatial career Mentorship leadership and career advice Mentorship leadership and career advice      
QField

QField

2024-09-1849:071

In this episode, Marco Bernasconi, co-founder and CEO of OPENGIS.ch, introduces us to QField, an open-source mobile application designed for field data collection in conjunction with QGIS. Marco shares his journey in developing QField and discusses its seamless integration with QGIS, allowing users to capture, survey, and manage geospatial data on various mobile devices. We also discuss the technical aspects of QField, including its user-friendly interface, the ability to connect with external sensors, and the recent introduction of QField Cloud for enhanced data synchronization and management. Marco highlights the application’s diverse use cases, from citizen science initiatives to archaeological documentation and utility inspections, demonstrating its potential to transform data collection processes across various industries. More information on Qfield:  https://qfield.org/ https://qfield.cloud/ Or https://www.opengis.ch/#contact      On a personal note, I have been working as a freelance Geospatial consultant for some time now and one of my projects is slowly winding down, which is why I am looking for new projects to get involved in! If you need expertise in Geospatial consultancy, GIS management or the marketing of geospatial products and services Please reach out! https://www.linkedin.com/in/danielodonohue/ or contact me here info@mapscpaing.com  
Analyst To Engineer

Analyst To Engineer

2024-08-2841:051

This is the story of Priscilla Cole, and what she did when she discovered that her ambitions were bigger than the tools she was using! Connect with Priscilla here! https://www.linkedin.com/in/priscilla-cole-5892549/   Recommended Listening The Way You Talk About Your Skills Is Costing You Money   Geospatial Consulting As A Business And Career   Mid-Life Career Change   Getting Where You Want To Go In Your Career   Applying For A Job, Getting Picked and Negotiating   Mentorship Leadership And Career Advice  
In this episode, I'm joined by Konstantine Klemmer, a researcher at Microsoft, to dive deep into the fascinating world of GeoAI. Konstantine introduces us to Satclip, a cutting-edge model that encodes geographic locations based on satellite images. We discuss how Satclip works, the data it uses, and its potential applications, particularly in low-resource settings and predictive modeling. Whether you're into AI, geography, or just curious about the intersection of these fields, this episode is packed with insights. Key Takeaways: What is Satclip?: Learn about Satclip's location encoding, a neural network that converts geographic coordinates into numerical representations based on satellite images. Data and Training: Understand how Satclip is trained using Sentinel-2 satellite images and how it captures unique geographic features. Applications: Discover how Satclip can be used in low-resource environments, such as on edge devices, and how it enhances other models by providing geographic context. The Future of GeoAI: Explore the potential future directions for Satclip, including more detailed regional models and the integration of multiple data modalities. Connect with Konstantine https://www.linkedin.com/in/konstantinklemmer/ Try Satclip https://github.com/microsoft/satclip   Recommended Listening https://mapscaping.com/podcast/computer-vision-and-geoai/ https://mapscaping.com/podcast/planet-imaging-everything-every-day-almost/        
In this episode, I welcome Jason Gilman, a Principal Software Engineer at Element 84, to explore the exciting world of natural language geocoding. Key Topics Discussed: Introduction to Natural Language Geocoding: Jason explains the concept of natural language geocoding and its significance in converting textual descriptions of locations into precise geographical data. This involves using large language models to interpret a user's natural language input, such as "the coast of Florida south of Miami," and transform it into an accurate polygon that represents that specific area on a map. This process automates and simplifies how users interact with geospatial data, making it more accessible and user-friendly. The Evolution of AI and ML in Geospatial Work: Over the last six months, Jason has shifted focus to AI and machine learning, leveraging large language models to enhance geospatial data processing. Challenges and Solutions: Jason discusses the challenges of interpreting natural language descriptions and the solutions they've implemented, such as using JSON schemas and OpenStreetMap data. Applications and Use Cases: From finding specific datasets to processing geographical queries, the applications of natural language geocoding are vast. Jason shares some real-world examples and potential future uses. Future of Geospatial AIML: Jason touches on the broader implications of geospatial AI and ML, including the potential for natural language geoprocessing and its impact on scientific research and everyday applications. Interesting Insights: The use of large language models can simplify complex geospatial queries, making advanced geospatial analysis accessible to non-experts. Integration of AI and machine learning with traditional geospatial tools opens new avenues for research and application, from environmental monitoring to urban planning. Quotes: "Natural language geocoding is about turning a user's textual description of a place on Earth into a precise polygon." "The combination of vision models and large language models allows us to automate complex tasks that previously required manual effort." Additional Resources: Element 84 Website State of the Map US Conference Talk on YouTube Blog Posts on Natural Language Geocoding Connect with Jason: Visit Element 84's website for more information and contact details. Google "Element 84 Natural Language Geocoding" for additional resources and talks.
loading
Comments (4)

Daniel Wurzbacher

Dan, Seems worthwhile to follow up this episode with one exploring advice to mentees. The field seems to accept plenty of midlife career changers, and those folks often wonder how to ask for help without destroying any chance of building a respectable professional reputation down the road.

Mar 22nd
Reply

🤨

Hands down, this guy's work literally made people's careers!

Mar 9th
Reply

Aliakbar Karamvand

what a career 👏

Nov 8th
Reply

Houman Almassi

👌

Feb 15th
Reply