As strides in artificial intelligence continue to reshape virtually every industry, it’s no surprise that more and more construction, engineering, and inspection stakeholders are curious about what impact AI may have on their industries. In this article, we’ll zero in on three technologies with the most potential to impact construction, then touch on what’s preventing AI from taking over the most tedious tasks on your project monitoring to-do list.
There are three main AI technologies experts believe will impact construction-related work most significantly in the coming years. They are large language models, neural radiance fields, and recognition.
LLMs—recently made infamous by ChatGPT—are powerful engines that rapidly synthesize data and answer questions. Tools like these are also quite adept at combining structured and unstructured information. For example, in construction, LLM engines can quickly generate summaries of uploaded field reports.
But LLMs do so much more than automatically summarize manually-gathered notes. These engines have real potential to integrate many different systems, platforms, and data sources. Recently, for example, LLMs have been asked to generate code that asks APIs or other programming interfaces to utilize other models and libraries.
In other words, the construction industry could eventually reap benefits from LLM far beyond text summarization and generation. In time, this type of technology could perform truly complex tasks, such as automatically generating detailed field reports from scratch by reviewing text and visual data, including visual data gathered by NeRF technology.
NeRF technology generates 3D (and even 4D) models of an object, structure, or scene using limited 2D images. What’s more, NeRF can integrate footage over space and time from the exact same perspective, showing users how some aspect or area of a site has changed over time. In some instances, geo-referenced reality capture can be performed using hardware as simple as a smartphone, 360 camera, or drone.
In construction, this reality mapping technology allows stakeholders to turn back the clock and see through walls that were sealed six months—or even six years—ago. Thanks to such technology, an engineer can pinpoint the layout of a series of pipes without the cost, delay, or disruption of demolition using an as-built digital twin. And what’s more, NeRF technology offers measurability, allowing stakeholders the opportunity to remotely visualize a job site and inspect and measure it with total confidence.
NeRF technology, like LLM, also has the potential to integrate with other machine learning tools, including recognition.
Related: How to Choose the Right Reality Capture Tool, Process & Team
Finally, strides in recognition tools are helping machines identify objects, images, and so much more. These technologies can rapidly recognize floors, ceilings, windows, fire extinguishers… virtually anything that appears in an image. And when NeRF and recognition work together, AI can begin recognizing whatever’s been visualized in a 3D model of an airport runway, a record-breaking skyscraper, or even a quick-serve restaurant.
While recognition is powerful in and of itself, it becomes infinitely more helpful when we can use it alongside other machine learning models.
Related: Remote Construction Monitoring: The Key to Building Sweden's Tallest Office Tower
Experts now believe that a new recognition model from Meta called Segment Anything could bring even more change to the construction industry. That’s because this technology makes it easier than ever before to deal with images in terms of objects, not pixels. In the past, when technology recognized images, these tools were actually labeling pixels into different things clustered together into one object. But now, Segment Anything instantly identifies each discrete object in an image.
While this advance may seem minor, it has big implications for the construction industry. That’s because if recognition tools can truly segment and identify objects, they become infinitely more powerful when combined with LLMs and NeRF models. Suddenly, these three technologies can work together, offering stakeholders measurability from NeRF, object clarity from recognition, and LLM that can tap into all that data to count objects and quantify work in progress.
In other words, these technologies could become a key component of AI-powered visual progress monitoring, visual production management, visual quality assurance and quality control, and physical asset inspection. Using this new recognition technology alongside LLM and NeRF, we can not only identify fire extinguishers within a “reality mapped” structure but also quantify how many extinguishers there are and where they’ve been laid out. We can also begin to ask questions such as, “What is the total linear feet of drywall installed on the first floor so far?”
In a perfect world, these three AI tools—when combined with schedules—could help synthesize data and share findings in a more natural, human way, perhaps answering construction risk management questions such as, “Which team is most behind schedule?” And it’s true that one day, LLM may also be able to tap into recognition technology to review NeRF visualizations and answer questions such as, “Which of my parking lots most needs repairs?”
Another dream scenario? Retail and multi-site construction stakeholders could ask LLM, “Show me all my storefront signage,” and then receive a report with pictures of the recognized objects identified from NeRF-generated models.
So why can’t a machine do this quite yet? Why can’t a construction-focused LLM start handling every last detail of your project monitoring today? One of the biggest challenges to integrating AI into everyday construction workflows is that every person wants different information. And beyond that, many of these technologies are only a couple of years old.
To create a frustration-free tool that monitors construction progress and responds to every last question just the way we want it answered, these models and technologies need to be further developed (both separately and together) to interpret, synthesize, and communicate information about our projects in a truly valuable way.
Reconstruct brings the job site directly to stakeholders by generating as-built, precise 2D floor plans and 3D models of a building or structure as built. Within Reconstruct’s Visual Command Center, even the most remote of stakeholders can monitor progress, perform a facility condition assessment, overlay design and BIM against what’s been built, use 4D BIM to visualize schedule, or tap into Reconstruct Project Snapshots to rewind construction to an earlier date or time.