In this post, we are experimenting with the ability of major Generative AI models to analyze and deduce information from a VFR approach chart. The French aviation authorities (DGAC) publish and regularly maintain VFR approach charts that help general aviation pilots plan and conduct their approaches safely.
Typically, a pilot needs this information before attempting to approach an airfield. It contains useful details about the approach and airfield frequencies, the type of accepted aircraft, field elevation, runway length, pattern altitude, and Visual Reference Points.
Here is an example of a French VFR approach plate for Lannion Côte de Granit Airfield.
In today's experiment, we will try to extract information from this chart using Anthropic's Claude 3.5 sonnet model.
Prior to writing this blog post, we have done a few experiments by uploading the PDF files of the approach charts. This came with mixed results as only the written text is taken into account by the AI. This also applies to OpenAI's Chat GPT 4o.
However, for all visual information, we were more successful at directly uploading PNGs and asking models to respond to queries based on them. We will demonstrate a few cases in this blog post.
Extracting Visual Approach Briefing Information
We asked Claude 3.5 the following question after uploading the LFRO PNG representation of the approach chart:
"Given this approach chart, can you give me a short briefing about the runway's length, orientation, and slopes, frequencies to contact; field elevations, pattern altitude, entry points, and any specifics that would allow me to initiate my approach in day VFR? I intend to approach the airfield from the south."
Claude responded with the following as shown in the picture.
Claude's response was mostly correct. It recognized frequencies, entry points, and runway lengths. However, it failed to recognize the pattern orientation and pattern altitude. It also claimed that the pattern is right-hand for both runways, while only runway 11 has a right-hand pattern. It was not able to conduct reasoning on the visual reference point to use when approaching from the south and gave both choices OL (North) and S (South).
We have conducted the same experiment with GPT4o, also capable of vision. The result is very similar with a few notable differences as shown below.
GPT-4 did recognize different left and right patterns depending on the active runway; however, it was not able to recognize the entry points that Claude was able to. Subtle but important difference.
The Need for Human Supervision
For now, our conclusion is that the models, although capable of extracting information from the approach charts, are not fully capable of providing advisories on their own. There is a need to curate and verify the data, especially in a highly regulated and demanding domain like aviation. With Pilot Briefer, we intend to find a solution to that problem by designing hybrid systems that can get us 80% of the way, combining the "opinions" of different models and curating the data based on official and reliable sources, including human supervision. It is possible to do so since the volume of data is limited.
The current state of the art is promising but does have room for further improvements. Having redundant systems is a great architectural pattern, and it should apply to Generative AI-based models.