Home / Blogs

Google Bard Fails to Answer Satellite Internet Questions

Bard made false, inconsistent statements, overlooked pertinent evidence, and overlooked expertise.

In an earlier post, I asked whether electronically steered antennas (ESAs) would replace parabolic antennas in satellite ground stations. I did some research and concluded that it is likely that they will. Next, I discussed the same question with ChatGPT and, while it made several false statements, it made a relevant point that I had overlooked. The relevant addition was positive, but the errors were troublesome, so I decided to try ChatGPT’s competitor Google Bard.

Since the P in Chat GPT stands for “pre-trained,” I began by asking Bard if it was more up-to-date than ChatGPT, and it replied, “Bard is able to access the Internet in real-time, while ChatGPT-4 is limited to a dataset that only goes up to late 2021.” It added that it had last been updated on June 7, 2023.

That was encouraging since antenna technology has improved since late 2021 and at least two new products incorporating ESAs have been announced since that time, BlueHalo’s mobile ground station and Thinkom’s gateway array.

But my optimism faded when I began by asking Bard to “list the advantages of electronically steered antennas over parabolic antennas in ground stations for LEO Internet service constellations” and then asked it to “list the advantages of parabolic antennas over electronically steered antennas.” The results of the two queries are shown here.

The replies are less complete and specific than those given by ChatGPT, and they are inconsistent—Bard credits both ESAs and parabolic antennas as being cheaper. Furthermore, there was no indication that Bard was aware of the BlueHalo or Thinkom products, so I asked three leading questions:

  • Have there been recent satellite ground station innovations?
  • Any new hardware?
  • Any innovative new products?

It did not mention either BlueHalo or Thinkom, but after the third try, it suggested that I try a Google search for “Innovative new products for satellite ground station,” and that returned many links, including one to the Thinkom array of antennas.

My next query demonstrated an extreme lack of awareness. I asked if China’s Long March 5B rocket would be used to launch their GuoWang satellite constellation. Bard answered that it would and added that it “will be the first time that a country has launched such a large constellation of satellites.”

Evidently, GuoWang is more frequently associated with Chinese propaganda than the SpaceX Starlink constellation in the Bard training set. (Garbage in, garbage out).

Red and blue text indicates errors (source)

A final example illustrates Bard’s inability to recognize expertise. I asked Bard, “Who is Larry Press” and it answered with the error-ridden professional biography shown here. The red text highlights statements that are false, and the blue highlights work done in collaboration with others, not by me alone.

One can understand some of the errors. For example, I got my Ph.D. from UCLA, not USC, but I did teach at USC for a while, and that may be more frequently mentioned in the training set. Similarly, I have been on several editorial boards, but I no longer am. I am an expert on my own biography, and a simple Google Search would have found a somewhat dated but accurate short biography on the Web.

I was also disappointed in the promise that Bard would provide links to references. It only cites references when it “directly quotes at length from a webpage.” I only saw one reference during my experimentation, and it was irrelevant.

“Artificial intelligence” programs have succeeded in many specific tasks over the years. As an undergraduate, I had the privilege of taking a course from Herbert Simon and learned about his chess-playing and theorem-proving programs. Arthur Samuel’s checker playing program learned to beat him, and expert systems assist doctors in making some diagnoses. I even wrote an interactive program based on concept acquisition models that assisted researchers in multivariate data analysis. (It could be an Internet service today).

Artificial neural nets have also had success in specific tasks from recognizing zip codes on envelopes to playing Go and Chess and helping robots do back flips, and they will succeed at other tasks like helping Microsoft Windows users or writing poetry. (I submitted haikus written by Bard and ChatGPT to the AI detector GPTZero, and it reported that they were “likely to have been written entirely by a human”).

But none of these programs are generally intelligent. The illusion of intelligence results from the ability to generate responsive grammatical sentences in a conversation, but that is a low bar. Many years ago, I placed Teletype terminals in a public library, and people conversed, sometimes for hours, with ELIZA, a simple simulation of a non-directive therapist. The common practice of referring to errors as “hallucinations” furthers the illusion of intelligence.

Update Sep 27, 2023:

I could not recall the verb “scrub” for navigating through a timeline, so I asked ChatGPT and Bard for help.

ChatGPT got it;
https://chat.openai.com/c/6896ce83-d324-4a04-bb99-357c0da90ddf

Bard blew it:
https://bard.google.com/chat/525ec7146864588a

NORDVPN DISCOUNT - CircleID x NordVPN
Get NordVPN  [74% +3 extra months, from $2.99/month]
By Larry Press, Professor of Information Systems at California State University

He has been on the faculties of the University of Lund, Sweden and the University of Southern California, and worked for IBM and the System Development Corporation. Larry maintains a blog on Internet applications and implications at cis471.blogspot.com and follows Cuban Internet development at laredcubana.blogspot.com.

Visit Page

Filed Under

Comments

The AI Bubble John Poole  –  Jul 26, 2023 4:56 PM

Thanks for sharing your personal experiences with Bard and ChatGPT.  OpenAI recently told a federal court in the state of Georgia, that a radio
talk show host can’t sue OpenAI for defaming him to a reporter since it is widely known that its ChatGPT chatbot provides false information and the platform is incapable of actual malice.
https://www.law360.com/media/articles/1703569/openai-says-chatgpt-couldn-t-defame-talk-radio-host
Caveat Emptor!

Comment Title:

  Notify me of follow-up comments

We encourage you to post comments and engage in discussions that advance this post through relevant opinion, anecdotes, links and data. If you see a comment that you believe is irrelevant or inappropriate, you can report it using the link at the end of each comment. Views expressed in the comments do not represent those of CircleID. For more information on our comment policy, see Codes of Conduct.

CircleID Newsletter The Weekly Wrap

More and more professionals are choosing to publish critical posts on CircleID from all corners of the Internet industry. If you find it hard to keep up daily, consider subscribing to our weekly digest. We will provide you a convenient summary report once a week sent directly to your inbox. It's a quick and easy read.

Related

Topics

Threat Intelligence

Sponsored byWhoisXML API

DNS

Sponsored byDNIB.com

Domain Names

Sponsored byVerisign

Brand Protection

Sponsored byCSC

Cybersecurity

Sponsored byVerisign

IPv4 Markets

Sponsored byIPv4.Global

New TLDs

Sponsored byRadix