Last weekend I tried out Google's latest AI model (Gemini, the AI formally known as Bard) to see how much of an improvement it makes over the woeful Google Assistant, and this gives an interesting insight into automation v manual work.
Bear in mind that Google is the 6th largest company on the planet, and search is their baby - there should be no better company at it.
The question was 'what time is Gladiators on tonight?' (great to watch with the kids BTW!), it should have taken the context of the question as I'm in the UK, looked up a result, parsed and delivered it back.
It delivered with absolute confidence that Gladiators had finished, and that I can watch it on catch up (error #1), so I stated it was wrong, giving no more context, it then replied with confidence that it was actually on TV and would be shown after the Six Nations (error #2), as that wasn't on last weekend. I stated it was wrong again, and it then went and found the right answer.
This experience raises two noteworthy points:
1. The fact that Google, a tech giant, struggled with such a basic task raises concerns about the prioritization of speed over quality in the AI development race.
2. Effectively harnessing AI requires users to guide it accurately, emphasising the critical role of human interaction.
In my world of IT security, where constant innovation in tools and attack methodologies occurs, AI is often hailed as either the ultimate threat or the ultimate savior. I call BS on both notions. While technology can empower both malicious and benevolent actors, the essence of security remains inherently human.
Every day, we use AI to help build sophisticated cyber attacks against our customers so that we can outpace the real attackers - but to do so, we have a team of experts that build the human nuance into each attack and take the specific context of the client to make the attack effective.
Don't trust either your security or your testing solely to automated tools - you need, and always will, human experts to truly test your security.
If the concern of AI and hacking is keeping you up at night, or you're unsure about your company's security, let's chat.
What other examples have you got of AI failing the simplest of tasks?
#AIsecurity #redteaming #penetrationtesting #artificialintelligence #googlebard #googlegemini #covertswarm
Cybersecurity Professor; Recovering Entrepreneur
1moI remember we were both passionate about machine learning when no one else was ☺️