Address
33-17, Q Sentral.
2A, Jalan Stesen Sentral 2, Kuala Lumpur Sentral,
50470 Federal Territory of Kuala Lumpur
Contact
+603-2701-3606
[email protected]
Address
33-17, Q Sentral.
2A, Jalan Stesen Sentral 2, Kuala Lumpur Sentral,
50470 Federal Territory of Kuala Lumpur
Contact
+603-2701-3606
[email protected]
Banks and financial services are facing more risks because criminals are using advanced artificial intelligence (AI) technologies. These technologies can create fake audio and video clips that look and sound like real customers or company bosses, tricking security systems. There has been a big increase in these fake (“deepfake”) incidents, especially in the financial technology sector, with a 700% jump from last year.
Another worrying trend is the use of AI by criminals to make smarter malware. This malware can change its own code to keep from being caught by standard security systems. It can steal important information like usernames and passwords, which is a big threat to financial safety.
To deal with these AI threats, banks are also using AI to strengthen their defenses. Banks have been using AI for a while to spot fake transactions by looking for unusual patterns. Now, they’re using even more sophisticated AI to keep up with criminals. For example, Mastercard’s new AI system can check a trillion data points to confirm transactions, which really helps in spotting fraud.
AI is now also used to look at security risks as they happen. For instance, FBD Insurance uses AI software to keep an eye on up to 15,000 IT events every second. This helps them react quickly to any security problems. This move to real-time monitoring is a big improvement over older systems that only noticed threats after they happened.
Although AI can be very helpful, it also comes with new problems. Criminals might try to mess with AI systems by giving them wrong data, which can mess up the AI models and lead to wrong decisions about threats. Banks need to be careful and make sure their AI systems are strong and protected from these tricks.
Despite all the progress with AI, we still need human cybersecurity experts. AI systems, especially those that create things like deepfakes, need people to watch over them and fix any mistakes. Human experts are very important for handling the risks that come with using AI in cybersecurity.
Learn why it’s important to use AI to fight against AI-driven threats in finance. Get the latest updates on AI in cybersecurity, the ongoing challenges of using AI for security, and why human experts are still essential.
Sure, here are three frequently asked questions (FAQs) that could complement the article on AI’s role in cybersecurity for financial services:
Answer: A deepfake is a synthetic media in which a person in an existing image or video is replaced with someone else’s likeness using artificial intelligence technologies. These can include audio and video clips that look and sound very realistic. In the context of financial fraud, deepfakes might mimic the voice or appearance of a trusted individual to trick security systems or people into granting access to sensitive information.
Answer: AI detects fraudulent transactions by analyzing patterns and anomalies in data. Financial institutions use AI systems to review vast amounts of transaction data in real time. The AI is trained to recognize what normal transactions look like and can flag transactions that deviate from these patterns. Advanced AI systems, like the one used by Mastercard, can analyze trillions of data points to accurately identify potentially fraudulent activities.
Answer: Human oversight is crucial because AI systems can sometimes make mistakes or “hallucinate” false outputs. Also, AI models can be tampered with or manipulated by feeding them incorrect data, which can lead to faulty decisions. Human cybersecurity experts are needed to supervise AI operations, verify its findings, and intervene when necessary to correct errors and refine the system’s responses. This combination of AI and human expertise ensures more robust defenses against cyber threats.
Sources Financial Times