Connect with us

Business

Building an impactful security training programme to handle the volume and sophistication of today’s AI-enabled cyberattacks

By Alexia Pedersen, SVP International of O’Reilly 

Digital assaults that are underpinned by AI are quickly becoming one of the most predominant issues on the planet, with the National Cyber Security Centre warning that the use of AI for malicious purposes will significantly shape the threat landscape as we know it today. Whether it’s sophisticated phishing emails or deepfake videos, this technology is enabling relatively unskilled threat actors to carry out more effective access and information-gathering operations than ever before.   

On top of this, O’Reilly’s research highlights that nearly a quarter (24%) of learning professionals within British tech companies say cybersecurity is the digital skill most are lacking. As such, the vast majority (88%) of companies plan to spend more than £25,000 in the next twelve months to fill crucial roles, with cybersecurity top of the priority list.  

Ultimately, the dual crisis of AI-enabled threats and a widening skills gap is not one that companies can hire their way out of. So, how can organisations and their employees keep pace with the sophistication and volume of attacks? And will the EU AI Act help? 

The evolving regulatory landscape

While the EU’s AI Act is a significant step forward in regulating AI to ensure its safe and ethical development, there is a long way to go before we can secure our digital future.

Today, the Act focuses on security, transparency, and accountability to mitigate the risks associated with AI. By imposing stringent security requirements on high-risk AI systems – like those used in critical infrastructure – the Act ensures these systems are designed to be accurate, robust, and secure against unauthorised access and manipulation. It also requires these systems to have robust cybersecurity measures in place, including regular security assessments, vulnerability management, and incident response plans. 

Furthermore, the Act mandates transparency in the development and deployment of AI systems – providing clear information about the system’s capabilities, limitations, and potential risks. Meanwhile, companies developing and deploying high-risk AI systems will be held accountable for any harm caused by their systems. This creates a strong incentive for organisations to prioritise cybersecurity and ensure the security of their AI systems. 

The AI Act also emphasises the importance of mitigating bias and discrimination in AI systems. This includes ensuring that AI systems are trained on diverse and representative data to avoid unfair outcomes.  By promoting fairness and non-discrimination, the AI Act indirectly contributes to a more secure digital environment.

As the regulatory environment continues to evolve, organisations have a responsibility to educate their staff on the ever-evolving risks posed by AI-enabled cyberattacks. We recommend keeping the following key steps in mind for building an impactful, AI-related security training programme.

  1. Identify the key stakeholders that can drive the programme forward

Firstly, it’s deciding who should take charge. Ideally, the leadership of your programme should be a collaborative effort between IT and those responsible for learning and development. With IT specialists providing the technical expertise, ensuring the content is relevant and appropriately complex, while learning professionals contribute their knowledge of learning strategies, programme design, and evaluation to ensure effective delivery.

However, given the complexities of today’s threat landscape, it’s important that leadership is also involved to align the programme with the organisation’s strategic goals. Emerging roles like Digital Transformation Leaders and Chief AI Officers, are becoming increasingly critical stakeholders and involving them in this process will help support change management as a new initiative gets rolled out.

  1. Align your unique organisational needs with your programme  

The next key step is to assess your organisation’s current needs and skill gaps against future needs. By engaging with all stakeholders, from leadership to employees and IT specialists, organisations will gain a comprehensive understanding of their unique technology landscape. Focus on the relevancy, variety, and flexibility of available high-quality learning content when rolling out a news skills programme. This approach will guarantee the programme addresses current industry trends and incorporates your organisation’s professional IT certifications, while also anticipating future needs. 

  1.  Maximising impact with a blended learning approach 

A blended learning approach is important. After all, your education programme must cater to a variety of learning styles and paces, so a combination of theoretical learning and hands-on practice is important to provide staff with robust and thorough knowledge.

Your programme should therefore integrate a mix of learning channels including digital learning, webinars, workshops, and one-on-one mentorship. Self-paced e-learning modules, for example, will allow for flexibility while scheduled sessions offer real-time interaction. At the same time, workshops, mentoring, and on-the-job practice will offer more opportunities for experiential learning. Ultimately, a mix of content to suit different learning styles and abilities will make the training accessible, engaging and inclusive for all designated participants. 

  1.  Data and insights: Ways to measure success

Once up and running, continuous monitoring and evaluation of skill development will enable you to gauge the effectiveness and make refinements where needed. Success for your training programme can be gauged through various methods, with a key one being regular, technical assessments or certifications to verify the development of skills.

At the same time, you should conduct regular reviews to build a culture of learning, checking in with managers to assess progress and adapt as needed. Longer-term, you should also measure changes in performance metrics post-training, such as the reduction in IT-related errors or increased productivity in assigned tasks. In addition, build engagement plans and activities to maintain this momentum. This combination will allow you to improve the programme in real time and address your employees’ dynamic learning needs. 

Looking ahead, business leaders need to put adequate investment behind the development of education programmes that educate staff on the risks posed by AI-enabled cyberattacks. This should be driven by IT and learning professionals, given the combination of their indispensable expertise will maximise effectiveness. 

Both stakeholders must spend time pinpointing a diverse range of employees to drive forward their training programme, as well as identifying their company’s unique operational needs to ensure training is tailored and highly relevant. As an example, in Q2 2024, Check Point Research reported a 30% YoY increase in cyberattacks globally, reaching over 1,600 attacks per organisation per week. As AI initiatives continue to expand, awareness and skills in cybersecurity will be essential.

Whether you are developing AI solutions in-house, purchasing third-party technology with embedded AI, or partnering with AI tools, it’s critical to have a plan in place and implement comprehensive security training across the organisation. Only when armed with this foundational knowledge will learning professionals and IT leaders be empowered to identify the most suitable L&D partner that can support their unique needs and objectives. 

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Business

How the BPO sector is tackling the surge in fraud across US banking

Source: Finance Derivative

Hans Zachar, Group Chief Information Officer at Nutun

Fraud in the U.S. banking industry is on the rise, driven by the rapid shift towards digital banking by traditional banks coupled with the emergence of neobanks. This trend is not only increasing costs, but also eroding consumer trust and negatively impacting customer experience (CX). According to the latest annual LexisNexis® True Cost of Fraud™ Study: Financial Services and Lending Report — U.S. and Canada Edition, 63% of financial firms reported a fraud increase of at least 6% over the past year, with digital channels contributing to half of all fraud losses.

The study also highlighted the steep financial toll, revealing that for every dollar lost to fraud, North American financial institutions incur $4.41 in total costs. U.S. investment firms and credit lenders have seen the financial impact of fraud rise by 9% year-over-year. Alarmingly, 79% of respondents noted that fraud has also made it harder to earn consumer trust.

The fraudster’s playbook

With the wealth of personal customer data out there, fraudsters are becoming more adept at breaching security verification checks. For example, with customer data showing up in multiple breaches, fraudsters can collate data across sources to build a more complete picture of a person, placing them in a better position to answer knowledge-based authentication questions, often better than the individual.

Despite the increased awareness, there has been a recent shift in modus operandi where criminals impersonate the fraud department from a customer’s bank, asking them to share their one-time pin (OTP). They know your name, address, and credit card digits, and generate an SMS from the bank to get the OTP. With this information, they can access a customer’s account and engage in account origination and transactional fraud.

The situation is worse than ever, with the TransUnion State of Omnichannel Fraud Report for H2 2024 indicating that the sector experienced $3.2 billion in lender exposure to suspected synthetic identities for U.S. auto loans, credit cards, retail credit cards and personal loans at the end of June 2024, which was the highest level ever recorded.

How technology is reshaping fraud landscapes

Technology is aiding and abetting criminals, with artificial intelligence (AI) increasingly used to circumvent multi-factor authentication (MFA). For instance, fraudsters now create deepfakes across voice and video channels to pass biometric authentication. The 2023 Sumsub Identity Fraud Report, revealed a 10-fold increase in the number of deepfakes detected globally across all industries from 2022 to 2023, with a staggering 1740% deepfake surge in North America. The report identified AI-powered fraud, money-muling networks, fake IDs, account takeovers and forced verification as the top risks.

In this regard, Deloitte’s Center for Financial Services predicts that GenAI could enable fraud losses to reach $40 billion in the United States by 2027, up from $12.3 billion in 2023, representing a compound annual growth rate of 32%.

In response, banking institutions are combining a risk-based and data-driven approach to fraud management, leveraging the capabilities of cutting-edge technologies like AI, machine learning (ML) and biometric and behavior-based authentication methods. However, banks need to balance the cost of implementing more effective and stringent fraud risk mitigation and management without compromising customer service and CX. In this regard, many banks are investing in advanced technologies to monitor transactions in real-time and leverage more sophisticated processes to better understand risks at an individual transaction level on an account by better understanding flow and originating IP addresses.

With these insights, the bank can decide what to do with a transaction, either validating it, sending an automated SMS to confirm the action, or diverting the transaction to a customer call or contact center for authentication.

However, despite the technology that banks have in place, the volumes are causing backlogs in the contact centers, which is affecting CX and creating friction in the customer journey. Banks need the capabilities to interact with customers in more efficient and cost-effective ways to tackle the full volume of potentially fraudulent transactions. For these reasons, many banks and lenders are turning to the global Business Processing Outsourcing (BPO) sector to tap into readily available CX and security skills, expertise and technological capabilities.

The importance of BPO banking for financial institutions in the digital era

Banks need a BPO provider that not only has a comprehensive understanding of the financial sector, but also effectively manages costs by utilising the most efficient and budget-friendly methods to engage with customers, focusing on text and voice interactions. After a fraudulent transaction has occurred, banks require a robust system for managing disputes and supporting backend investigations. Banks must track transactions across different regions and time zones since there is no interbank switch available for fraud detection, often relying on human resources to compile transaction details and provide feedback to distressed customers.

To provide compassionate and empathetic support after a fraud case, it is essential to have well-trained agents equipped with real-time information who can guide affected customers through the entire process. A poor experience or a lack of care can significantly impact customer retention rates. However, establishing these capabilities and developing agent expertise within in-house contact centers can be expensive, especially as fraud incidents continue to rise.

Banks that discover a global BPO provider possessing a powerful combination of fraud detection technology, omnichannel engagement features, trained and experienced agents, and fraud investigators will gain significant advantages such as continuous monitoring and industry leading issue resolution. This approach achieves an equal balance between cost-effective and efficient fraud mitigation with high-quality customer service, while adhering to stringent data privacy and regulatory standards.

Continue Reading

Business

Using technology to safeguard against fraud this holiday season

Source: Finance Derivative

Tristan Prince, Product Director, Fraud & Financial Crime, Experian

The holiday season brings with it a surge in consumer spending, with UK shoppers expected to part with an impressive £28 billion this year. Unfortunately, this increased activity also draws the attention of cybercriminals looking to exploit vulnerabilities in security systems and personal data.

For financial institutions, the stakes have never been higher. With identity fraud on the rise and new regulations from the Payment Systems Regulator, there is a pressing need to ramp up fraud prevention measures. This season, businesses must leverage innovative technologies to protect their customers and ensure a safe shopping experience.

Fraud is on the rise

In recent years, the prevalence of fraud has reached new levels. Identity fraud alone has seen a 21% increase during the holiday season since 2021, with last year’s figures showing that 83% of all fraud cases were identity-related.

This alarming trend continues in 2024, with a 12.5% increase in identity fraud cases recorded in just the first half of the year. These statistics highlight a troubling reality: fraud is evolving, becoming more sophisticated and harder to detect.

Technology: the key to fighting fraud

Despite these challenges, financial institutions are not powerless. Advanced technology is playing a pivotal role in strengthening defences against fraud. From artificial intelligence (AI) to collaborative data networks, companies now have powerful tools at their disposal to outwit even the most determined criminals.

Artificial intelligence: a game-changer

AI has emerged as a cornerstone in modern fraud prevention strategies. By analyzing massive datasets in real time, AI can quickly identify unusual activity and potential fraud.

Here’s how AI is reshaping fraud detection:

  • Real-time monitoring
    AI systems continuously monitor transactions, instantly identifying irregular patterns that could indicate fraud. This allows institutions to intervene before any damage is done.
  • Behavioral insights
    By examining customer behaviour, AI can detect deviations from typical spending habits, such as unexpected purchases or login attempts from unusual locations. These insights not only help prevent fraud but also improve the experience for legitimate customers by reducing unnecessary disruptions.
  • Strengthened identity checks
    AI-powered tools verify customer identities by cross-referencing data from various sources, ensuring transactions are carried out by the right individuals while minimizing delays.

Data sharing: strength in unity

In addition to AI, collaborative data sharing between financial institutions is proving to be a powerful weapon against fraud. By pooling insights on fraudulent activities and suspicious trends, companies can create a unified front to tackle threats more effectively.

The benefits of data collaboration:

  • Broader visibility: Sharing information helps institutions detect fraud patterns that might otherwise go unnoticed within their own systems.
  • Faster action: Real-time data exchange ensures that when one company flags a suspicious transaction, others can respond immediately, preventing further attacks.

Holiday security: a shared responsibility

The fight against fraud is a continuous battle. Although technology has made significant inroads in preventing financial crime, fraudsters are constantly refining their methods. This requires financial institutions to remain agile and invest in the latest innovations.

Encouragingly, advancements in fraud prevention are already yielding results. For example, the financial services sector successfully blocked £710 million worth of unauthorized fraud in the first half of 2024, thanks to cutting-edge solutions like AI and data-sharing networks.

Making the holidays safe for everyone

As the festive season gets underway, businesses must prioritize the safety of their customers. Through strategic use of technology, financial institutions can outpace fraudsters and protect consumers during one of the busiest shopping periods of the year.

By embracing innovation, fostering collaboration, and maintaining vigilance, companies can ensure that shoppers feel secure, and the spirit of the season remains intact. Together, we can make this festive season safer for everyone.

Continue Reading

Business

The Evolution of AI in Trading: Building Smarter Partnerships Between Humans and Machines

In these uncertain times where what we are seeing is increasing and perhaps most importantly , unprecedented volatility in the financial markets, it is no surprise that the integration of AI in trading has become a focal point of industry discussion. Today, we’re witnessing a fundamental shift in how traders approach markets against the backdrop of an exponential growth in data complexity.

You get a sense that it’s the same story on trading desks worldwide. One can not deny that the sheer volume and velocity of market-moving information has now surpassed human cognitive capacity. All this means is that we’re at a critical inflection point.

If you look back, it’s clear that ever since the first algorithmic trading systems took seed, we’ve been moving toward this moment. But as with most things in financial technology, the reality is somewhat more nuanced.

The Reality of Real-Time Analysis

Initially, many believed AI would simply replace human traders. But yet perhaps what we need here is some perspective. It is my view that we can expect AI to augment rather than replace human decision-making in trading. Think of it like this – in this scenario, machines will help handle the heavy lifting of data processing and analysis while traders focus on final strategy.

Now, there’s a reason why leading trading houses are investing heavily in AI capabilities and it is simply because successful trading will increasingly rely on human-AI partnerships. At least that’s what our experience with the major trading institutions we work with indicates.

Risk Management in the AI Era

Let’s briefly look at risk management and AI’s capacity for processing vast amounts of market data is nothing short of remarkable. What we’ve found using our own systems in-house is that risk management becomes more proactive when powered by AI. Again and again, we have been seeing how machine learning models can identify potential risks before they materialise, helping a trader to make better trading decisions and spotting new opportunities which may otherwise not have surfaced.

So there it is. The keys to effective risk management lie in combining AI’s processing power with human judgment. And the good news is despite these technological advancements, it can not be overstated just how important human experience remains.

The Evolution of The Human-AI Partnership

In this light, as long as we rely on markets driven by human behaviour, we’ll need human insight. And so, defining what is classed as effective AI integration is becoming vital, as is helping traders to understand both AI’s capabilities and limitations.

From our point of view it has been fascinating to witness the different reactions to embedding AI capabilities in trading – from keen early-adopters willing to take a chance on something new all the way down to dinosaurs prefer to rely on traditional methods and will inevitably be left behind as the race for AI supremacy intensifies.

Increasingly, we’re seeing successful traders embrace AI as a partner rather than a replacement. At the end of the day, markets are complex adaptive systems and those who will win will be those who use AI to enhance human decision-making.

As for the future, one cannot argue against the fact that AI will play an increasingly important role in trading. Even that feels like an understatement.  Everywhere you look, trading firms are investing in AI capabilities – some far more quickly and deeply than others – and it’s without a doubt that this trend will continue exponentially.

Author Bio

Wilson Chan is the Founder of Permutable AI, a London-based fintech pioneering AI solutions for financial markets. With roots at Merrill Lynch and Bank of America, he bridges institutional trading expertise with cutting-edge technology. Their latest innovation, the Trading Co-Pilot, delivers real-time event-driven insights for traders, combining geopolitical, macroeconomic, and supply-side data.

Continue Reading

Copyright © 2021 Futures Parity.