In today’s digital landscape, few businesses can go without relying on cloud computing to build a rock-solid IT infrastructure. Boosted efficiency, reduced expenses, and increased scalability are just some of the reasons behind its increasing popularity.

In case you aren’t familiar with the concept, cloud computing refers to running software and services on the internet using data stored on outside sources. So, instead of owning and maintaining their infrastructure locally and physically, businesses access cloud-based services as needed.

And what is found in the cloud? Well, any crucial business data that you can imagine. Customer information, business applications, data backups, and the list can go on.

Given this data’s sensitivity, cloud computing security is of utmost importance.

Unfortunately, cloud computing isn’t the only aspect that keeps evolving. So do the risks, issues, and challenges threatening its security.

Let’s review the most significant security issues in cloud computing and discuss how to address them adequately.

Understanding Cloud Computing Security Risks

Cloud computing security risks refer to potential vulnerabilities in the system that malicious actors can exploit for their own benefit. Understanding these risks is crucial to selecting the right cloud computing services for your business or deciding if cloud computing is even the way to go.

Data Breaches

A data breach happens when unauthorized individuals access, steal, or publish sensitive information (names, addresses, credit card information). Since these incidents usually occur without the organization’s knowledge, the attackers have ample time to do severe damage.

What do we mean by damage?

Well, in this case, damage can refer to various scenarios. Think everything from using the stolen data for financial fraud to sabotaging the company’s stock price. It all depends on the type of stolen data.

Whatever the case, companies rarely put data breaches behind them without a severely damaged reputation, significant financial loss, or extensive legal consequences.

Data Loss

The business world revolves around data. That’s why attackers target it. And why companies fight so hard to preserve it.

As the name implies, data loss occurs when a company can no longer access its previously stored information.

Sure, malicious attacks are often behind data loss. But this is only one of the causes of this unfortunate event.

The cloud service provider can also accidentally delete your vital data. Physical catastrophes (fires, floods, earthquakes, tornados, explosions) can also have this effect, as can data corruption, software failure, and many other mishaps.

Account Hijacking

Using (or reusing) weak passwords as part of cloud-based infrastructure is basically an open invitation for account hijacking.

Again, the name is pretty self-explanatory – a malicious actor gains complete control over your online accounts. From there, the hijacker can access sensitive data, perform unauthorized actions, and compromise other associated accounts.

Insecure APIs

In cloud computing, communication service providers (CSPs) offer their customers numerous Application Programming Interfaces (APIs). These easy-to-use interfaces allow customers to manage their cloud-based services. But besides being easy to use, some of these APIs can be equally easy to exploit. For this reason, cybercriminals often prey on insecure APIs as their access points for infiltrating the company’s cloud environment.

Denial of Service (DoS) Attacks

Denial of service (DoS) attacks have one goal – to render your network or server inaccessible. They do so by overwhelming them with traffic until they malfunction or crash.

It’s clear that these attacks can cause severe damage to any business. Now imagine what they can do to companies that rely on those online resources to store business-critical data.

Insider Threats

Not all employees will have your company’s best interest at heart, not to mention ex-employees. If these individuals abuse their authorized access, they can wreak havoc on your networks, systems, and data.

Insider threats are more challenging to spot than external attacks. After all, these individuals know your business inside out, positioning them to cause serious damage while staying undetected.

Advanced Persistent Threats (APTs)

With advanced persistent threats (APTs), it’s all about the long game. The intruder will infiltrate your company’s cloud environment and fly under the radar for quite some time. Of course, they’ll use this time to steal sensitive data from your business’s every corner.

Challenges in Cloud Computing Security

Security challenges in cloud computing refer to hurdles your company might hit while implementing cloud computing security.

Shared Responsibility Model

A shared responsibility model is precisely what it sounds like. The responsibility for maintaining security falls on several individuals or entities. In cloud computing, these parties include the CSP and your business (as the CSP’s consumer). Even the slightest misunderstanding concerning the division of these responsibilities can have catastrophic consequences for cloud computing security.

Compliance With Regulations and Standards

Organizations must store their sensitive data according to specific regulations and standards. Some are industry-specific, like HIPAA (Health Insurance Portability and Accountability Act) for guarding healthcare records. Others, like GDPR (General Data Protection Regulation), are more extensive. Achieving this compliance in cloud computing is more challenging since organizations typically don’t control all the layers of their infrastructure.

Data Privacy and Protection

Placing sensitive data in the cloud comes with significant exposure risks (as numerous data breaches in massive companies have demonstrated). Keeping this data private and protected is one of the biggest security challenges in cloud computing.

Lack of Visibility and Control

Once companies move their data to the cloud (located outside their corporate network), they lose some control over it. The same goes for their visibility into their network’s operations. Naturally, since companies can’t fully see or control their cloud-based resources, they sometimes fail to protect them successfully against attacks.

Vendor Lock-In and Interoperability

These security challenges in cloud computing arise when organizations want to move their assets from one CSP to another. This move is often deemed too expensive or complex, forcing the organization to stay put (vendor lock-in). Migrating data between providers can also cause different applications and systems to stop working together correctly, thus hindering their interoperability.

Security of Third-Party Services

Third-party services are often trouble, and cloud computing is no different. These services might have security vulnerabilities allowing unauthorized access to your cloud data and systems.

Issues in Cloud Computing Security

The following factors have proven as major security issues in cloud computing.

Insufficient Identity and Access Management

The larger your business, the harder it gets to establish clearly-defined roles and assign them specific permissions. However, Identity and Access Management (IAM) is vital in cloud computing. Without a comprehensive IAM strategy, a data breach is just waiting to happen.

Inadequate Encryption and Key Management

Encryption is undoubtedly one of the most effective measures for data protection. But only if it’s implemented properly. Using weak keys or failing to rotate, store, and protect them adequately is a one-way ticket to system vulnerabilities.

So, without solid encryption and coherent key management strategies, your cloud computing security can be compromised in no time.

Vulnerabilities in Virtualization Technology

Virtualization (running multiple virtual computers on the hardware elements of a single physical computer) is becoming increasingly popular. Consider the level of flexibility it allows (and at what cost!), and you’ll understand why.

However, like any other technology, virtualization is prone to vulnerabilities. And, as we’ve already established, system vulnerabilities and cloud computing security can’t go hand in hand.

Limited Incident Response Capabilities

Promptly responding to a cloud computing security incident is crucial to minimizing its potential impact on your business. Without a proper incident report strategy, attackers can run rampant within your cloud environment.

Security Concerns in Multi-Tenancy Environments

In a multi-tenancy environment, multiple accounts share the same cloud infrastructure. This means that an attack on one of those accounts (or tenants) can compromise the cloud computing security for all the rest. Keep in mind that this only applies if the CSP doesn’t properly separate the tenants.

Addressing Key Concerns in Cloud Computing Security

Before moving your data to cloud-based services, you must fully comprehend all the security threats that might await. This way, you can implement targeted cloud computing security measures and increase your chances of emerging victorious from a cyberattack.

Here’s how you can address some of the most significant cloud computing security concerns:

  • Implement strong authentication and access controls (introducing multifactor authentication, establishing resource access policies, monitoring user access rights).
  • Ensure data encryption and secure key management (using strong keys, rotating them regularly, and protecting them beyond CSP’s measures).
  • Regularly monitor and audit your cloud environments (combining CSP-provided monitoring information with your cloud-based and on-premises monitoring information for maximum security).
  • Develop a comprehensive incident response plan (relying on the NIST [National Institute of Standards and Technology] or the SANS [SysAdmin, Audit, Network, and Security] framework).
  • Collaborate with cloud service providers to successfully share security responsibilities (coordinating responses to threats and investigating potential threats).

Weathering the Storm in Cloud Computing

Due to the importance of the data they store, cloud-based systems are constantly exposed to security threats. Compare the sheer number of security risks to the number of challenges and issues in addressing them promptly, and you’ll understand why cloud computing security sometimes feels like an uphill battle.

Since these security threats are ever-evolving, staying vigilant, informed, and proactive is the only way to stay on top of your cloud computing security. Pursue education in this field, and you can achieve just that.

Related posts

Wired: Think Twice Before Creating That ChatGPT Action Figure
OPIT - Open Institute of Technology
OPIT - Open Institute of Technology
May 12, 2025 6 min read

Source:

  • Wired, published on May 01st, 2025

People are using ChatGPT’s new image generator to take part in viral social media trends. But using it also puts your privacy at risk—unless you take a few simple steps to protect yourself.

By Kate O’Flaherty

At the start of April, an influx of action figures started appearing on social media sites including LinkedIn and X. Each figure depicted the person who had created it with uncanny accuracy, complete with personalized accessories such as reusable coffee cups, yoga mats, and headphones.

All this is possible because of OpenAI’s new GPT-4o-powered image generator, which supercharges ChatGPT’s ability to edit pictures, render text, and more. OpenAI’s ChatGPT image generator can also create pictures in the style of Japanese animated film company Studio Ghibli—a trend that quickly went viral, too.

The images are fun and easy to make—all you need is a free ChatGPT account and a photo. Yet to create an action figure or Studio Ghibli-style image, you also need to hand over a lot of data to OpenAI, which could be used to train its models.

Hidden Data

The data you are giving away when you use an AI image editor is often hidden. Every time you upload an image to ChatGPT, you’re potentially handing over “an entire bundle of metadata,” says Tom Vazdar, area chair for cybersecurity at Open Institute of Technology. “That includes the EXIF data attached to the image file, such as the time the photo was taken and the GPS coordinates of where it was shot.”

OpenAI also collects data about the device you’re using to access the platform. That means your device type, operating system, browser version, and unique identifiers, says Vazdar. “And because platforms like ChatGPT operate conversationally, there’s also behavioral data, such as what you typed, what kind of images you asked for, how you interacted with the interface and the frequency of those actions.”

It’s not just your face. If you upload a high-resolution photo, you’re giving OpenAI whatever else is in the image, too—the background, other people, things in your room and anything readable such as documents or badges, says Camden Woollven, group head of AI product marketing at risk management firm GRC International Group.

This type of voluntarily provided, consent-backed data is “a gold mine for training generative models,” especially multimodal ones that rely on visual inputs, says Vazdar.

OpenAI denies it is orchestrating viral photo trends as a ploy to collect user data, yet the firm certainly gains an advantage from it. OpenAI doesn’t need to scrape the web for your face if you’re happily uploading it yourself, Vazdar points out. “This trend, whether by design or a convenient opportunity, is providing the company with massive volumes of fresh, high-quality facial data from diverse age groups, ethnicities, and geographies.”

OpenAI says it does not actively seek out personal information to train models—and it doesn’t use public data on the internet to build profiles about people to advertise to them or sell their data, an OpenAI spokesperson tells WIRED. However, under OpenAI’s current privacy policy, images submitted through ChatGPT can be retained and used to improve its models.

Any data, prompts, or requests you share helps teach the algorithm—and personalized information helps fine tune it further, says Jake Moore, global cybersecurity adviser at security outfit ESET, who created his own action figure to demonstrate the privacy risks of the trend on LinkedIn.

Uncanny Likeness

In some markets, your photos are protected by regulation. In the UK and EU, data-protection regulation including the GDPR offer strong protections, including the right to access or delete your data. At the same time, use of biometric data requires explicit consent.

However, photographs become biometric data only when processed through a specific technical means allowing the unique identification of a specific individual, says Melissa Hall, senior associate at law firm MFMac. Processing an image to create a cartoon version of the subject in the original photograph is “unlikely to meet this definition,” she says.

Meanwhile, in the US, privacy protections vary. “California and Illinois are leading with stronger data protection laws, but there is no standard position across all US states,” says Annalisa Checchi, a partner at IP law firm Ionic Legal. And OpenAI’s privacy policy doesn’t contain an explicit carve-out for likeness or biometric data, which “creates a grey area for stylized facial uploads,” Checchi says.

The risks include your image or likeness being retained, potentially used to train future models, or combined with other data for profiling, says Checchi. “While these platforms often prioritize safety, the long-term use of your likeness is still poorly understood—and hard to retract once uploaded.”

OpenAI says its users’ privacy and security is a top priority. The firm wants its AI models to learn about the world, not private individuals, and it actively minimizes the collection of personal information, an OpenAI spokesperson tells WIRED.

Meanwhile, users have control over how their data is used, with self-service tools to access, export, or delete personal information. You can also opt out of having content used to improve models, according to OpenAI.

ChatGPT Free, Plus, and Pro users can control whether they contribute to future model improvements in their data controls settings. OpenAI does not train on ChatGPT Team, Enterprise, and Edu customer data⁠ by default, according to the company.

Read the full article below:

Read the article
LADBible and Yahoo News: Viral AI trend could present huge privacy concerns, says expert
OPIT - Open Institute of Technology
OPIT - Open Institute of Technology
May 12, 2025 4 min read

Source:


You’ve probably seen them all over Instagram

By James Moorhouse

Experts have warned against participating in a viral social media trend which sees people use ChatGPT to create an action figure version of themselves.

If you’ve spent any time whatsoever doomscrolling on Instagram or TikTok or dare I say it, LinkedIn recently, you’ll be all too aware of the viral trend.

Obviously, there’s nothing more entertaining and frivolous than seeing AI generated versions of your co-workers and their cute little laptops and piña coladas, but it turns out that it might not be the best idea to take part.

There may well be some benefits to artificial intelligence but often it can produce some pretty disturbing results. Earlier this year, a lad from Norway sued ChatGPT after it falsely claimed he had been convicted of killing two of his kids.

Unfortunately, if you don’t like AI, then you’re going to have to accept that it’s going to become a regular part of our lives. You only need to look at WhatsApp or Facebook messenger to realise that. But it’s always worth saying please and thank you to ChatGPT just in case society does collapse and the AI robots take over, in the hope that they treat you mercifully. Although it might cost them a little more electricity.

Anyway, in case you’re thinking of getting involved in this latest AI trend and sharing your face and your favourite hobbies with a high tech robot, maybe don’t. You don’t want to end up starring in your own Netflix series, à la Black Mirror.

Tom Vazdar, area chair for cybersecurity at Open Institute of Technology, spoke with Wired about some of the dangers of sharing personal details about yourself with AI.

Every time you upload an image to ChatGPT, you’re potentially handing over ‘an entire bundle of metadata’ he revealed.

Vazdar added: “That includes the EXIF data attached to the image file, such as the time the photo was taken and the GPS coordinates of where it was shot.

“Because platforms like ChatGPT operate conversationally, there’s also behavioural data, such as what you typed, what kind of images you asked for, how you interacted with the interface and the frequency of those actions.”

Essentially, if you upload a photo of your face, you’re not just giving AI access to your face, but also the whatever is in the background, such as the location or other people that might feature.

Vazdar concluded: “This trend, whether by design or a convenient opportunity, is providing the company with massive volumes of fresh, high-quality facial data from diverse age groups, ethnicities, and geographies.”

While we’re at it, maybe stop using ChatGPT for your university essays and general basic questions you can find the answer to on Google as well. The last thing you need is AI knowing you don’t know how to do something basic if it does takeover the world.

Read the full article below:

Read the article