Do you really know what’s inside your iOS and Android apps?

It’s time to review your code as it is possible that no-code or low-code features that are employed on iOS or Android applications might not be as safe as you believed. That’s the major conclusion from a study that revealed Russian software is being utilized in apps developed by organizations like the US Army, the CDC, and the UK Labor Party and other organizations.

When Washington transforms into Siberia

The issue is that the software developed by a company known as Pushwoosh is being used in thousands of apps by thousands of companies. This includes those from the Centers for Disease Control and Prevention (CDC) and says they were misled to think that Pushwoosh was situated in Washington but the company’s founder is actually located in Siberia as explained by Reuters. A look at the Pushwoosh Twitter account shows its claim to be located in Washington, DC.

The company offers data processing and code support which can be integrated into apps to monitor the activities of smartphone users online and to send customized notifications. CleverTap Braze, One Signal, and Firebase provide similar services. To be fair, Reuters has no evidence that the information that it collects was misused. However, it is true that the firm is headquartered in Russia is an issue, since the data will be subject to the local laws regarding data that could create the risk of security.

Stay informed of the most recent IT thoughts, trends, concepts about how-to’s, and more by reading Computerworld’s newsletters. [

Of course not however it’s highly unlikely that anyone involved in the handling of data that’s sensitive would be willing to accept that risk.

What’s the story behind it?

There are many reasons to be skeptical of Russia in the present I’m pretty sure that every nation has its own third-party component creators that may or might not put security first. The trick is to determine the ones that do and those that don’t.

The reason for code such as this is from Pushwoosh is utilized in apps is straightforward: it’s about time and money. Mobile application development can be expensive and to cut down on the cost of development, certain applications will make use of common third-party code to perform certain tasks. This reduces costs, and as we’re moving quickly towards low-code or no-code development environments, we’ll likely to be seeing more of this type of brick model approach to app development.

It’s okay because modular code can provide huge benefits to developers, apps, and companies, but it also raises a problem that every business using third-party software should examine.

Who is the owner of your code?

What is the extent to which the code safe? What information is collected by the code, and where is it stored and what authority does the user (or the company whose name appears on the application) have to safeguard and manage, erase, or delete the data?

Other issues to consider When the code is being used is it frequently updated? Are the codes themselves safe? What level of rigor has been employed in testing software? Does the program contain any unidentified tracker code for scripts? What kind of encryption is employed and where are the files is it stored?

The issue is that if the answer to one of the above questions is “don’t know” or “none” The information is in danger. This is why it is imperative to conduct thorough security checks for using any component code that is modular.

Data compliance teams need to test the material thoroughly The data compliance teams must test this material thoroughly – “minimal” testing is not enough.

I also think that a method whereby the data collected is anonymous is a good idea. This way, if information leaks out and is misused, the risk is greatly reduced. (The risk of using custom-made technologies that do not provide adequate protection to information during the exchange is that the information when it is collected becomes a security risk.)

The consequences of Cambridge Analytica illustrate why obfuscation is necessary in this connected world.

Apple definitely seems to be aware of this danger. Pushwoosh is included in more than 8000 iOS and Android applications. It is crucial to remember that the company claims that the information it gathers will not be stored by Russia however this might not shield it from being stolen according to the experts quoted by Reuters clarify.

In a way, it’s not that important because safety is about anticipating risk instead of waiting for the danger to happen. With the number of businesses that go into bankruptcy after being attacked It is best to be prepared instead of sorry with security policies.

That’s why companies that have teams of developers that depend on code from the shelf must make sure that any third-party software is compliant to the company’s security policy. Since it’s the code that you wrote, that’s your brand name and any misuse of the information due to inadequate tests for compliance is your issue.

1 thought on “Do you really know what’s inside your iOS and Android apps?”

Leave a Comment