Researchers Beat Google's Bouncer 44
An anonymous reader writes "When earlier this year Google introduced Bouncer — an automated app scanning service that analyzes apps by running them on Google's cloud infrastructure and simulating how they will run on an Android device — it shared practically nothing about how it operates, in the hopes of making malicious app developers' scramble for a while to discover how to bypass it. As it turned out, several months later security researchers Jon Oberheide and Charlie Miller discovered — among other things — just what kind of virtual environment Bouncer uses (the QEMU processor emulator) and that all requests coming from Google came from a specific IP block, and made an app that was instructed to behave as a legitimate one every time it detected this specific virtual environment. Now two more researchers have effectively proved that Bouncer can be rather easily fooled into considering a malicious app harmless."
Re: (Score:1, Flamebait)
Apparently spelling isn't your forte either. And I'm not even going to get started on how much of a racist piece of shit you are.
Re: (Score:3, Insightful)
Please, STOP FEEDING THE FUCKING TROLLS!!! Ignore them For God's sake, don't quote them!!! Jesus, man, what the fuck is wrong with you? Anonymous troll is at -1 so you gave him a voice! Mods, please downmod every response to the troll, including mine but especially the parent's, who stupidly quoted the racist bullshit. Fucking trollbiters are often as bad as the fucking trolls.
pretty easy to fix, though (Score:5, Insightful)
It seems like they just found that the sandbox Google simulates the apps in is a little sloppy in its simulation (IP addresses are predictable), so it's easy to tell you're inside the sandbox. But they could fix that part pretty easily.
Was hoping for something more halting-problem-esque, since it's really difficult to "scan an app for malware" in general.
Re: (Score:2)
Whereas with the "is this malware" problem you're not.
One workaround is sandboxing. From the "halting problem" perspective, sandboxing would be the like setting a time limit so that all programs will halt by a certain time.
Re: (Score:2)
I don't think searching for malware is equivalent to solving the halting problem. For e.g. for a game it's enough to check where it wants to write; if it wants to write outside of it's own directory than it raises red flags. Basically it's enough to analyse what kind of APIs it uses. (The OS sandbox should provide an API that jails your writes to a certain directory.)
Meh (Score:2)
Re: (Score:3, Funny)
I thought bunch of nerds gave a drubbing to a bouncer at Google-sponsored party.
Just out of curiousity, when have a bunch of nerds -- ever -- given a drubbing to a bouncer? (Physical drubbings only please, chicken-shit revenge tactics don't count...)
Re: (Score:3, Informative)
Not all nerds are weak. A guy in my CompSci course actually worked as a bouncer. Really nice guy too - not just someone who was out to beat people up. A bunch of drunken nerds could take a single bouncer if they actually had the motivation. Bouncers tend to have backup though.
Re: (Score:1)
Re: (Score:2)
Let's grow up, shall we?
Community Relations (Score:5, Insightful)
"A renowned security researcher who claims he discovered a flaw in iOS was kicked out of Apple's iOS Developers program."
Just sayin'.
Re: (Score:2)
Yes, because he didn't apprise Apple of the research beforehand. That makes a pretty big difference than having the company be aware you are doing the research and give its blessing.
Re: (Score:2, Informative)
Actually, he did. (Assuming we're talking about Charlie Miller). He did several times and was promptly ignored. I'm sure if you google it, you'll find that out real quick.
Then he made an application that abused said bug silently to prove a point, since nobody was listening.
Re:Community Relations (Score:4, Insightful)
If Apple wants to seriously engage the security community there ought to be a way for the researchers to submit proof of concept apps to the app store to see if their current review process can catch them (obviously the reviewers would need to be blinded as to the identity of the submitter). They could improve their review process, catch security issues, AND avoid the negative press of booting a developer like this.
Re: (Score:2)
My impression was that they kicked him out for submitting the app to the store (for customers to purchase), not for finding the vulnerability.
As did the guys who were testing Bouncer. They put an SMS blocker app on the Google Play Store and repeatedly updated it, adding more malicious behavior each time.
Re: (Score:2)
Re: (Score:1)
He was kicked out for making Apple look bad by allowing any security flaw ever to become public
That's what they want you to think (Score:1)
Actually any malware that's "smart" enough to fool Bouncer is left alone while the NSA, FBI, and MPAA are alerted. Black helicopters full of hot women in black latex arrive...
News Flash (Score:1)
News Flash: Any automated security system can be beaten.
In further news, using technology to secure against technology is only as effective as the minds behind it.
Tune in at 11.
Require signed apps? (Score:2)
Re: (Score:2, Troll)
They already do that, unless the user decides to turn it off.
Any other ideas you want to share that are already in use?
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Of course it's because of the slew of people wouldn't bother paying for apps they could build and install themselves. Or those who would take it for their own, fork it, make it better, make it worse...
Surprised??? (Score:1)
1. Inside Google - A bunch of college boys (no girls, as they are not smart enough for google), very, extremely good at solving entry interview quiz and questions, but extremely poor and incompetent at actually doing what they were hired to do, DEVELOPING.
2. Outside Google - A bunch of software developers, usually old, with a lot of experience, some of them even PhD, and who are actually DEVELOPING a software that google buys, because their bunch is so incompetent...
So to sumariz
Re: (Score:2)
Hmmmmmmm.... (Score:3)
It's almost as though they're trying to achieve security by making information about their service very obscure. Has anyone ever tried this before?
it's another layer (Score:2)
As long as you know what you're doing, obscurity can work just fine as another layer of protection.
The problem is that most people choosing obscurity aren't secure to start with, so it's the *only* layer of protection.
This is the best approach to security... (Score:1)
Google was aware of and blessed the research, and has been apprised of its results so that it can make changes and better secure Google Play against malicious individuals.
2 big mistakes (Score:2)
1 - not using random proxies
2 - not going out of their way to make the VMs look like real machines. This is already a problem with PC viruses, many of them are designed not to infect a VM to slow analysis.
"Bouncer" provably cannot win (Score:2)
So if Google wants to keep malware out, Bouncer is fundamentally the wrong approach.